-
Notifications
You must be signed in to change notification settings - Fork 51
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Should implementing __imatmul__
really be enforced?
#509
Comments
I think it's just a simple consistency issue. If all operators have in-place versions, then
I'd say yes - that seems like the only consistent option for the semantics to me. It won't be used much, and yes it won't be faster than the out-of-place version. But I don't think that's a good enough reason to not add it. The semantics are well-defined. |
I don't think it is required. The spec says in-place operators "may" be supported with Also, I thought I remembered it only being required for instances where the left-hand array shape wouldn't change (i.e., the right-hand operator is square), but I can't find that now. |
Yeah just to back up Aaron's comment. From the Python docs:
|
Also https://data-apis.org/array-api/latest/design_topics/copies_views_and_mutation.html#copyview-mutability. When views are avoided, Should the spec say that if views are implemented then in-place operators should actually act in-place? To me it's not necessarily clear that it says that. |
@asmeurer, that is what tensorflow does, but I think it is misleading. In NumPy:
is always an error. This ensures that we don't have the one odd "in-place" operator which isn't in-place at all. I do think that is important. So the question is whether a library can expect |
OK, I am unsure that it is very useful to support it, but it seems OK. There is one broadcasting corner case that needs to be decided on in NumPy (which must have been part of the reason why I was unsure this is actually useful):
is a bit strange. But at least in-place behavior may be useful for stacks of small matrices (in some future). |
Can you explain what's going on in your example? How do you end up with |
Note that the PEP only mentions broadcasting for >2 dimensions, and we also say the same in the spec (because contracted dimensions should never broadcast). |
It's an example if NumPy were to implement this (doesn't work today). This demonstrates the behavior that it would have import numpy as np
a = np.array([1., 2.])
a1 = a.copy()
np.matmul(a1, a1, out=a1) The issue is the scalar result gets broadcast and assigned to the whole array. In any event we are discussing potentially raising an error in this case (as it is odd behavior). |
Oh I understand now. I never realized that In general, contracted dimensions shouldn't broadcast, and to me it makes sense to extend this to the "out" broadcasting that happens in the ufunc (and especially so if that ends up being the semantics for The spec says:
So it would already disallow this behavior. But it's probably worth adding a note clarifying what is required for |
__imatmul__
be really be enforced?__imatmul__
really be enforced?
I was not around when this may have been discussed in numpy/numpy#21912. In-place matmul is a weird beast, but for NumPy it would be strange to make
@=
and out-of-place operator.Because of this, NumPy doesn't define it currently, and I am unsure that it should be done. Would NumPy have the tight restriction that the shape must fit and we actually assign back to
a
(i.e. truly in-place at all operators?).That would make sense, but even then, it would even be slower and use as much memory as before anyway.
The text was updated successfully, but these errors were encountered: