You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi! I think there's a bug at this line in the forward function. Specifically, if the attention tensor attn_out is as follows for an input image with shape (channels, h(=2), w(=3)) and self-attention channels dv = 2:
# attention values of the 6 pixelsAtttensor([[-3.5002, -1.2102],
[-4.3694, -1.5107],
[-4.7621, -1.6465],
[-4.9178, -1.7003],
[-2.2335, -0.7722],
[-5.0056, -1.7307]], grad_fn=<SliceBackward>)
I can confirm that this is a bug and that you solved it. can disregard my original question. The issue was because on reshape we needed to have H*W as the last dimension
Hi! I think there's a bug at this line in the
forward
function. Specifically, if the attention tensorattn_out
is as follows for an input image with shape(channels, h(=2), w(=3))
and self-attention channelsdv = 2
:you should not reshape it directly using
but instead you should use
The output difference:
vs.
Hope it helps!
The text was updated successfully, but these errors were encountered: