You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
you index the output of the transformer with [0]. Does this not take the output of the first layer of the transformer decoder, instead of the last layer? And is this behaviour expected?
The text was updated successfully, but these errors were encountered:
@CarlDegio hello,I changed 0 to -1 and used the output of the last layer, but there is no obvious difference in the effect. May I ask that your experiment is effective?
I did not try the effect of multi-layer decoder. I just removed the decoder forward propagation that was not used in the original code to speed up the training. @uuu686
JINXER000
pushed a commit
to JINXER000/act_interleaved
that referenced
this issue
Sep 19, 2024
In
act/detr/models/detr_vae.py
Line 131 in 694c606
act/detr/models/detr_vae.py
Line 136 in 694c606
The text was updated successfully, but these errors were encountered: