-
Notifications
You must be signed in to change notification settings - Fork 287
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Exemplar MAE with DDP does not work #1775
Comments
Hi, thanks for raising the issue! This is indeed wrong in the example. You have to set If for some reason you cannot use
Finally, if you want to reproduce results from the paper I suggest you follow the more complete implementation here:
Regarding the positional embedding, MAE uses a fixed 2D sin-cos positional embedding and the corresponding parameter has set
If I remember correctly DDP expects that all parameters receive an update even if |
@mcleod-matthew-gene In the paper you will see that they also use sinusoidal positonal embeddings Masked Autoencoders Are Scalable Vision Learners – Appendix A.1, ViT architecture
Therefore I would also suggest to proceed in suggested way above. |
Hello all,
Thanks for the great open source package. I noticed the example for MAE with pytorch lightning training DDP simply does not work? There is an to be an issue with unused parameters. i.e.
If you print which parameters are do not have a gradient, you'll see they are
vit.pos_embed
,vit.head.weight
,vit.head.bias
. The unused head parameter makes sense, but I don't see why thevit.pos_embed
would be unused.I'd really appreciate if could confirm this is an issue with the example on main and if fixing this will be on the roadmap.
Thanks!
The text was updated successfully, but these errors were encountered: