You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi @LTH14! Congrats on your nice work being accepted by CVPR.
Just as the title, I'm confused why you choose to use an encoder-decoder architecture like MAE? Have you ever tried using a encoder only arch like BEiT?
The text was updated successfully, but these errors were encountered:
We haven't tried an encoder-only structure like BEiT. The reason why we chose the MAE-style enc-dec structure is simply that they were the sota method at that time. Also, an encoder-decoder structure enables us to decouple the representation learning from the generation.
Hi @LTH14! Congrats on your nice work being accepted by CVPR.
Just as the title, I'm confused why you choose to use an encoder-decoder architecture like MAE? Have you ever tried using a encoder only arch like BEiT?
The text was updated successfully, but these errors were encountered: