Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Some questions about training #9

Open
zhLawliet opened this issue Mar 29, 2022 · 2 comments
Open

Some questions about training #9

zhLawliet opened this issue Mar 29, 2022 · 2 comments

Comments

@zhLawliet
Copy link

zhLawliet commented Mar 29, 2022

@ziniuwan
1: you mentioned that “Use the last checkpoint of stage 1 to initialize the model and starts training stage 2.”
but there is "# We empirically choose not to load the pretrained decoder weights from stage1 as it yields better performance." in code, so will not use the pretrained weights of decode form stage1. howerver the MODEL.ENCODER.BACKBONE is "cnn" in stage1, wihch is "ste" in stage2, so we will also not use the pretrained weights of encode form stage1. so neither encode nor decode is used,why do we still need to do pre-training for stage1?

image

2: what does this operation mean?, i think it is similar to "x = x + self.pos_embed"
x = x.reshape(-1, seqlen, N, C) + self.temp_embed[:,:seqlen,:,:] in vision_transformer.py

@zhLawliet zhLawliet reopened this Mar 29, 2022
@zhLawliet
Copy link
Author

@ziniuwan where can we find the supplmentary Material of the paper?
image

@zhLawliet
Copy link
Author

zhLawliet commented Apr 24, 2022

@ziniuwan i find the ROT_JITTER is 0, Is this profile the one you trained on? and the sample_freq is 8, why not it is 1?
image

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant