-
Notifications
You must be signed in to change notification settings - Fork 11
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
code can not achieve result in paper #5
Comments
did you change any thing in the script? Can you show me the script you used for Visda? Thanks. |
So, I think it's important to releasing baseline script/code for your backbone is different from standard ViT backbone |
what do you mean by "baseline script/code"? |
I change data_loader to my dataload function, which include CUDA_VISIBLE_DEVICES=7 python main.py --train_batch_size 64 --dataset visda --name visda --source_name train --target_name target --root_path /home/wendong/dataset/Vis2017 --source_list data/office/webcam_list.txt --target_list data/office/amazon_list.txt --test_list data/office/amazon_list.txt --num_classes 12 --model_type ViT-B_16 --pretrained_dir checkpoin |
sorry, it was a misexpression. I mean code for implementing 【source only】 result. |
Thanks. Let me test office-31 and office-home and let you know the result soon. |
Thanks for your work. code for [source only] is important for implementing a transformer backbone network😂(especially for office datasets). |
I use |
I'm running the sourceOnly code on both office and office-home datasets and will post my results here once finished. The script I use is like this: The results are far from what are reported in the paper, please advise potential reasons for the failure of reproduction. |
This is so weird. I will double check it and let you know ASAP. Thanks. |
Hi, can you @hellowangqian @ShiyeLi follow the following requirement to rebuild your environment and try again? Thanks. |
Sure, I'll set up a new environment following yours and re-run the experiments to see what happens. May I ask @viyjy if you use the same code in the repo (e.g., cloning from the repo as I did) to reproduce the results above? I ask this to check the possibility that some unnoticed changes have been made when you uploaded your code to GitHub. |
Yes, the result in #5 (comment) is obtained by downloading the code from this repo and run it again. What kind of machine are you using? |
Ubuntu 20.04 + Nvidia Titan RTX GPU |
Are you using a single GPU to train the model? |
Yes. |
The only difference is that my ubuntu version is 18.04, but I don't think it makes a difference to the result. |
Thanks for clarifying the details. I'll spend more time investigating the issue. |
hi , I download the code and dataset again from this repo and run it without any modification. however , still can not reproduce baseline result in paper. My environment is a little bit different from yours.(caused by my CUDA version in servers.) But i don't think it's the primary cause. |
Thanks, let check this. |
@ShiyeLi Hi, would you please send the data.zip in Datasets to [email protected]? I wrongly deleted it from my google drive yesterday. Thanks. |
I have send this zip file, have you receive that? |
Yes, thanks. |
Sorry for the late reply, I still cannot reproduce your results. May I know which pre-trained ViT are you using? Thanks. |
I use the pre-trained model 'ViT-B_16.npz' you provided in this repo. |
Thanks. I will test this on another machine and will let you know by the end of today. |
@hellowangqian @ShiyeLi which Pretrained ViT are you using? |
I used ViT-B_16.npz previously. Now I can reproduce the SourceOnly results for OfficeHome in the paper by using ImageNet21K_ViT-B_16.npz. |
Thanks. I have tested this code on another machine by downloading the repo, building the environment, and downloading the dataset all from scratch, but still cannot reproduce the issues you reported. |
I guess all results reported in the paper are based on the ImageNet21K_ViT-B_16.npz pre-trained model, right? If so, It's expected to have lower performance when ViT-B_16.npz (pre-trained on ImageNet1K) is used. If you can get better results than what I shared above using ViT-B_16.npz (ImageNet1K), could you please share them here for reference? Thanks. |
Thanks, I haven't tried reproducing TVT* yet. What I got was for SourceOnly* (sourceOnly with ViT-B_16.npz). Since I've managed to reproduce SourceOnly results, my SourceOnly* results above should be close to yours if you have these results. |
Sure, please let me know if you have any questions. BTW, I will upload the swin-T and swin-S version soon, which are more comparable to ResNet-50 and ResNet-101. |
I run some script in ‘script.txt’, but get result far away from result reported in paper.
Here are some result:
Office-home
Pr->Cl:52.37
Cl->Pr:78.64
Cl->Ar:72.39
Office-31
AD:94.98
AW:94.21
DW:76.78
Visda17
Train->Val:80.78
ps: apex didn't work well for some installation error, but i think it have no effect on result.(have effect on GPU memory and accumulation.)
(【
fused_weight_gradient_mlp_cuda
module not found. gradient accumulation fusion with weight gradient computation disabled.】 I didn't find a resolution for this error)The text was updated successfully, but these errors were encountered: