-
Notifications
You must be signed in to change notification settings - Fork 123
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
When the network reaches convergence, The loss has been around 2.0 #47
Comments
And the mIOU is 0.323895 |
Because you set the flag |
@hellochick I get it.My batch size is 2 because my GPU is limited. So I will try again. Thanks a lot. |
@hellochick So I should run by python train.py --is-training or python train.py --is-training --train-beta-gamma |
Hi, i have the same situation that the loss is about 2.0. Did you use pretrained model? @Hukongtao |
@ningscapr I didn't use pretrained model. When I use the pretrained model, the initial loss is 0.2, but the training is done, the loss is about 2.0. I don't know why |
@Hukongtao Hi, have you solved your issue? I followed hellochick's advice del the --update-mean-var but found the loss was still around 2.0 after 90k iterations. I use cityscapes dataset and the batchsize is 2. |
@waterputty No, I didn't |
@Hukongtao My latest finding is the loss went down to 1.7 after 150k iterations. I think for the cityscapes it may needs much more iterations than 60k in the code, or the learning rate must be set much bigger than 1e-3 which is also from the original code. But I'm not sure. @hellochick Could you give more advice. Thx a lot. |
@waterputty Have you solved your issue? I am still confused |
@hellochick can you show your train args? your read.me do not have training introduction |
@DenceChen i am also having the same problem my training args are : BATCH_SIZE = 1 i cant set up the Batch for less because it wont run even on a 1080TI gpu with 11gb of memory |
I'm having the same issue. The best I'm able to get is 1.7 loss and ~.3 mIOU. @hellochick I want to set this as a baseline, so I'd like to get up to the best possible. |
@AmeetR I think It's hard to reproduce this repo. I ran it 1 month ago, debugged it, and failed. |
python train.py --is-training --update-mean-var --train-beta-gamma
to run the network, but when the network reaches convergence, The loss has been around 2.0.
How did you get 0.2?
@hellochick
Thank you very much for your reply.
The text was updated successfully, but these errors were encountered: