-
-
Notifications
You must be signed in to change notification settings - Fork 73
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The loss on COCO dataset of training stage1 dosen't decrease #5
Comments
The learning rate = global learning rate * layer learning rate. |
I also notice there is error in your code? @habbakuk1103 |
I have solved the problem. Thank you for your suggestion. |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Hi, I run the train_coco_word2_1_pool.m, but after more than 10 epochs, I found the train result is still bad(as below). I didn't change any hyper parameters but I don't know why it doesn't work.
Can you tell me what make this result happen? And I found the learning rate in your code is 0.1, but it's reported as 0.001 in the paper. Which lr is correct and better in this task?
The text was updated successfully, but these errors were encountered: