-
Notifications
You must be signed in to change notification settings - Fork 18.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
strange training log of imagenet #1102
Comments
I also checked the training accuracy, it gets to 100%!!! |
I had tried it in other data set, it looked fine, the only thing I'm feeling not good is that the accuracy not reach to my expectation. |
How did you generate the data? Did you use convert_imageset.cpp to resize the image and then convert it to leveldb? |
I generate the data by call the tools convert_imageset.bin and compute_image_mean.bin that located in build directory, the detail steps followed the official tutorial. |
And when you type convert_imageset.bin abd compute_image_mean.bin in command line with the option --h, it will show the usage info, I followed the info and did it. |
I following the tutorial strictly as well... weird... |
Well, I forget the details! I think maybe the two approaches are same, but I suggest first resize the images to 256*256, in this way, we can save time for convert_imageset.bin. I have little remember that I maybe using this approach! But I can't sure, because when I do this preprocessing, I tried many times. |
problem solved, it's the problem of the data... |
Hi,
I am using the default model (imagenet_train_val.prototxt) to train imagenet. After many iterations, I got something strange:
I0917 15:36:23.184696 2254 solver.cpp:270] Test score #0: 0.00699999
I0917 15:36:23.184826 2254 solver.cpp:270] Test score #1: 15.3443
I0917 15:36:24.644654 2254 solver.cpp:195] Iteration 30000, loss = 0.00245684
I0917 15:36:24.644719 2254 solver.cpp:365] Iteration 30000, lr = 0.01
I0917 15:36:59.604066 2254 solver.cpp:195] Iteration 30020, loss = 0.00182616
I0917 15:36:59.604195 2254 solver.cpp:365] Iteration 30020, lr = 0.01
I0917 15:37:34.563117 2254 solver.cpp:195] Iteration 30040, loss = 0.000589138
I0917 15:37:34.563244 2254 solver.cpp:365] Iteration 30040, lr = 0.01
I0917 15:38:09.522680 2254 solver.cpp:195] Iteration 30060, loss = 0.00313978
I0917 15:38:09.522809 2254 solver.cpp:365] Iteration 30060, lr = 0.01
I0917 15:38:44.481019 2254 solver.cpp:195] Iteration 30080, loss = 0.00256942
I0917 15:38:44.481150 2254 solver.cpp:365] Iteration 30080, lr = 0.01
I0917 15:39:19.437052 2254 solver.cpp:195] Iteration 30100, loss = 0.000853064
I0917 15:39:19.437180 2254 solver.cpp:365] Iteration 30100, lr = 0.01
I0917 15:39:54.397054 2254 solver.cpp:195] Iteration 30120, loss = 0.00521982
I0917 15:39:54.397181 2254 solver.cpp:365] Iteration 30120, lr = 0.01
The test score #0 stays around 0.005 from the every beginning and does not change much during the long process.
Moreover, the test score #1 is around 15 which is much greater than 0.005.
Did anyone have this problem? what's the potential causes for this?
Can someone share me the training log?
thanks,
The text was updated successfully, but these errors were encountered: