-
Notifications
You must be signed in to change notification settings - Fork 4.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Floating point exception #159
Comments
I encountered a similar problem. Solving... |
I got the same error and it turned out that I was feeding in empty boxes array. Filtering out roidb properly fixed my problem. |
what does "filtering out roidb properly" mean?Would you please give us more details? |
I've got the same error. By changing the RNG_SEED default value I get error in different iterations. Have you guys found the solution yet? @weichengkuo , I would be thankful if you please elaborate a little bit more. Where should I filter the empty boxes? Thanks! |
take a look #65 |
It's possible that some layer of your faster RCNN receive no boxes at some iteration. I ran into this error multiple times and it's often due to empty boxes. Filtering roidb means to remove the roidb elements that could cause this problem. |
how to solve, please? |
pad 0 the original image to reasonable aspect ratio (600*1000) will solve this problem. |
@morusu So where do we need to modify to 'pad 0s the original image' ? |
How to fix the code to do 'pad 0 the original image', or still need to preprocess the images first?. |
@buaaliyi @LiberiFatali preprocess the images first, pad 0 to images' right-side or down-side to reasonable aspect ratio will be fine. |
I got this error while using old code. This problem is solved for me by applying
in |
@LiberiFatali Thanks, your solution solved my problem! |
@vra Where did you apply the filter_roidb function? It is already called in train_net() function (fast_rcnn/train.py). I am facing the same problem as @morusu described. Suddenly my loss goes to nan (overflow encountered in exp). I am using PascalVoc dataset and have no clue about the problem. Anyone solved this issue? Thank you! |
Hi @fernandorovai , |
@vra Hello, does it go well when you add the filter_roidb to train.py? In my case, there is the function of filter_roidb, but I have the problem of 'floating point exception'. I tried to change the learning rate and the RNG_SEED, but it does not go well. |
@hyunjun-jo hello,I have the same problem,too.I tried to change the learning rate and the RNG_SEED,but it does not go well,too.Have you solved the problem? thx |
@morusu @wait1988 @weichengkuo @smasoudn @smichalowski I0312 16:25:25.883342 2983 sgd_solver.cpp:106] Iteration 0, lr = 0.0005 I try to change lr from 0.001 to 0.0001,but it didn't work.I also change RNG_SEED,and it also didn't work. |
Have anyone solved the problem? I get the same error at iteration 5800 while using the learning rate at 0.001 and at iteration 18800 while using 0.0001..If someone have solved the problem, please help me to solve it. |
I have solved my 'Floating point exception (core dumped)' problem by modifying the function 'is_valid' in function 'filter_roidb' in file da-faster-rcnn-master/lib/fast_rcnn/train.py: def filter_roidb(roidb):
|
after thousands iterations, faster-rcnn throw a error "Floating point exception " at ./experiments/scripts/faster_rcnn_end2end.sh . I search the error saying about i/0 or i%0, anyone encountered this?
The text was updated successfully, but these errors were encountered: