-
Notifications
You must be signed in to change notification settings - Fork 566
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
CUDA error: an illegal memory access was encountered #405
Comments
Please run python |
fatal: Not a git repository (or any parent up to mount point /data2)
TorchVision: 0.9.0 |
Please try the blow suggestions:
to
Because there may exit the bug in cuda function- |
Thank you very much for your advice. Unfortunately, I've tried all of your suggestions, and the results show that sometimes they work and sometimes they don't. In order to test smoothly, my solution was to skip the images that would cause errors. Obviously this is not a perfect solution, so I hope someone can fundamentally solve this problem. Thank you! |
I guess it is because some sub-image in DOTA2.0 contain many objects, which causes some cuda operators to take up a lot of memory. |
This could be the reason, and it also happened when I used DOTA1.0. |
Hi, I've faced this too, any workaround? (I ran using V100 GPUs) |
A successful solution: set smaller
|
@yangxue0827 I still get this error even with that change:
Even tried with the V100's that are configured with 32GB of memory instead of 16GB (AWS p3dn.24xlarge instances) |
set smaller nms_pre also occur this bug, any updates now? |
|
Its worked for me, great!!! |
I have the same problem with you. I am also a V100. I have no problem in training, but this problem will occur when I test,Have you solved it yet |
I met the same problem when verifying on the test set. I tried to modify the image size to (960, 960) and nms_pre=1000, but sometimes it works and sometimes it doesn't. May I ask if you have solved this problem now? |
I met the same problem and it didn't work with suitble nms_pre. So any update of this bug now? |
I met the same problem too, anyone can help? |
Seems there is no feasible solution |
First of all, thank you very much for your work.
But it always reports the following error when testing rotated_reppoints::
File "/data2/S/RepPoints_oriented/mmrotate-0.2.0/mmrotate/models/dense_heads/rotated_reppoints_head.py", line 1157, in _get_bboxes_single
scale_factor)
RuntimeError: CUDA error: an illegal memory access was encountered
How can this be solved? I tried different mmrotate versions and the solutions mentioned in issues, but none of them worked. I really hope to get your help Thank you
The text was updated successfully, but these errors were encountered: