-
Notifications
You must be signed in to change notification settings - Fork 9.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Enhance] Continue to speed up training. #6974
Conversation
Codecov Report
@@ Coverage Diff @@
## dev #6974 +/- ##
==========================================
- Coverage 62.34% 62.34% -0.01%
==========================================
Files 327 327
Lines 26129 26129
Branches 4424 4424
==========================================
- Hits 16290 16289 -1
- Misses 8970 8971 +1
Partials 869 869
Flags with carried forward coverage won't be shown. Click here to find out more.
Continue to review full report at Codecov.
|
Does |
Yes, it works. I'll add this both in train.py and test.py. |
* [Enhance] Speed up training time. * set in cfg
* [Enhance] Speed up training time. * set in cfg
* [Enhance] Speed up training time. * set in cfg
Motivation
This PR continues on #6867.
Not only limit the opencv multi-processing but also set OMP and MKL threads to 1 if not set in the environment.
Also, switch the start method from spawn to fork to speed up the start time.
Comparison
V100 x8
system info:
YOLOX-s
launcher: slurm
workers per GPU: 8
file client: s3
Faster RCNN
launcher: slurm
workers per GPU: 2
file client: s3
A100 x8
YOLOX-s
launcher: torch
workers per GPU: 8
file client: hard disk