-
Notifications
You must be signed in to change notification settings - Fork 5.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
is trainer_count work when infer with CPU? #7218
Comments
Which paddle version do you use? You can use |
paddle version info:
|
从版本信息看,with_avx是ON的时候,MKL库不应该被关闭(请问您是从哪儿安装的paddle)。
其中THREADS=cpu核数/trainer_count,如果cpu核数 < trainer_count, THREADS=1. |
谢谢,性能好很多了 |
Merged
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
I use pre-trained ResNet50 model to do infer job on CPU machine, the code is mainly copied from PaddlePaddle/models/image_classification.
I found that when i set trainer_count = 4, it became slower than i set trainer_count = 1. the test_data size is 64.
when trainer_count = 1, the infer job cost 0.3s per img in a machine with 32 CPU cores, almost the same as cost time(0.3) in my notebook. Not faster?
the
top
command shows that 16 cores is busy and the other 16 cores are idle. so i run two other infer job likepython infer.py
at the same time. it cost 0.5s per img per job, about twice times better than one job only. But, can paddle do it self?The text was updated successfully, but these errors were encountered: