-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Accelerating the inference of the trained model for COCO-WholeBody #520
Comments
Hi thanks for your interest! The first step is to do profiling. Identify the bottleneck and improve it. Steps:
You can post the profiling image here Related: #73 (comment) |
It is possible. Since only 28 keypoints are needed, not all heatmaps have to be transformed to keypoints. The post-processing (from heatmaps to keypoints) can be accelerated. You may need to add a new keypoint head, and modify the post-processing part.
|
@jin-s13, I use a copy of |
@jin-s13, I have make rough transform Question: Is it normal acceleration or I can accelerate inference more? |
Not very sure about that. But you may follow @innerlee to do profiling, and check which part costs the most of time. You may also modify the network structure, especially the keypoint head to decrease the output channel from 133 to 28. By the way, have you tried these tips to speed up inference? |
Profiling is a good idea, but I still believe that I can accelerate inference by modifying the network structure. Yes, I tried tips to speed up inference. |
Then modify |
jin-s13, thank you very much for the help! |
Hi!
I want to accelerate the inference of the trained model for COCO-WholeBody.
I have ideas:
This model returned 133 keypoints. So I think that have a possibility to accelerate the trained model by decreasing returned amount of keypoints. For example 17 for body, 6 for feet, 3 for face, and 2 for hands = 28 of 133.
The question: Is it make sense? If YES, then how to customize the output of the trained model?
If I write my personal script for the inference powered by MMLab API, is it accelerate the process?
The text was updated successfully, but these errors were encountered: