You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have two versions of SSDLite (ssd_mobilenet_v1_coco) with same backbone
One version is implemented all components (feature extractor, bbox decoder, and NSM) separately. Feature extractor is implemented by CoreML and both bbox decoder and NSM is by bit of swift code running on CPU. Other version is combined every module in MLModel as you explained
I checked the performance of two versions on iPhoneX and got a result hard to understand.
Processing time of the first version is two times faster than the order one. Although the first one uses more CPU resource to calculate real bbox coordinates and NMS (uses 100 ~ 110% of CPU resource. the other consumes 45 ~ 50%) it's strange in common sense.
Could you give me any opinion or understanding about this status?
Thank you.
The text was updated successfully, but these errors were encountered:
minhohihi
changed the title
Question about performance of SSDLite wit Core ML
Question about performance of SSDLite with Core ML
Dec 28, 2018
I didn't actually do any speed tests (hehe) but 2x speed difference is quite a bit! I wouldn't have expected that...
Doing these operations inside the model will run them on the GPU, or some on the GPU and some on the CPU. My guess is that doing this on the GPU might be slower than on the CPU. Because on the CPU it's just a single loop while on the GPU it's split up into separate layers that all do different calculations.
It might be useful to do a follow-up blog post that compares these three methods on different devices:
Do the decoding and NMS on the CPU afterwards.
Do everything inside the model.
Do the decoding in a custom layer (so it can run on the GPU) and do NMS on the CPU afterwards.
First of all, always thank you for your blog.
I have a question about performance.
I have two versions of SSDLite (ssd_mobilenet_v1_coco) with same backbone
One version is implemented all components (feature extractor, bbox decoder, and NSM) separately. Feature extractor is implemented by CoreML and both bbox decoder and NSM is by bit of swift code running on CPU. Other version is combined every module in MLModel as you explained
I checked the performance of two versions on iPhoneX and got a result hard to understand.
Processing time of the first version is two times faster than the order one. Although the first one uses more CPU resource to calculate real bbox coordinates and NMS (uses 100 ~ 110% of CPU resource. the other consumes 45 ~ 50%) it's strange in common sense.
Could you give me any opinion or understanding about this status?
Thank you.
The text was updated successfully, but these errors were encountered: