You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'm thinking hard and long about selecting a platform to deploy an assembly-of yolo models on-to.
And surface-pro with the snapdragon elite chip comes to mind. But i have no confidence in it.
I hoped Qualcomm would have had all NPU drivers 100% ready when they started shipping the Copilot Laptops. BUT that's not the case. DirectML is not 100% supported. Not all ONNX models run without errors, only Qualcomm approved ones from their AI Hub. The latest version of the demo supports running some models on the GPU (Press G). A Yolov8s-seg.onnx model, not approved by Qualcomm runs ok on the GPU, but with errors on the NPU.
And something about the speed, my 5 year old Dell precision, with an nVidia RTX 3000 graphics adapter runs these yolo models as fast as the Qualcomm NPU. I guess the speed Qualcomm advertises is from running quantized (int8) models.
I made some sample code to show how to use the NPU to run a yolo model on mp4 files.
Currently it runs in real time on my Snapdragon X Elite Dev Box. Twice as fast as the Yolov4 GPU DirectML sample while using less than half the power.
You can see it here: https://github.com/fobrs/yolov9_npu
Should I make a pull request?
The text was updated successfully, but these errors were encountered: