GPU setup for running SLEAP on a large dataset of videos #1942
Replies: 1 comment
-
Hi @hemanyaradadia, Sorry about the delay! That's a large dataset indeed! Looking at some specs:
It's hard to say how these will play out in a real setting, but it might be the case that the A4500 could have a bit more performance with its tensor cores somehow. On the other hand, the A40 has a lot more memory, so you'll be able to increase the batch size more to increase throughput. As you mentioned, another advantage is that if you can do virtual parallelization (GPU fractionalization) to run multiple models in parallel, the A40 has more memory and bandwidth to support that and will get you some gains in terms of data parallelism on the same device. Given that you have many small videos, one thing I'd strongly recommend is using the Python API to do the inference yourself rather than using the In all cases though, you'll likely be bottlenecked by the I/O, so definitely consider running as many tracking processes in parallel as your I/O path will permit. For example, if your data is sitting in the network, then make sure you're saturating (but not exceeding) your network card's bandwidth or your storage system's read rates. Let us know if you have any questions and keep us posted on what solution you land on! Talmo |
Beta Was this translation helpful? Give feedback.
-
Hello everyone,
I am working on projects involving massive datasets like >10k videos (several seconds long) and 5-week-long high-resolution videos, and I plan to use SLEAP for estimating pose from these videos. I am looking for suggestions for a computing infrastructure suitable for such large-scale projects.
What would be the best GPU setup to run inference on many videos as fast as possible (with a trained model)? For example, a single powerful GPU like A40 or many less powerful ones like A4500? Does virtual parallelization improve performance on a single GPU?
Beta Was this translation helpful? Give feedback.
All reactions