Replies: 3 comments 5 replies
-
👋 Hello @MarcoMagl, thank you for reaching out to the Ultralytics community 🚀! We understand that the pre-processing phase can be time-consuming and appreciate your detailed description. For performance issues, especially related to data transfer and caching, please provide a minimum reproducible example if possible. This will greatly assist our team in diagnosing the issue more effectively. In the meantime, make sure to check out our Docs for potential optimizations and guidance on improving the data pipeline. You might also want to review our Tips for Best Training Results to see if there are any applicable strategies. To stay connected and exchange ideas with other users, join our Discord community 🎧. For more detailed discussions, you can also explore our Discourse forum or participate in conversations on our Subreddit. UpgradeEnsure you are using the latest pip install -U ultralytics Make sure your environment meets these requirements, ideally in a Python>=3.8 environment with PyTorch>=1.8. EnvironmentsConsider running YOLOv8 in one of our verified environments to potentially enhance pre-processing speeds:
StatusThis is an automated message and an Ultralytics engineer will assist you soon 😊. Thank you for your patience and cooperation! |
Beta Was this translation helpful? Give feedback.
-
@MarcoMagl to improve data transfer speed, consider using faster storage solutions or optimizing your data pipeline. You might also try reducing image size or format for quicker loading. For more detailed profiling, tools like PyTorch's built-in profiler could be helpful. If you need further assistance, our documentation at https://docs.ultralytics.com/guides/yolo-common-issues/ may offer additional insights. |
Beta Was this translation helpful? Give feedback.
-
Hi @glenn-jocher, thanks for answering. |
Beta Was this translation helpful? Give feedback.
-
Hello to everybody,
I am a bit stuck on an investigation I am doing.
When launching yolo (whether it is using the software stack we have on our HPC without a venv, using a conda environment, or using the ultralytics container), the different steps prior to the startup of the training are quite long.
In particular, I am worried about the caching of the images that is quite long.
I have around 1700 images, and it takes more than one minute to move data from the storage to the RAM of the GPU I am using. I also tried to first move the data from the storage to the local memory /dev/shm of the node on which I run experiments but it does not help.
I have around 25 to 30 it/s when it comes to moving data. Checking the io with htop also shows me that the writing processes are quite slow.
Has anyone got experience in profiling this startup phase? I tried using the scalene profiler but it did not help.
Also, has anyone found some tricks on the hyperparameters settings to speed up this data transfer without compromising the subsequent trainings? I tried different number of workers but it did not help.
Links to related issues would also be greatly appreciated.
Beta Was this translation helpful? Give feedback.
All reactions