Slow training of pruned transducer is normal? #391
-
Hi, I'm training a model with my own data by combining tedlium's The GPU is a RTX3090, I'm not bottlenecked by IO (minimal time spent between end of and start of batch iteration loop), my env is:
So my question is should I expect this training speed? Is this inherent with using RNNT? 50 batches (max-duration 500) take more than 2 minutes (closer to 3). |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 23 replies
-
With V100 32GB, when the max duration is 300, it takes about 33 seconds to 55 seconds for 50 batches for the LibriSpeech corpus. You may use https://github.com/benfred/py-spy to find out which one is the time consuming part. |
Beta Was this translation helpful? Give feedback.
With V100 32GB, when the max duration is 300, it takes about 33 seconds to 55 seconds for 50 batches for the LibriSpeech corpus.
You may use https://github.com/benfred/py-spy to find out which one is the time consuming part.