You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
On imagenet_resnet workload, I encounter the following warning when running with DDP and pytorch framework.
/u/najroldi/miniconda3/envs/alpe/lib/python3.8/site-packages/torch/autograd/init.py:251: UserWarning: Grad strides do not match bucket view strides. This may indicate grad was not created according to the gradient layout contract, or that the param's strides changed since DDP was constructed. This is not an error, but may impair performance.
grad.sizes() = [512, 2048, 1, 1], strides() = [2048, 1, 2048, 2048]
Is this a known problem? Is there a known fix?
Related to this, I have noticed that the prize_qualification_baseline logs start with python3 submission_runner.py. For self reporting results, and for the final scoring, should we use DDP?
@msaroufim have you encountered this warning before?
Full message in this log.
So far it seems like it happens only on ResNet. I just noticed this warning has been in our test logs for >3 months, I think all of which we were running on PyTorch 2.1. Trying to find the first instance this warning occurred..
It looks like this warning only gets printed in the first step. I'm not sure how it affects the speed of the rest of the training. I had last compared the PyTorch and JAX run times in Dec 2023. The ResNet runs in PyTorch also show this warning in the logs for the timing runs. ResNet is about 6% slower in PyTorch vs JAX which I think is within tolerance.
Description
On
imagenet_resnet
workload, I encounter the following warning when running with DDP andpytorch
framework.Is this a known problem? Is there a known fix?
Related to this, I have noticed that the prize_qualification_baseline logs start with
python3 submission_runner.py
. For self reporting results, and for the final scoring, should we use DDP?Steps to Reproduce
The text was updated successfully, but these errors were encountered: