Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[rocm6.1] Update Triton dependency logic #47

Merged
merged 2 commits into from
Mar 22, 2024

Conversation

jithunnair-amd
Copy link

@jithunnair-amd jithunnair-amd commented Mar 21, 2024

@jithunnair-amd
Copy link
Author

Test results of PyTorch2.0 and triton wheels generated from http://rocm-ci.amd.com/job/rocm-pytorch-manylinux-wheel-builder/154, installed in rocm/pytorch-private:manylinux-hipclang-13484:

[root@ed0329edef16 apex]# pip3 install -f https://compute-artifactory.amd.com/artifactory/compute-pytorch-rocm/compute-rocm-dkms-no-npi-hipclang/13484/ torch==2.0.1 torchvision==0.15.2
Looking in links: https://compute-artifactory.amd.com/artifactory/compute-pytorch-rocm/compute-rocm-dkms-no-npi-hipclang/13484/
Collecting torch==2.0.1
  Downloading https://compute-artifactory.amd.com/artifactory/compute-pytorch-rocm/compute-rocm-dkms-no-npi-hipclang/13484/torch-2.0.1%2Brocm6.1-cp39-cp39-linux_x86_64.whl (2095.9 MB)
...
Collecting pytorch-triton-rocm==2.0.2+f84c1f1e62 (from torch==2.0.1)
  Downloading https://compute-artifactory.amd.com/artifactory/compute-pytorch-rocm/compute-rocm-dkms-no-npi-hipclang/13484/pytorch_triton_rocm-2.0.2%2Bf84c1f1e62-cp39-cp39-linux_x86_64.whl (206.3 MB)
...

[root@ed0329edef16 pytorch-micro-benchmarking]# python3 micro_benchmarking_pytorch.py --network resnet50 --compile
INFO: running forward and backward for warmup.
/opt/python/cp39-cp39/lib/python3.9/site-packages/torch/_inductor/compile_fx.py:90: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance.
  warnings.warn(
INFO: running the benchmark..
OK: finished running benchmark..
--------------------SUMMARY--------------------------
Microbenchmark for network : resnet50
Num devices: 1
Dtype: FP32
Mini batch size [img] : 64
Time per mini-batch : 0.11212306022644043
Throughput [img/sec] : 570.8014022338267

@jithunnair-amd jithunnair-amd force-pushed the rocm6.1_with_triton_hash branch from be72ec9 to 608d4f4 Compare March 22, 2024 21:24
@jithunnair-amd jithunnair-amd marked this pull request as ready for review March 22, 2024 21:24
@jithunnair-amd jithunnair-amd changed the title Update Triton dependency logic [rocm6.1] Update Triton dependency logic Mar 22, 2024
@jithunnair-amd jithunnair-amd merged commit 4f4f6bd into rocm6.1 Mar 22, 2024
2 checks passed
@jithunnair-amd jithunnair-amd deleted the rocm6.1_with_triton_hash branch March 22, 2024 21:30
jithunnair-amd added a commit that referenced this pull request Apr 11, 2024
* Use hashed triton version

* skip triton dependency for PyTorch < 2.0

(cherry picked from commit 4f4f6bd)
jithunnair-amd added a commit that referenced this pull request Jul 1, 2024
* Use hashed triton version

* skip triton dependency for PyTorch < 2.0

(cherry picked from commit 4f4f6bd)

Add ROCM_VERSION to triton dependency (#48)

(cherry picked from commit c74ab20)

Use ROCm version with patch for triton dependency (#49)
jithunnair-amd added a commit that referenced this pull request Jul 4, 2024
* Use hashed triton version

* skip triton dependency for PyTorch < 2.0

(cherry picked from commit 4f4f6bd)

Add ROCM_VERSION to triton dependency (#48)

(cherry picked from commit c74ab20)

Use ROCm version with patch for triton dependency (#49)
jithunnair-amd added a commit that referenced this pull request Jul 12, 2024
* Use hashed triton version

* skip triton dependency for PyTorch < 2.0

(cherry picked from commit 4f4f6bd)

Add ROCM_VERSION to triton dependency (#48)

(cherry picked from commit c74ab20)

Use ROCm version with patch for triton dependency (#49)
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant