From d9d4066a482159d280f672867989af6aec10d658 Mon Sep 17 00:00:00 2001 From: Edward Yang Date: Tue, 12 Nov 2024 14:56:37 -0800 Subject: [PATCH] Add pyre fixme for downstream type errors from D65753120 Summary: X-link: https://github.com/pytorch/captum/pull/1439 X-link: https://github.com/facebook/Ax/pull/3055 X-link: https://github.com/ctrl-labs/src2/pull/38515 X-link: https://github.com/ctrl-labs/src2/pull/38514 bypass-github-export-checks "The check is bugged, I exported all the required exports" Reviewed By: jermenkoo Differential Revision: D65826205 fbshipit-source-id: b08f00caaac8f3cc235bc280c4fe3f99089a0753 --- fbgemm_gpu/experimental/gemm/test/fp8_gemm_test.py | 2 ++ 1 file changed, 2 insertions(+) diff --git a/fbgemm_gpu/experimental/gemm/test/fp8_gemm_test.py b/fbgemm_gpu/experimental/gemm/test/fp8_gemm_test.py index 1c78f065d..89f510e00 100644 --- a/fbgemm_gpu/experimental/gemm/test/fp8_gemm_test.py +++ b/fbgemm_gpu/experimental/gemm/test/fp8_gemm_test.py @@ -156,6 +156,8 @@ def _quantize_matmul_fp8( expected_result = a @ b.T if use_bias: + # pyre-fixme[6]: For 1st argument expected `Union[bool, complex, + # float, int, Tensor]` but got `Optional[Tensor]`. expected_result += bias self.assertTrue( torch.allclose(result, expected_result, atol=2e-1, rtol=5e-2)