From c5ddc78e016a21e4dbe793aeab5073f6449c583a Mon Sep 17 00:00:00 2001 From: Cody Hao Yu Date: Sat, 14 Sep 2019 00:42:51 -0700 Subject: [PATCH] eliminate unused variables --- topi/python/topi/cuda/dense.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/topi/python/topi/cuda/dense.py b/topi/python/topi/cuda/dense.py index 5136ee694783d..d25f43bd26f76 100644 --- a/topi/python/topi/cuda/dense.py +++ b/topi/python/topi/cuda/dense.py @@ -234,7 +234,7 @@ def schedule_dense_large_batch(cfg, s, C, outs): # Schedule for A's shared memory load num_thread_x = cfg['tile_x'].size[2] - ty, xi = s[AA].split(s[AA].op.axis[0], nparts=num_thread_x) + ty, _ = s[AA].split(s[AA].op.axis[0], nparts=num_thread_x) _, xi = s[AA].split(s[AA].op.axis[1], factor=num_thread_x * 4) tx, xi = s[AA].split(xi, nparts=num_thread_x) s[AA].bind(ty, tvm.thread_axis("threadIdx.y")) @@ -243,7 +243,7 @@ def schedule_dense_large_batch(cfg, s, C, outs): # Schedule for B' shared memory load num_thread_y = cfg['tile_y'].size[2] - ty, xi = s[BB].split(s[BB].op.axis[0], nparts=num_thread_y) + ty, _ = s[BB].split(s[BB].op.axis[0], nparts=num_thread_y) _, xi = s[BB].split(s[BB].op.axis[1], factor=num_thread_y * 4) tx, xi = s[BB].split(xi, nparts=num_thread_y) s[BB].bind(ty, tvm.thread_axis("threadIdx.y"))