From 3fff8df760b93605655a3d59894e3db4ab58be84 Mon Sep 17 00:00:00 2001 From: "Zewen (Evan) Li" Date: Wed, 1 May 2024 15:26:46 -0700 Subject: [PATCH] remove more cudnn dependencies --- cmake/Modules/FindcuDNN.cmake | 243 ------------------ cmake/dependencies.cmake | 2 - core/plugins/CMakeLists.txt | 1 - examples/int8/ptq/README.md | 5 +- examples/int8/qat/README.md | 5 +- examples/torchtrt_runtime_example/Makefile | 2 +- examples/torchtrt_runtime_example/README.md | 9 +- notebooks/Resnet50-CPP.ipynb | 13 - py/ci/build_whl.sh | 2 +- py/ci/soname_excludes.params | 1 - .../ci_workspaces/WORKSPACE.win.release.tmpl | 6 - 11 files changed, 10 insertions(+), 279 deletions(-) delete mode 100644 cmake/Modules/FindcuDNN.cmake diff --git a/cmake/Modules/FindcuDNN.cmake b/cmake/Modules/FindcuDNN.cmake deleted file mode 100644 index 593a9fcacf..0000000000 --- a/cmake/Modules/FindcuDNN.cmake +++ /dev/null @@ -1,243 +0,0 @@ -# Source: -# https://github.com/arrayfire/arrayfire/blob/master/CMakeModules/FindcuDNN.cmake -# -# Fetched the original content of this file from -# https://github.com/soumith/cudnn.torch -# -# Original Copyright: -# Distributed under the OSI-approved BSD 3-Clause License. See accompanying -# file Copyright.txt or https://cmake.org/licensing for details. -# -# Copyright (c) 2021, ArrayFire -# All rights reserved. -# -# This file is distributed under 3-clause BSD license. -# The complete license agreement can be obtained at: -# http://arrayfire.com/licenses/BSD-3-Clause -# -# FindcuDNN -# ------- -# -# Find cuDNN library -# -# This module creates imported target cuDNN::cuDNN upon successfull -# lookup of cuDNN headers and libraries. -# -# Valiables that affect result: -# , , : as usual -# -# Usage -# ----- -# add_exectuable(helloworld main.cpp) -# target_link_libraries(helloworld PRIVATE cuDNN::cuDNN) -# -# Note: It is recommended to avoid using variables set by the find module. -# -# Result variables -# ---------------- -# -# This module will set the following variables in your project: -# -# ``cuDNN_INCLUDE_DIRS`` -# where to find cudnn.h. -# -# ``cuDNN_LINK_LIBRARY`` -# the libraries to link against to use cuDNN. Priot to cuDNN 8, this is a huge monolithic -# library. However, since cuDNN 8 it has been split into multiple shared libraries. If -# cuDNN version 8 if found, this variable contains the shared library that dlopens the -# other libraries: cuDNN_*_INFER_LINK_LIBRARY and cuDNN_*_TRAIN_LINK_LIBRARY as needed. -# For versions of cuDNN 7 or lower, cuDNN_*_INFER_LINK_LIBRARY and cuDNN_*_TRAIN_LINK_LIBRARY -# are not defined. -# -# ``cuDNN_ADV_INFER_LINK_LIBRARY`` -# the libraries to link directly to use advanced inference API from cuDNN. -# ``cuDNN_ADV_INFER_DLL_LIBRARY`` -# Corresponding advanced inference API Windows DLL. This is not set on non-Windows platforms. -# ``cuDNN_ADV_TRAIN_LINK_LIBRARY`` -# the libraries to link directly to use advanced training API from cuDNN. -# ``cuDNN_ADV_TRAIN_DLL_LIBRARY`` -# Corresponding advanced training API Windows DLL. This is not set on non-Windows platforms. -# -# ``cuDNN_CNN_INFER_LINK_LIBRARY`` -# the libraries to link directly to use convolutional nueral networks inference API from cuDNN. -# ``cuDNN_CNN_INFER_DLL_LIBRARY`` -# Corresponding CNN inference API Windows DLL. This is not set on non-Windows platforms. -# ``cuDNN_CNN_TRAIN_LINK_LIBRARY`` -# the libraries to link directly to use convolutional nueral networks training API from cuDNN. -# ``cuDNN_CNN_TRAIN_DLL_LIBRARY`` -# Corresponding CNN training API Windows DLL. This is not set on non-Windows platforms. -# -# ``cuDNN_OPS_INFER_LINK_LIBRARY`` -# the libraries to link directly to use starndard ML operations API from cuDNN. -# ``cuDNN_OPS_INFER_DLL_LIBRARY`` -# Corresponding OPS inference API Windows DLL. This is not set on non-Windows platforms. -# ``cuDNN_OPS_TRAIN_LINK_LIBRARY`` -# the libraries to link directly to use starndard ML operations API from cuDNN. -# ``cuDNN_OPS_TRAIN_DLL_LIBRARY`` -# Corresponding OPS inference API Windows DLL. This is not set on non-Windows platforms. -# -# ``cuDNN_FOUND`` -# If false, do not try to use cuDNN. -# ``cuDNN_VERSION`` -# Version of the cuDNN library found -# ``cuDNN_VERSION_MAJOR`` -# Major Version of the cuDNN library found -# ``cuDNN_VERSION_MINOR`` -# Minor Version of the cuDNN library found - -find_package(PkgConfig) -pkg_check_modules(PC_CUDNN QUIET cuDNN) - -find_package(CUDAToolkit QUIET) - -find_path(cuDNN_INCLUDE_DIRS - NAMES cudnn.h - HINTS - ${cuDNN_ROOT_DIR} - ${PC_CUDNN_INCLUDE_DIRS} - ${CUDA_TOOLKIT_INCLUDE} - PATH_SUFFIXES include - DOC "cuDNN include directory path." ) - -if(cuDNN_INCLUDE_DIRS) - file(READ ${cuDNN_INCLUDE_DIRS}/cudnn.h CUDNN_VERSION_FILE_CONTENTS) - string(REGEX MATCH "define CUDNN_MAJOR * +([0-9]+)" - CUDNN_MAJOR_VERSION "${CUDNN_VERSION_FILE_CONTENTS}") - list(LENGTH CUDNN_MAJOR_VERSION cudnn_ver_matches) - if(${cudnn_ver_matches} EQUAL 0) - file(READ ${cuDNN_INCLUDE_DIRS}/cudnn_version.h CUDNN_VERSION_FILE_CONTENTS) - string(REGEX MATCH "define CUDNN_MAJOR * +([0-9]+)" - CUDNN_MAJOR_VERSION "${CUDNN_VERSION_FILE_CONTENTS}") - endif() - string(REGEX REPLACE "define CUDNN_MAJOR * +([0-9]+)" "\\1" - CUDNN_MAJOR_VERSION "${CUDNN_MAJOR_VERSION}") - string(REGEX MATCH "define CUDNN_MINOR * +([0-9]+)" - CUDNN_MINOR_VERSION "${CUDNN_VERSION_FILE_CONTENTS}") - string(REGEX REPLACE "define CUDNN_MINOR * +([0-9]+)" "\\1" - CUDNN_MINOR_VERSION "${CUDNN_MINOR_VERSION}") - string(REGEX MATCH "define CUDNN_PATCHLEVEL * +([0-9]+)" - CUDNN_PATCH_VERSION "${CUDNN_VERSION_FILE_CONTENTS}") - string(REGEX REPLACE "define CUDNN_PATCHLEVEL * +([0-9]+)" "\\1" - CUDNN_PATCH_VERSION "${CUDNN_PATCH_VERSION}") - set(cuDNN_VERSION_MAJOR ${CUDNN_MAJOR_VERSION}) - set(cuDNN_VERSION_MINOR ${CUDNN_MINOR_VERSION}) - set(cuDNN_VERSION ${CUDNN_MAJOR_VERSION}.${CUDNN_MINOR_VERSION}) -endif() - -# Choose lib suffix to be exact major version if requested -# otherwise, just pick the one read from cudnn.h header -if(cuDNN_FIND_VERSION_EXACT) - set(cudnn_ver_suffix "${cuDNN_FIND_VERSION_MAJOR}") -else() - set(cudnn_ver_suffix "${CUDNN_MAJOR_VERSION}") -endif() - -if(cuDNN_INCLUDE_DIRS) - get_filename_component(libpath_cudart "${CUDA_CUDART_LIBRARY}" PATH) - - macro(af_find_cudnn_libs cudnn_lib_name_infix) - if("${cudnn_lib_name_infix}" STREQUAL "") - set(LIB_INFIX "") - else() - string(TOUPPER ${cudnn_lib_name_infix} LIB_INFIX) - endif() - find_library(cuDNN${LIB_INFIX}_LINK_LIBRARY - NAMES - libcudnn${cudnn_lib_name_infix}.so.${cudnn_ver_suffix} - libcudnn${cudnn_lib_name_infix}.${cudnn_ver_suffix}.dylib - cudnn${cudnn_lib_name_infix} - PATHS - ${cuDNN_ROOT_DIR} - ${PC_CUDNN_LIBRARY_DIRS} - $ENV{LD_LIBRARY_PATH} - ${libpath_cudart} - ${CMAKE_INSTALL_PREFIX} - PATH_SUFFIXES lib lib64 bin lib/x64 bin/x64 - DOC "cudnn${cudnn_lib_name_infix} link library." ) - mark_as_advanced(cuDNN${LIB_INFIX}_LINK_LIBRARY) - - if(WIN32 AND cuDNN_LINK_LIBRARY) - find_file(cuDNN${LIB_INFIX}_DLL_LIBRARY - NAMES cudnn${cudnn_lib_name_infix}64_${cudnn_ver_suffix}${CMAKE_SHARED_LIBRARY_SUFFIX} - PATHS - ${cuDNN_ROOT_DIR} - ${PC_CUDNN_LIBRARY_DIRS} - $ENV{PATH} - ${libpath_cudart} - ${CMAKE_INSTALL_PREFIX} - PATH_SUFFIXES lib lib64 bin lib/x64 bin/x64 - DOC "cudnn${cudnn_lib_name_infix} Windows DLL." ) - mark_as_advanced(cuDNN${LIB_INFIX}_DLL_LIBRARY) - endif() - endmacro() - - af_find_cudnn_libs("") # gets base cudnn shared library - if(cuDNN_VERSION_MAJOR VERSION_GREATER 8 OR cuDNN_VERSION_MAJOR VERSION_EQUAL 8) - af_find_cudnn_libs("_adv_infer") - af_find_cudnn_libs("_adv_train") - af_find_cudnn_libs("_cnn_infer") - af_find_cudnn_libs("_cnn_train") - af_find_cudnn_libs("_ops_infer") - af_find_cudnn_libs("_ops_train") - endif() -endif() - -# pytorch compatibility layer -set(CUDNN_LIBRARY_PATH ${cuDNN_LINK_LIBRARY}) -set(CUDNN_INCLUDE_PATH ${cuDNN_INCLUDE_DIRS}) - -find_package_handle_standard_args(cuDNN - REQUIRED_VARS cuDNN_LINK_LIBRARY cuDNN_INCLUDE_DIRS - VERSION_VAR cuDNN_VERSION) - -mark_as_advanced(cuDNN_LINK_LIBRARY cuDNN_INCLUDE_DIRS cuDNN_DLL_LIBRARY) - -if(cuDNN_FOUND) - if(NOT TARGET cuDNN::cuDNN) - add_library(cuDNN::cuDNN SHARED IMPORTED) - if(WIN32) - set_target_properties(cuDNN::cuDNN - PROPERTIES - IMPORTED_LINK_INTERFACE_LANGUAGE "C" - INTERFACE_INCLUDE_DIRECTORIES "${cuDNN_INCLUDE_DIRS}" - IMPORTED_LOCATION "${cuDNN_DLL_LIBRARY}" - IMPORTED_IMPLIB "${cuDNN_LINK_LIBRARY}" - ) - else(WIN32) - set_target_properties(cuDNN::cuDNN - PROPERTIES - IMPORTED_LINK_INTERFACE_LANGUAGE "C" - INTERFACE_INCLUDE_DIRECTORIES "${cuDNN_INCLUDE_DIRS}" - IMPORTED_LOCATION "${cuDNN_LINK_LIBRARY}" - ) - endif(WIN32) - if(cuDNN_VERSION_MAJOR VERSION_GREATER 8 OR cuDNN_VERSION_MAJOR VERSION_EQUAL 8) - macro(create_cudnn_target cudnn_target_name) - string(TOUPPER ${cudnn_target_name} target_infix) - add_library(cuDNN::${cudnn_target_name} SHARED IMPORTED) - if(WIN32) - set_target_properties(cuDNN::${cudnn_target_name} - PROPERTIES - IMPORTED_LINK_INTERFACE_LANGUAGE "C" - INTERFACE_INCLUDE_DIRECTORIES "${cuDNN_INCLUDE_DIRS}" - IMPORTED_LOCATION "${cuDNN_${target_infix}_DLL_LIBRARY}" - IMPORTED_IMPLIB "${cuDNN_${target_infix}_LINK_LIBRARY}" - ) - else(WIN32) - set_target_properties(cuDNN::${cudnn_target_name} - PROPERTIES - IMPORTED_LINK_INTERFACE_LANGUAGE "C" - INTERFACE_INCLUDE_DIRECTORIES "${cuDNN_INCLUDE_DIRS}" - IMPORTED_LOCATION "${cuDNN_${target_infix}_LINK_LIBRARY}" - ) - endif(WIN32) - endmacro() - create_cudnn_target(adv_infer) - create_cudnn_target(adv_train) - create_cudnn_target(cnn_infer) - create_cudnn_target(cnn_train) - create_cudnn_target(ops_infer) - create_cudnn_target(ops_train) - endif() - endif(NOT TARGET cuDNN::cuDNN) -endif(cuDNN_FOUND) diff --git a/cmake/dependencies.cmake b/cmake/dependencies.cmake index a2d994eb1a..4689f52734 100644 --- a/cmake/dependencies.cmake +++ b/cmake/dependencies.cmake @@ -7,11 +7,9 @@ endif() # If the custom finders are needed at this point, there are good chances that they will be needed when consuming the library as well install(FILES "${CMAKE_SOURCE_DIR}/cmake/Modules/FindTensorRT.cmake" DESTINATION "${CMAKE_INSTALL_LIBDIR}/cmake/torchtrt/Modules") -install(FILES "${CMAKE_SOURCE_DIR}/cmake/Modules/FindcuDNN.cmake" DESTINATION "${CMAKE_INSTALL_LIBDIR}/cmake/torchtrt/Modules") # CUDA find_package(CUDAToolkit REQUIRED) -find_package(cuDNN REQUIRED) # Headers are needed somewhere # libtorch find_package(Torch REQUIRED) diff --git a/core/plugins/CMakeLists.txt b/core/plugins/CMakeLists.txt index 2b1eaaf73f..90b039c94a 100644 --- a/core/plugins/CMakeLists.txt +++ b/core/plugins/CMakeLists.txt @@ -23,7 +23,6 @@ target_link_libraries(${lib_name} TensorRT::nvinfer_plugin torch core_util - cuDNN::cuDNN PRIVATE Threads::Threads ) diff --git a/examples/int8/ptq/README.md b/examples/int8/ptq/README.md index 4ac147bb6e..329d4d021d 100644 --- a/examples/int8/ptq/README.md +++ b/examples/int8/ptq/README.md @@ -149,14 +149,13 @@ tar -xvzf libtorch_tensorrt.tar.gz unzip libtorch.zip ``` -> If cuDNN and TensorRT are not installed on your system / in your LD_LIBRARY_PATH then do the following as well +> If TensorRT is not installed on your system / in your LD_LIBRARY_PATH then do the following as well ```sh cd deps -mkdir cudnn && tar -xvzf --directory cudnn --strip-components=1 mkdir tensorrt && tar -xvzf --directory tensorrt --strip-components=1 cd .. -export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$(pwd)/deps/torch_tensorrt/lib:$(pwd)/deps/libtorch/lib:$(pwd)/deps/tensorrt/lib:$(pwd)/deps/cudnn/lib64:/usr/local/cuda/lib +export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$(pwd)/deps/torch_tensorrt/lib:$(pwd)/deps/libtorch/lib:$(pwd)/deps/tensorrt/lib:/usr/local/cuda/lib ``` 2) Build and run `ptq` diff --git a/examples/int8/qat/README.md b/examples/int8/qat/README.md index f47aea15b0..992121c705 100644 --- a/examples/int8/qat/README.md +++ b/examples/int8/qat/README.md @@ -43,14 +43,13 @@ tar -xvzf libtorch_tensorrt.tar.gz unzip libtorch.zip ``` -> If cuDNN and TensorRT are not installed on your system / in your LD_LIBRARY_PATH then do the following as well +> If TensorRT is not installed on your system / in your LD_LIBRARY_PATH then do the following as well ```sh cd deps -mkdir cudnn && tar -xvzf --directory cudnn --strip-components=1 mkdir tensorrt && tar -xvzf --directory tensorrt --strip-components=1 cd .. -export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$(pwd)/deps/torch_tensorrt/lib:$(pwd)/deps/libtorch/lib:$(pwd)/deps/tensorrt/lib:$(pwd)/deps/cudnn/lib64:/usr/local/cuda/lib +export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$(pwd)/deps/torch_tensorrt/lib:$(pwd)/deps/libtorch/lib:$(pwd)/deps/tensorrt/lib:/usr/local/cuda/lib ``` 2) Build and run `qat` diff --git a/examples/torchtrt_runtime_example/Makefile b/examples/torchtrt_runtime_example/Makefile index 9f8df64763..07fcc4c291 100644 --- a/examples/torchtrt_runtime_example/Makefile +++ b/examples/torchtrt_runtime_example/Makefile @@ -1,7 +1,7 @@ CXX=g++ DEP_DIR=$(PWD)/deps INCLUDE_DIRS=-I$(DEP_DIR)/libtorch/include -I$(DEP_DIR)/torch_tensorrt/include -LIB_DIRS=-L$(DEP_DIR)/torch_tensorrt/lib -L$(DEP_DIR)/libtorch/lib # -Wl,-rpath $(DEP_DIR)/tensorrt/lib -Wl,-rpath $(DEP_DIR)/cudnn/lib64 +LIB_DIRS=-L$(DEP_DIR)/torch_tensorrt/lib -L$(DEP_DIR)/libtorch/lib # -Wl,-rpath $(DEP_DIR)/tensorrt/lib LIBS=-Wl,--no-as-needed -ltorchtrt_runtime -Wl,--as-needed -ltorch -ltorch_cuda -ltorch_cpu -ltorch_global_deps -lbackend_with_compiler -lc10 -lc10_cuda SRCS=main.cpp diff --git a/examples/torchtrt_runtime_example/README.md b/examples/torchtrt_runtime_example/README.md index 9effed5046..15b740b0b0 100644 --- a/examples/torchtrt_runtime_example/README.md +++ b/examples/torchtrt_runtime_example/README.md @@ -26,17 +26,16 @@ tar -xvzf libtorch_tensorrt.tar.gz unzip libtorch-cxx11-abi-shared-with-deps-[PYTORCH_VERSION].zip ``` -> If cuDNN and TensorRT are not installed on your system / in your LD_LIBRARY_PATH then do the following as well +> If TensorRT is not installed on your system / in your LD_LIBRARY_PATH then do the following as well ```sh cd deps -mkdir cudnn && tar -xvzf --directory cudnn --strip-components=1 mkdir tensorrt && tar -xvzf --directory tensorrt --strip-components=1 cd .. -export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$(pwd)/deps/torch_tensorrt/lib:$(pwd)/deps/libtorch/lib:$(pwd)/deps/tensorrt/lib:$(pwd)/deps/cudnn/lib64:/usr/local/cuda/lib +export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$(pwd)/deps/torch_tensorrt/lib:$(pwd)/deps/libtorch/lib:$(pwd)/deps/tensorrt/lib:/usr/local/cuda/lib ``` -This gives maximum compatibility with system configurations for running this example but in general you are better off adding `-Wl,-rpath $(DEP_DIR)/tensorrt/lib -Wl,-rpath $(DEP_DIR)/cudnn/lib64` to your linking command for actual applications +This gives maximum compatibility with system configurations for running this example but in general you are better off adding `-Wl,-rpath $(DEP_DIR)/tensorrt/lib` to your linking command for actual applications 2) Build and run `torchtrt_runtime_example` @@ -48,6 +47,6 @@ To build and run the app cd examples/torchtrt_runtime_example make # If paths are different than the ones below, change as necessary -export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$(pwd)/deps/torch_tensorrt/lib:$(pwd)/deps/libtorch/lib:$(pwd)/deps/tensorrt/lib:$(pwd)/deps/cudnn/lib64:/usr/local/cuda/lib +export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$(pwd)/deps/torch_tensorrt/lib:$(pwd)/deps/libtorch/lib:$(pwd)/deps/tensorrt/lib:/usr/local/cuda/lib ./torchtrt_runtime_example $PWD/examples/torchtrt_runtime_example/norm.jit ``` diff --git a/notebooks/Resnet50-CPP.ipynb b/notebooks/Resnet50-CPP.ipynb index 3d5e66ee51..198ebc9911 100755 --- a/notebooks/Resnet50-CPP.ipynb +++ b/notebooks/Resnet50-CPP.ipynb @@ -172,19 +172,6 @@ "unzip libtorch-cxx11-abi-shared-with-deps-1.11.0+cu113.zip" ] }, - { - "cell_type": "code", - "execution_count": null, - "id": "198553b7", - "metadata": {}, - "outputs": [], - "source": [ - "%%bash\n", - "cd deps\n", - "wget https://github.com/pytorch/TensorRT/releases/download/v1.1.0/libtorchtrt-v1.1.0-cudnn8.2-tensorrt8.2-cuda11.3-libtorch1.11.0.tar.gz\n", - "tar -xvzf libtorchtrt-v1.1.0-cudnn8.2-tensorrt8.2-cuda11.3-libtorch1.11.0.tar.gz" - ] - }, { "cell_type": "markdown", "id": "309696b6", diff --git a/py/ci/build_whl.sh b/py/ci/build_whl.sh index 7dabaecda3..35090445f0 100755 --- a/py/ci/build_whl.sh +++ b/py/ci/build_whl.sh @@ -101,7 +101,7 @@ libtorchtrt() { TORCHTRT_VERSION=$(cd ${PROJECT_DIR} && ${PY_DIR}/bin/python3 -c "import versions; versions.torch_tensorrt_version_release()") TRT_VERSION=$(cd ${PROJECT_DIR} && ${PY_DIR}/bin/python3 -c "import versions; versions.tensorrt_version()") TORCH_VERSION=$(${PY_DIR}/bin/python -c "from torch import __version__;print(__version__.split('+')[0])") - cp ${PROJECT_DIR}/bazel-bin/libtorchtrt.tar.gz ${PROJECT_DIR}/py/wheelhouse/libtorchtrt-${TORCHTRT_VERSION}-cudnn${CUDNN_VERSION}-tensorrt${TRT_VERSION}-cuda${CUDA_VERSION}-libtorch${TORCH_VERSION}-x86_64-linux.tar.gz + cp ${PROJECT_DIR}/bazel-bin/libtorchtrt.tar.gz ${PROJECT_DIR}/py/wheelhouse/libtorchtrt-${TORCHTRT_VERSION}-tensorrt${TRT_VERSION}-cuda${CUDA_VERSION}-libtorch${TORCH_VERSION}-x86_64-linux.tar.gz } libtorchtrt_pre_cxx11_abi() { diff --git a/py/ci/soname_excludes.params b/py/ci/soname_excludes.params index a5eecb7c9a..53f7f48a65 100644 --- a/py/ci/soname_excludes.params +++ b/py/ci/soname_excludes.params @@ -30,7 +30,6 @@ --exclude libcuda.so.515 --exclude libcublasLt.so.11 --exclude libnvinfer.so.8 ---exclude libcudnn.so.8 --exclude libcublas.so.12 --exclude libcublasLt.so.12 --exclude libcublas.so.12.1.3.1 diff --git a/toolchains/ci_workspaces/WORKSPACE.win.release.tmpl b/toolchains/ci_workspaces/WORKSPACE.win.release.tmpl index acc4ecd51c..01bbd2130e 100644 --- a/toolchains/ci_workspaces/WORKSPACE.win.release.tmpl +++ b/toolchains/ci_workspaces/WORKSPACE.win.release.tmpl @@ -73,12 +73,6 @@ http_archive( # Locally installed dependencies (use in cases of custom dependencies or aarch64) #################################################################################### -new_local_repository( - name = "cudnn", - path = "C:/", - build_file = "@//third_party/cudnn/local:BUILD" -) - new_local_repository( name = "tensorrt", path = "C:/",