From 347528a687d52cc32355da238271e60091db2b21 Mon Sep 17 00:00:00 2001 From: irexyc Date: Mon, 15 Sep 2025 03:23:43 +0000 Subject: [PATCH] remove cudnn --- builder/manywheel/scripts/install_cuda.sh | 48 ++------------------- cmake/Modules/FindCUDNN.cmake | 51 ----------------------- src/turbomind/utils/cuda_utils.h | 10 ----- 3 files changed, 4 insertions(+), 105 deletions(-) delete mode 100644 cmake/Modules/FindCUDNN.cmake diff --git a/builder/manywheel/scripts/install_cuda.sh b/builder/manywheel/scripts/install_cuda.sh index 8434fe3dba..5cacfe12c2 100755 --- a/builder/manywheel/scripts/install_cuda.sh +++ b/builder/manywheel/scripts/install_cuda.sh @@ -3,7 +3,7 @@ set -ex function install_118 { - echo "Installing CUDA 11.8 and cuDNN 8.7 and NCCL 2.15" + echo "Installing CUDA 11.8 and NCCL 2.15" rm -rf /usr/local/cuda-11.8 /usr/local/cuda # install CUDA 11.8.0 in the same container wget -q https://developer.download.nvidia.com/compute/cuda/11.8.0/local_installers/cuda_11.8.0_520.61.05_linux.run @@ -12,16 +12,6 @@ function install_118 { rm -f cuda_11.8.0_520.61.05_linux.run rm -f /usr/local/cuda && ln -s /usr/local/cuda-11.8 /usr/local/cuda - # cuDNN license: https://developer.nvidia.com/cudnn/license_agreement - mkdir tmp_cudnn && cd tmp_cudnn - wget -q https://developer.download.nvidia.com/compute/redist/cudnn/v8.7.0/local_installers/11.8/cudnn-linux-x86_64-8.7.0.84_cuda11-archive.tar.xz -O cudnn-linux-x86_64-8.7.0.84_cuda11-archive.tar.xz - tar xf cudnn-linux-x86_64-8.7.0.84_cuda11-archive.tar.xz - cp -a cudnn-linux-x86_64-8.7.0.84_cuda11-archive/include/* /usr/local/cuda/include/ - cp -a cudnn-linux-x86_64-8.7.0.84_cuda11-archive/lib/* /usr/local/cuda/lib64/ - cd .. - rm -rf tmp_cudnn - ldconfig - # NCCL license: https://docs.nvidia.com/deeplearning/nccl/#licenses mkdir tmp_nccl && cd tmp_nccl wget -q https://developer.download.nvidia.com/compute/redist/nccl/v2.15.5/nccl_2.15.5-1+cuda11.8_x86_64.txz @@ -34,7 +24,7 @@ function install_118 { } function install_121 { - echo "Installing CUDA 12.1 and cuDNN 8.9 and NCCL 2.18.1" + echo "Installing CUDA 12.1 and NCCL 2.18.1" rm -rf /usr/local/cuda-12.1 /usr/local/cuda # install CUDA 12.1.0 in the same container wget -q https://developer.download.nvidia.com/compute/cuda/12.1.0/local_installers/cuda_12.1.0_530.30.02_linux.run @@ -43,16 +33,6 @@ function install_121 { rm -f cuda_12.1.0_530.30.02_linux.run rm -f /usr/local/cuda && ln -s /usr/local/cuda-12.1 /usr/local/cuda - # cuDNN license: https://developer.nvidia.com/cudnn/license_agreement - mkdir tmp_cudnn && cd tmp_cudnn - wget -q https://developer.download.nvidia.com/compute/cudnn/redist/cudnn/linux-x86_64/cudnn-linux-x86_64-8.9.2.26_cuda12-archive.tar.xz -O cudnn-linux-x86_64-8.9.2.26_cuda12-archive.tar.xz - tar xf cudnn-linux-x86_64-8.9.2.26_cuda12-archive.tar.xz - cp -a cudnn-linux-x86_64-8.9.2.26_cuda12-archive/include/* /usr/local/cuda/include/ - cp -a cudnn-linux-x86_64-8.9.2.26_cuda12-archive/lib/* /usr/local/cuda/lib64/ - cd .. - rm -rf tmp_cudnn - ldconfig - # NCCL license: https://docs.nvidia.com/deeplearning/nccl/#licenses mkdir tmp_nccl && cd tmp_nccl wget -q https://developer.download.nvidia.com/compute/redist/nccl/v2.18.1/nccl_2.18.1-1+cuda12.1_x86_64.txz @@ -65,7 +45,7 @@ function install_121 { } function install_124 { - echo "Installing CUDA 12.4 and cuDNN 8.9 and NCCL 2.25.1" + echo "Installing CUDA 12.4 and NCCL 2.25.1" rm -rf /usr/local/cuda-12.4 /usr/local/cuda # install CUDA 12.4.1 in the same container wget -q https://developer.download.nvidia.com/compute/cuda/12.4.1/local_installers/cuda_12.4.1_550.54.15_linux.run @@ -74,16 +54,6 @@ function install_124 { rm -f cuda_12.4.1_550.54.15_linux.run rm -f /usr/local/cuda && ln -s /usr/local/cuda-12.4 /usr/local/cuda - # cuDNN license: https://developer.nvidia.com/cudnn/license_agreement - mkdir tmp_cudnn && cd tmp_cudnn - wget -q https://developer.download.nvidia.com/compute/cudnn/redist/cudnn/linux-x86_64/cudnn-linux-x86_64-8.9.2.26_cuda12-archive.tar.xz -O cudnn-linux-x86_64-8.9.2.26_cuda12-archive.tar.xz - tar xf cudnn-linux-x86_64-8.9.2.26_cuda12-archive.tar.xz - cp -a cudnn-linux-x86_64-8.9.2.26_cuda12-archive/include/* /usr/local/cuda/include/ - cp -a cudnn-linux-x86_64-8.9.2.26_cuda12-archive/lib/* /usr/local/cuda/lib64/ - cd .. - rm -rf tmp_cudnn - ldconfig - # NCCL license: https://docs.nvidia.com/deeplearning/nccl/#licenses mkdir tmp_nccl && cd tmp_nccl wget -q https://developer.download.nvidia.com/compute/redist/nccl/v2.25.1/nccl_2.25.1-1+cuda12.4_x86_64.txz @@ -96,7 +66,7 @@ function install_124 { } function install_128 { - echo "Installing CUDA 12.8 and cuDNN 8.9 and NCCL 2.25.1" + echo "Installing CUDA 12.8 and NCCL 2.25.1" rm -rf /usr/local/cuda-12.8 /usr/local/cuda # install CUDA 12.8.1 in the same container wget -q https://developer.download.nvidia.com/compute/cuda/12.8.1/local_installers/cuda_12.8.1_570.124.06_linux.run @@ -105,16 +75,6 @@ function install_128 { rm -f cuda_12.8.1_570.124.06_linux.run rm -f /usr/local/cuda && ln -s /usr/local/cuda-12.8 /usr/local/cuda - # cuDNN license: https://developer.nvidia.com/cudnn/license_agreement - mkdir tmp_cudnn && cd tmp_cudnn - wget -q https://developer.download.nvidia.com/compute/cudnn/redist/cudnn/linux-x86_64/cudnn-linux-x86_64-8.9.2.26_cuda12-archive.tar.xz -O cudnn-linux-x86_64-8.9.2.26_cuda12-archive.tar.xz - tar xf cudnn-linux-x86_64-8.9.2.26_cuda12-archive.tar.xz - cp -a cudnn-linux-x86_64-8.9.2.26_cuda12-archive/include/* /usr/local/cuda/include/ - cp -a cudnn-linux-x86_64-8.9.2.26_cuda12-archive/lib/* /usr/local/cuda/lib64/ - cd .. - rm -rf tmp_cudnn - ldconfig - # NCCL license: https://docs.nvidia.com/deeplearning/nccl/#licenses mkdir tmp_nccl && cd tmp_nccl wget -q https://developer.download.nvidia.com/compute/redist/nccl/v2.25.1/nccl_2.25.1-1+cuda12.8_x86_64.txz diff --git a/cmake/Modules/FindCUDNN.cmake b/cmake/Modules/FindCUDNN.cmake deleted file mode 100644 index 7e7fc0c939..0000000000 --- a/cmake/Modules/FindCUDNN.cmake +++ /dev/null @@ -1,51 +0,0 @@ -# taken from https://github.com/pytorch/pytorch/blob/master/cmake/Modules_CUDA_fix/FindCUDNN.cmake -# Find the CUDNN libraries -# -# The following variables are optionally searched for defaults -# CUDNN_ROOT: Base directory where CUDNN is found -# CUDNN_INCLUDE_DIR: Directory where CUDNN header is searched for -# CUDNN_LIBRARY: Directory where CUDNN library is searched for -# CUDNN_STATIC: Are we looking for a static library? (default: no) -# -# The following are set after configuration is done: -# CUDNN_FOUND -# CUDNN_INCLUDE_PATH -# CUDNN_LIBRARY_PATH -# - -include(FindPackageHandleStandardArgs) - -set(CUDNN_ROOT $ENV{CUDNN_ROOT_DIR} CACHE PATH "Folder containing NVIDIA cuDNN") -if (DEFINED $ENV{CUDNN_ROOT_DIR}) - message(WARNING "CUDNN_ROOT_DIR is deprecated. Please set CUDNN_ROOT instead.") -endif() -list(APPEND CUDNN_ROOT $ENV{CUDNN_ROOT_DIR} ${CUDA_TOOLKIT_ROOT_DIR}) - -# Compatible layer for CMake <3.12. CUDNN_ROOT will be accounted in for searching paths and libraries for CMake >=3.12. -list(APPEND CMAKE_PREFIX_PATH ${CUDNN_ROOT}) - -set(CUDNN_INCLUDE_DIR $ENV{CUDNN_INCLUDE_DIR} CACHE PATH "Folder containing NVIDIA cuDNN header files") - -find_path(CUDNN_INCLUDE_PATH cudnn.h - HINTS ${CUDNN_INCLUDE_DIR} - PATH_SUFFIXES cuda/include cuda include) - -option(CUDNN_STATIC "Look for static CUDNN" OFF) -if (CUDNN_STATIC) - set(CUDNN_LIBNAME "libcudnn_static.a") -else() - set(CUDNN_LIBNAME "cudnn") -endif() - -set(CUDNN_LIBRARY $ENV{CUDNN_LIBRARY} CACHE PATH "Path to the cudnn library file (e.g., libcudnn.so)") -if (CUDNN_LIBRARY MATCHES ".*cudnn_static.a" AND NOT CUDNN_STATIC) - message(WARNING "CUDNN_LIBRARY points to a static library (${CUDNN_LIBRARY}) but CUDNN_STATIC is OFF.") -endif() - -find_library(CUDNN_LIBRARY_PATH ${CUDNN_LIBNAME} - PATHS ${CUDNN_LIBRARY} - PATH_SUFFIXES lib lib64 cuda/lib cuda/lib64 lib/x64) - -find_package_handle_standard_args(CUDNN DEFAULT_MSG CUDNN_LIBRARY_PATH CUDNN_INCLUDE_PATH) - -mark_as_advanced(CUDNN_ROOT CUDNN_INCLUDE_DIR CUDNN_LIBRARY) diff --git a/src/turbomind/utils/cuda_utils.h b/src/turbomind/utils/cuda_utils.h index 83a89dfe0b..3ef2e337e5 100644 --- a/src/turbomind/utils/cuda_utils.h +++ b/src/turbomind/utils/cuda_utils.h @@ -107,16 +107,6 @@ void syncAndCheck(const char* const file, int const line); throw std::runtime_error(msg.c_str()); \ } -#define checkCUDNN(expression) \ - { \ - cudnnStatus_t status = (expression); \ - if (status != CUDNN_STATUS_SUCCESS) { \ - std::cerr << "Error on file " << __FILE__ << " line " << __LINE__ << ": " << cudnnGetErrorString(status) \ - << std::endl; \ - std::exit(EXIT_FAILURE); \ - } \ - } - template void printMatrix(T* ptr, int m, int k, int stride, bool is_device_ptr);