mirror of
https://github.com/saymrwulf/onnxruntime.git
synced 2026-05-14 20:48:00 +00:00
### Description * Reverting default TensorRT version to 8.5 as temporary fix * Apart from that, this PR temporarily leaves this CI as a place to validate user behavior that uses TRT 8.5 with latest ORT ### Context * This CI pool equips 2xTesla M60 GPUs, which are no longer supported by TensorRT 8.6. * Currently, other CIs are using single-T4 VM but there's no VM with 2xT4 or other suitable dualGPU in the range. * Once we decide which VM instance for this CI to migrate to, TRT8.6 can be enabled on this CI * According to [Nvidia](https://docs.nvidia.com/deeplearning/tensorrt/release-notes/index.html): * TensorRT 8.5.3 was the last release supporting NVIDIA Kepler (SM 3.x) and NVIDIA Maxwell (SM 5.x) devices. *These devices are no longer supported in TensorRT 8.6*. NVIDIA Pascal (SM 6.x) devices are deprecated in TensorRT 8.6.
37 lines
1.7 KiB
Text
37 lines
1.7 KiB
Text
# Tag: nvidia/cuda:11.8.0-cudnn8-devel-ubuntu20.04
|
|
# Label: com.nvidia.cuda.version: 11.8.0
|
|
# Label: com.nvidia.cudnn.version: 8.7.0
|
|
# Ubuntu 20.04
|
|
FROM nvidia/cuda:11.8.0-cudnn8-devel-ubuntu20.04
|
|
|
|
ARG PYTHON_VERSION=3.8
|
|
ARG DEBIAN_FRONTEND=noninteractive
|
|
|
|
ADD scripts /tmp/scripts
|
|
RUN /tmp/scripts/install_ubuntu.sh -p $PYTHON_VERSION && /tmp/scripts/install_os_deps.sh && /tmp/scripts/install_python_deps.sh -p $PYTHON_VERSION && rm -rf /tmp/scripts
|
|
|
|
# Install TensorRT
|
|
RUN v="8.5.1-1+cuda11.8" &&\
|
|
apt-key adv --fetch-keys https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2004/x86_64/7fa2af80.pub &&\
|
|
apt-get update &&\
|
|
sudo apt-get install -y libnvinfer8=${v} libnvonnxparsers8=${v} libnvparsers8=${v} libnvinfer-plugin8=${v} \
|
|
libnvinfer-dev=${v} libnvonnxparsers-dev=${v} libnvparsers-dev=${v} libnvinfer-plugin-dev=${v} \
|
|
python3-libnvinfer=${v} libnvinfer-samples=${v}
|
|
|
|
WORKDIR /root
|
|
|
|
# Allow configure to pick up GDK and CuDNN where it expects it.
|
|
# (Note: $CUDNN_VERSION is defined by NVidia's base image)
|
|
RUN _CUDNN_VERSION=$(echo $CUDNN_VERSION | cut -d. -f1-2) && \
|
|
mkdir -p /usr/local/cudnn-$_CUDNN_VERSION/cuda/include && \
|
|
ln -s /usr/include/cudnn.h /usr/local/cudnn-$_CUDNN_VERSION/cuda/include/cudnn.h && \
|
|
mkdir -p /usr/local/cudnn-$_CUDNN_VERSION/cuda/lib64 && \
|
|
ln -s /etc/alternatives/libcudnn_so /usr/local/cudnn-$_CUDNN_VERSION/cuda/lib64/libcudnn.so && \
|
|
ln -s /usr/local/cudnn{-$_CUDNN_VERSION,}
|
|
|
|
ARG BUILD_USER=onnxruntimedev
|
|
ARG BUILD_UID=1000
|
|
WORKDIR /home/$BUILD_USER
|
|
RUN adduser --gecos 'onnxruntime Build User' --disabled-password $BUILD_USER --uid $BUILD_UID
|
|
USER $BUILD_USER
|
|
ENV CUDA_MODULE_LOADING "LAZY"
|