From 111a3bc805aeca59ec48cc220b304446d88fba6e Mon Sep 17 00:00:00 2001 From: Jon Janzen Date: Mon, 13 Nov 2023 09:59:07 -0800 Subject: [PATCH] Delete .circleci directory We've disabled CircleCI for this repo --- .circleci/config.yml | 1745 ------------------------------------------ 1 file changed, 1745 deletions(-) delete mode 100644 .circleci/config.yml diff --git a/.circleci/config.yml b/.circleci/config.yml deleted file mode 100644 index 2aad8ee43f..0000000000 --- a/.circleci/config.yml +++ /dev/null @@ -1,1745 +0,0 @@ -# Use the latest 2.1 version of CircleCI pipeline process engine. -# See: https://circleci.com/docs/2.0/configuration-reference -version: 2.1 - -commands: - install-bazel: - description: "Install bazelisk" - parameters: - platform: - type: string - default: "amd64" - version: - type: string - default: "v1.17.0" - steps: - - run: - name: Install bazelisk - command: | - sudo wget -q https://github.com/bazelbuild/bazelisk/releases/download/<< parameters.version >>/bazelisk-linux-<< parameters.platform >> -O /usr/bin/bazel - sudo chmod a+x /usr/bin/bazel - - install-cmake: - description: Install CMake. - parameters: - cache: - default: true - description: Whether or not to cache the installation. - type: boolean - cache-key: - default: v1 - description: String to use in cache key. Typically overridden when needed to bust cache. - type: string - install-dir: - default: $HOME/cmake - description: The path to install CMake to. - type: string - version: - default: 3.18.2 - description: The version of CMake to install. - type: string - steps: - - when: - condition: - equal: - - << parameters.cache >> - - true - steps: - - run: - command: mkdir -pv << parameters.install-dir >> - name: Prep cache restore - - restore_cache: - keys: - - cmake-<< parameters.cache-key >>-<< parameters.version >> - - run: - command: | - echo 'export PATH="<< parameters.install-dir >>/bin:$PATH"' >> $BASH_ENV - name: Add CMake to PATH - - run: - command: | - if which cmake; then - if cmake --version | grep "<< parameters.version >>"; then - echo "CMake is already installed." - exit 0 - else - echo "CMake is already installed but it is the wrong version." - fi - fi - - rm -rf << parameters.install-dir >>/* - echo "Installing the requested version of CMake." - baseUrl="https://github.com/Kitware/CMake/releases/download/" - url="${baseUrl}/v<< parameters.version >>/cmake-<< parameters.version >>-Linux-x86_64.tar.gz" - curl -sSL -o /tmp/cmake.tar.gz $url - tar -C << parameters.install-dir >> --strip-components 1 -zxf /tmp/cmake.tar.gz - name: Install CMake - - run: - command: rm -rf << parameters.install-dir >>/{doc,man} - name: Remove unnecessary files - - run: - command: | - ls -l << parameters.install-dir >>/bin - echo $PATH - cmake --version - name: Verify CMake installation - - when: - condition: - equal: - - << parameters.cache >> - - true - steps: - - save_cache: - key: cmake-<< parameters.cache-key >>-<< parameters.version >> - paths: - - << parameters.install-dir >> - - install-nvctk: - description: "Install NVIDIA Container Toolkit" - steps: - - run: - name: "Install NVIDIA Container Toolkit" - command: | - sudo apt-get update - sudo apt-get install -y nvidia-container-toolkit - sudo nvidia-ctk runtime configure --runtime=docker - sudo systemctl restart docker - - run: - name: "Test NVCTK" - command: | - sudo docker run --rm --runtime=nvidia --gpus all nvidia/cuda:11.6.2-base-ubuntu20.04 nvidia-smi - - install-cudnn: - description: "Install CUDNN 8.9.5" - parameters: - os: - type: string - default: "ubuntu2004" - platform: - type: string - default: "x86_64" - cudnn-version: - type: string - default: "8.9.5.30" - cuda-version: - type: string - default: "cuda12.0" - steps: - - run: - name: Install CUDNN - command: | - cd ~ - wget https://developer.download.nvidia.com/compute/cuda/repos/<< parameters.os >>/<< parameters.platform >>/cuda-<< parameters.os >>.pin - sudo mv cuda-<< parameters.os >>.pin /etc/apt/preferences.d/cuda-repository-pin-600 - sudo apt-key adv --fetch-keys https://developer.download.nvidia.com/compute/cuda/repos/<< parameters.os >>/<< parameters.platform >>/3bf863cc.pub - sudo add-apt-repository "deb https://developer.download.nvidia.com/compute/cuda/repos/<< parameters.os >>/<< parameters.platform >>/ /" - sudo apt-get update - - sudo apt-get install libcudnn8=<< parameters.cudnn-version >>-1+<< parameters.cuda-version >> - sudo apt-get install libcudnn8-dev=<< parameters.cudnn-version >>-1+<< parameters.cuda-version >> - - - run: - when: on_fail - name: Dump apt sources - command: cat /etc/apt/sources.list - - install-cuda: - description: "Install CUDA" - parameters: - os: - type: string - default: "ubuntu2004" - platform: - type: string - default: "x86_64" - architecture: - type: string - default: "amd64" - cuda-pkg-name: - type: string - default: "cuda-toolkit-12-1" - cuda-pkg-version: - type: string - default: "12-1" - cuda-version: - type: string - default: "12.1.1" - steps: - - run: - name: Install CUDA - command: | - cd ~ - wget https://developer.download.nvidia.com/compute/cuda/repos/<< parameters.os >>/<< parameters.platform >>/cuda-<< parameters.os >>.pin - sudo mv cuda-<< parameters.os >>.pin /etc/apt/preferences.d/cuda-repository-pin-600 - wget https://developer.download.nvidia.com/compute/cuda/<< parameters.cuda-version >>/local_installers/cuda-repo-ubuntu2004-<< parameters.cuda-pkg-version >>-local_<< parameters.cuda-version >>-520.61.05-1_<< parameters.architecture >>.deb - sudo dpkg -i cuda-repo-ubuntu2004-<< parameters.cuda-pkg-version >>-local_<< parameters.cuda-version >>-520.61.05-1_<< parameters.architecture >>.deb - sudo cp /var/cuda-repo-ubuntu2004-<< parameters.cuda-pkg-version >>-local/cuda-*-keyring.gpg /usr/share/keyrings/ - sudo apt-get update - sudo apt-get install -y cuda - - - run: - when: on_fail - name: Dump apt sources - command: cat /etc/apt/sources.list - - create-env: - description: "Install dependencies for Torch-TensorRT" - parameters: - os: - type: string - default: "ubuntu2004" - platform: - type: string - default: "x86_64" - architecture: - type: string - default: "amd64" - cuda-string-version: - type: string - default: "cuda12.0" - cudnn-version: - type: string - default: "8.9.5.30" - trt-version-short: - type: string - default: "8.6.1" - trt-version-long: - type: string - default: "8.6.1.6-1" - bazel-platform: - type: string - default: "amd64" - steps: - - install-cudnn: - os: << parameters.os >> - platform: << parameters.platform >> - cudnn-version: << parameters.cudnn-version >> - - run: - name: Install Tensorrt - command: | - cd ~ - sudo apt-key adv --fetch-keys https://developer.download.nvidia.com/compute/cuda/repos/<< parameters.os >>/<< parameters.platform >>/3bf863cc.pub - sudo add-apt-repository "deb https://developer.download.nvidia.com/compute/cuda/repos/<< parameters.os >>/<< parameters.platform >>/ /" - sudo apt-get update - - sudo apt-get install libnvinfer8=<< parameters.trt-version-long >>+<< parameters.cuda-string-version >> libnvinfer-plugin8=<< parameters.trt-version-long>>+<< parameters.cuda-string-version >> libnvinfer-dev=<< parameters.trt-version-long >>+<< parameters.cuda-string-version >> libnvinfer-plugin-dev=<< parameters.trt-version-long>>+<< parameters.cuda-string-version >> libnvinfer-headers-dev=<< parameters.trt-version-long>>+<< parameters.cuda-string-version >> libnvinfer-headers-plugin-dev=<< parameters.trt-version-long>>+<< parameters.cuda-string-version >> - - install-bazel: - platform: << parameters.bazel-platform >> - - setup-py-version: - description: "Set python version" - parameters: - python-version: - type: string - default: "3.10.9" - steps: - - run: - name: Set python version - command: | - pyenv install << parameters.python-version >> - pyenv global << parameters.python-version >> - - create-py-env: - description: "Install python dependencies" - parameters: - trt-version-long: - type: string - default: "8.6.1" - cudnn-version-long: - type: string - default: "8.9.5.30" - steps: - - run: - name: Set up python environment - command: | - pip3 install --upgrade pip - pip3 install wheel setuptools - pip3 install nvidia-pyindex - pip3 install tabulate - pip3 install tensorrt==<< parameters.trt-version-long >> nvidia-cudnn-cu12==<< parameters.cudnn-version-long >> - pip3 install pytest parameterized expecttest nox - - install-torch-from-index: - description: "Install Pytorch" - parameters: - torch-build: - type: string - default: "2.1.0.dev20230703+cu121" - torchvision-build: - type: string - default: "0.16.0.dev20230703+cu121" - torch-build-index: - type: string - default: "https://download.pytorch.org/whl/nightly/cu121" - torchvision-build-index: - type: string - default: "https://download.pytorch.org/whl/nightly/cu121" - steps: - - run: - name: Install Torch - command: | - pip3 install --upgrade pip - pip3 install torch==<< parameters.torch-build >> torchvision==<< parameters.torchvision-build >> --extra-index-url https://pypi.nvidia.com --extra-index-url << parameters.torch-build-index >> --extra-index-url << parameters.torchvision-build-index >> - - build-py: - description: "Build the torch-tensorrt python release (pre-cxx11-abi)" - parameters: - cuda-version: - type: string - torch-build-index: - type: string - platform: - type: string - default: "x86_64" - steps: - - run: - name: Build torch-tensorrt python release (pre-cxx11-abi) - command: | - export CUDA_HOME=/usr/local/cuda-<< parameters.cuda-version >>/ - mv toolchains/ci_workspaces/WORKSPACE.<< parameters.platform >> WORKSPACE - python3 -m pip install pip==22.0.2 - python3 -m pip wheel --no-deps --verbose --pre . --extra-index-url https://pypi.nvidia.com --extra-index-url << parameters.torch-build-index >> -w dist - python3 -m pip install dist/torch_tensorrt* - mkdir -p /tmp/dist/builds - cp dist/torch_tensorrt* /tmp/dist/builds - - build-py-legacy: - description: "Build the torch-tensorrt python legacy release (pre-cxx11-abi)" - parameters: - cuda-version: - type: string - torch-build-index: - type: string - platform: - type: string - default: "x86_64" - steps: - - run: - name: Build torch-tensorrt python legacy release (pre-cxx11-abi) - command: | - export CUDA_HOME=/usr/local/cuda-<< parameters.cuda-version >>/ - mv toolchains/ci_workspaces/WORKSPACE.<< parameters.platform >>.legacy WORKSPACE - mv toolchains/legacy/pyproject.toml pyproject.toml - python3 -m pip install wheel setuptools pyyaml - python3 -m pip install pybind11==2.6.2 - python3 setup.py bdist_wheel --legacy - python3 -m pip install dist/torch_tensorrt* - mkdir -p /tmp/dist/builds - cp dist/torch_tensorrt* /tmp/dist/builds - - build-py-cxx11-abi: - description: "Build the torch-tensorrt python release (cxx11-abi)" - parameters: - cuda-version: - type: string - torch-build-index: - type: string - platform: - type: string - default: "x86_64" - release: - type: boolean - default: false - steps: - - run: - name: Build setup - command: | - mv ~/project/toolchains/ci_workspaces/WORKSPACE.<< parameters.platform >> ~/project/WORKSPACE - - when: - condition: << parameters.release >> - steps: - - run: - name: Build torch-tensorrt python release package - command: | - export CUDA_HOME=/usr/local/cuda-<< parameters.cuda-version >>/ - python3 -m pip install pip==22.0.2 - python3 -m pip wheel --no-deps --verbose --pre . --extra-index-url https://pypi.nvidia.com --extra-index-url << parameters.torch-build-index >> --config-setting="--build-option=--release" --config-setting="--build-option=--use-cxx11-abi" -w dist - python3 -m pip install dist/torch_tensorrt* - mkdir -p /tmp/dist/builds - cp dist/torch_tensorrt* /tmp/dist/builds - - unless: - condition: << parameters.release >> - steps: - - run: - name: Build torch-tensorrt python package - command: | - export CUDA_HOME=/usr/local/cuda-<< parameters.cuda-version >>/ - python3 -m pip install pip==22.0.2 - python3 -m pip wheel --no-deps --verbose --pre . --extra-index-url https://pypi.nvidia.com --extra-index-url << parameters.torch-build-index >> --config-setting="--build-option=--use-cxx11-abi" -w dist - python3 -m pip install dist/torch_tensorrt* - mkdir -p /tmp/dist/builds - cp dist/torch_tensorrt* /tmp/dist/builds - - build-py-fx-only: - description: "Build the torch-tensorrt python release with only the fx backend" - parameters: - cuda-version: - type: string - torch-build-index: - type: string - platform: - type: string - default: "x86_64" - steps: - - setup-py-version - - run: - name: Build torch-tensorrt python release with only the fx backend - command: | - export CUDA_HOME=/usr/local/cuda-<< parameters.cuda-version >>/ - mv toolchains/ci_workspaces/WORKSPACE.<< parameters.platform >> WORKSPACE - python3 -m pip install pip==22.0.2 - python3 -m pip wheel --no-deps --verbose --pre . --extra-index-url https://pypi.nvidia.com --extra-index-url << parameters.torch-build-index >> --config-setting="--build-option=--fx-only" -w dist - python3 -m pip install dist/torch_tensorrt* - mkdir -p /tmp/dist/builds - cp dist/torch_tensorrt* /tmp/dist/builds - - build-py-ngc: - description: "Build the torch-tensorrt python release for NGC PyTorch (cxx11-abi)" - parameters: - release: - type: boolean - default: false - torch-base-image: - type: string - default: nvcr.io/nvidia/pytorch:latest - steps: - - setup-py-version - - run: - name: Log into docker - command: | - docker login --username="\$oauthtoken" --password=$NGC_TOKEN nvcr.io - - run: - name: Pull image - command: | - docker pull << parameters.torch-base-image >> - - run: - name: Create container - command: | - docker run -it -d --name ngc_build_container -v /home/circleci/project:/workspace -e BAZEL_VERSION="$(cat /home/circleci/project/.bazelversion)" -w /workspace << parameters.torch-base-image >> - - run: - name: Install build deps - command: | - export BAZEL_VERSION="$(cat /home/circleci/project/.bazelversion)" - docker exec ngc_build_container bash -c "wget https://github.com/bazelbuild/bazel/releases/download/$BAZEL_VERSION/bazel-$BAZEL_VERSION-linux-x86_64 -O /usr/bin/bazel && chmod a+x /usr/bin/bazel" - docker exec ngc_build_container bash -c "mv /workspace/docker/WORKSPACE.ngc /workspace/WORKSPACE" - docker exec ngc_build_container bash -c "git config --global --add safe.directory /workspace" - - when: - condition: << parameters.release >> - steps: - - run: - name: Build torch-tensorrt release build for NGC - command: | - docker exec ngc_build_container bash -c "cd /workspace/ && python3 -m pip wheel --no-deps --verbose --pre . --config-setting="--build-option=--release" --config-setting="--build-option=--use-cxx11-abi" -w dist" - - unless: - condition: << parameters.release >> - steps: - - run: - name: Build torch-tensorrt debug build for NGC - command: | - docker exec ngc_build_container bash -c "cd /workspace/py && python3 -m pip wheel --no-deps --verbose --pre . --config-setting="--build-option=--use-cxx11-abi" -w dist" - - run: - name: Collect builds - command: | - mkdir -p /tmp/dist/builds - cp /home/circleci/project/py/dist/* /tmp/dist/builds - - build-cmake: - description: "Build the torch-tensorrt using CMake" - parameters: - platform: - type: string - default: "x86_64" - cuda-version: - type: string - steps: - - run: - name: Build torch-tensorrt library with CMake - command: | - mkdir build - export PATH=$PATH:/usr/local/<< parameters.cuda-version >>/bin - ~/cmake/bin/cmake -S. -Bbuild \ - -DCMAKE_MODULE_PATH=cmake/Module \ - -DTorch_DIR=/opt/circleci/.pyenv/versions/3.10.9/lib/python3.10/site-packages/torch/share/cmake/Torch \ - -DTensorRT_ROOT=/usr \ - -DCMAKE_BUILD_TYPE=Debug - cmake --build build -- -j12 - mkdir -p /tmp/dist/builds - cp -r build/bin /tmp/dist/builds - cp -r build/lib /tmp/dist/builds - cp -r cpp/include /tmp/dist/builds - - - dump-test-env: - description: "Dump the test env to console" - steps: - - run: - name: Test torch - command: | - python3 -c "import torch; print(torch.cuda.is_available()); print(torch.cuda.device_count())" - - - run: - name: Get torch-tensorrt version information - command: | - python3 -c "import torch_tensorrt; torch_tensorrt.dump_build_info()" - - pull-test-models: - description: "Pull the test model set" - steps: - - run: - name: Pull test models - environment: - USE_HOST_DEPS: "1" - command: | - cd tests/modules - pip3 install -r requirements.txt - python3 hub.py - cd ~/project - - test-ts-core: - description: "Test torchscript backend c++ api" - parameters: - platform: - type: string - default: "x86_64" - steps: - - pull-test-models - - run: mkdir -p /tmp/artifacts - - run: - name: Run core / C++ tests - no_output_timeout: 15m - environment: - LD_LIBRARY_PATH: "/opt/circleci/.pyenv/versions/3.10.9/lib/python3.10/site-packages/torch_tensorrt.libs:/home/circleci/project/bazel-project/external/libtorch_pre_cxx11_abi/lib/:/home/circleci/project/bazel-project/external/tensorrt/lib/:/usr/local/cuda-12.1/lib64/:$LD_LIBRARY_PATH" - command: | - set -e - mv toolchains/ci_workspaces/WORKSPACE.<< parameters.platform >> WORKSPACE - bazel query 'kind(cc_*, tests(//tests))' --noshow_progress >> /tmp/test_manifest.txt - circleci tests split < /tmp/test_manifest.txt > /tmp/node_test_manifest.txt - bazel test $(cat /tmp/node_test_manifest.txt) --test_arg=--gtest_output=xml:/tmp/artifacts/test_results/ --jobs 4 --config ci_testing --config pre_cxx11_abi --noshow_progress --test_timeout=20000 - - run: - name: Collect logs - when: on_fail - command: | - mkdir -p /tmp/testlogs - cp -r bazel-testlogs /tmp/testlogs - sudo apt install tree - tree . > /tmp/testlogs/dir_structure.txt - - - store_test_results: - path: /tmp/artifacts - - store_artifacts: - path: /tmp/testlogs - - test-ts-py-api: - description: "Run L0 torch-tensorrt python tests" - steps: - - pull-test-models - - run: - name: Run L0 torch-tensorrt python tests - environment: - USE_HOST_DEPS: "1" - PYT_PATH: "/opt/circleci/.pyenv/versions/3.9.4/lib/python3.9/site-packages/" - LD_LIBRARY_PATH: "/usr/lib/x86_64-linux-gnu/:/usr/local/cuda-12.1/lib64/:$LD_LIBRARY_PATH" - command: | - set -e - mkdir -p /tmp/artifacts/test_results - cd tests/py/ts/ - pytest --junitxml=/tmp/artifacts/test_results/api/api_test_results.xml api/ - pytest --junitxml=/tmp/artifacts/test_results/models/models_test_results.xml models/ - pytest --junitxml=/tmp/artifacts/test_results/integrations/integrations_test_results.xml integrations/ - cd ~/project - - - store_test_results: - path: /tmp/artifacts - - store_artifacts: - path: /tmp/testlogs - -# =================== FX tests start ======================== # - test-fx_core: - description: "Test the fx core" - steps: - - run: - name: Run FX core tests - command: | - cd py/torch_tensorrt/fx/test - pushd core/ - pytest --junitxml=/tmp/artifacts/test_results/fx/core/test_results.xml - popd - - - store_test_results: - path: /tmp/artifacts - - store_artifacts: - path: /tmp/testlogs - - test-fx_converters_acc: - description: "Test the fx acc converters" - steps: - - run: - name: Run FX converter tests - command: | - set -e - cd py/torch_tensorrt/fx/test - TESTS_TO_RUN=$(circleci tests glob "converters/acc_op/test_*.py" | circleci tests split --split-by=timings) - pytest --junitxml=/tmp/artifacts/test_results/fx/converters/acc_op/test_results.xml $TESTS_TO_RUN - - - store_test_results: - path: /tmp/artifacts - - store_artifacts: - path: /tmp/testlogs - - test-fx_converters_aten: - description: "Test the fx aten converters" - steps: - - run: - name: Run FX converter tests - command: | - set -e - cd py/torch_tensorrt/fx/test - TESTS_TO_RUN=$(circleci tests glob "converters/aten_op/test_*.py" | circleci tests split --split-by=timings) - pytest --junitxml=/tmp/artifacts/test_results/fx/converters/aten_op/test_results.xml $TESTS_TO_RUN - - - store_test_results: - path: /tmp/artifacts - - store_artifacts: - path: /tmp/testlogs - - test-fx_converters_vanilla: - description: "Test the fx vanilla converters" - steps: - - run: - name: Run FX converter tests - command: | - cd py/torch_tensorrt/fx/test - pushd converters/vanilla/ - pytest --junitxml=/tmp/artifacts/test_results/fx/converters/vanilla/test_results.xml - popd - - - store_test_results: - path: /tmp/artifacts - - store_artifacts: - path: /tmp/testlogs - - test-fx_passes: - description: "Test the fx passes" - steps: - - run: - name: Run FX passes - command: | - cd py/torch_tensorrt/fx/test - pushd passes - list_passes=$(ls | grep -v test_setitem*) - pytest $list_passes --junitxml=/tmp/artifacts/test_results/fx/passes/test_results.xml - popd - - store_test_results: - path: /tmp/artifacts - - store_artifacts: - path: /tmp/testlogs - - test-fx_tools: - description: "Test the fx tools" - steps: - - run: - name: Run FX tools - command: | - cd py/torch_tensorrt/fx/test - pushd tools - pytest --junitxml=/tmp/artifacts/test_results/fx/tools/test_results.xml - popd - - store_test_results: - path: /tmp/artifacts - - store_artifacts: - path: /tmp/testlogs - - test-fx_trt_lower: - description: "Test the fx TRT lowering" - steps: - - run: - name: Run FX TRT lowering - command: | - cd py/torch_tensorrt/fx/test - pushd trt_lower - pytest --junitxml=/tmp/artifacts/test_results/fx/trt_lower/test_results.xml - popd - - store_test_results: - path: /tmp/artifacts - - store_artifacts: - path: /tmp/testlogs - - test-fx_tracer: - description: "Test all fx tracers" - steps: - - run: - name: Run FX tracer - command: | - cd py/torch_tensorrt/fx/test - pushd tracer - list_tracer=$(ls | grep -v test_dispatch_*) - pytest $list_tracer --junitxml=/tmp/artifacts/test_results/fx/tracer/test_results.xml - popd - - store_test_results: - path: /tmp/artifacts - - store_artifacts: - path: /tmp/testlogs - - test-fx_tracer_acc: - description: "Test the fx acc tracer only" - steps: - - run: - name: Run FX tracer - command: | - cd py/torch_tensorrt/fx/test - pushd tracer - list_tracer=$(ls | grep test_acc) - pytest $list_tracer --junitxml=/tmp/artifacts/test_results/fx/tracer/test_results.xml - popd - - store_test_results: - path: /tmp/artifacts - - store_artifacts: - path: /tmp/testlogs - - test-fx_quant: - description: "Test the fx quant" - steps: - - run: - name: Run FX quant tests - command: | - cd py/torch_tensorrt/fx/test - pushd quant/ - pytest --junitxml=/tmp/artifacts/test_results/fx/quant/test_results.xml - popd - - - store_test_results: - path: /tmp/artifacts - - store_artifacts: - path: /tmp/testlogs - - test-fx: - description: "Test the fx backend" - steps: - - run: - name: Run fx tests - command: | - mkdir -p /tmp/artifacts/test_results - - test-fx_converters_acc - - test-fx_converters_aten - - test-fx_converters_vanilla - - test-fx_passes - - test-fx_tools - - test-fx_trt_lower - - test-fx_tracer - - test-fx_core - - test-fx_quant - - store_test_results: - path: /tmp/artifacts - - store_artifacts: - path: /tmp/testlogs - - test-fx-no-aten: - description: "Test the fx backend without aten operators" - steps: - - run: - name: Run fx tests without aten ops - command: | - mkdir -p /tmp/artifacts/test_results - - test-fx_converters_acc - - test-fx_converters_vanilla - - test-fx_passes - - test-fx_tools - - test-fx_trt_lower - - test-fx_tracer_acc - - test-fx_core - - test-fx_quant - - store_test_results: - path: /tmp/artifacts - - store_artifacts: - path: /tmp/testlogs - -# =================== FX tests end ======================== # - -# =================== Dynamo tests start ======================== # - - test-dynamo-backend: - description: "Test Dynamo backend tests" - steps: - - run: - name: Run Dynamo backend tests - command: | - cd tests/py/dynamo/backend/ - TESTS_TO_RUN=$(circleci tests glob "test_*.py" | circleci tests split --split-by=timings) - pytest --junitxml=/tmp/artifacts/test_results/dynamo/backend/test_results.xml $TESTS_TO_RUN - - - store_test_results: - path: /tmp/artifacts - - store_artifacts: - path: /tmp/testlogs - - test-dynamo-shared_utilities: - description: "Test Dynamo shared utilities tests" - steps: - - run: - name: Run Dynamo lowering, partitioning, runtime tests - command: | - cd tests/py/dynamo/ - TESTS_TO_RUN=$(circleci tests glob "runtime/test_*.py" "partitioning/test_*.py" "lowering/test_*.py" | circleci tests split --split-by=timings) - pytest --junitxml=/tmp/artifacts/test_results/dynamo/shared_utilities/test_results.xml $TESTS_TO_RUN - - - store_test_results: - path: /tmp/artifacts - - store_artifacts: - path: /tmp/testlogs - - test-dynamo-models_torch_compile: - description: "Test the Dynamo models via torch_compile path" - steps: - - run: - name: Run Dynamo models via torch_compile path - command: | - cd tests/py/dynamo/models - pip3 install timm - pip3 install transformers - pytest test_models.py --junitxml=/tmp/artifacts/test_results/dynamo/backend/test_results.xml --ir torch_compile - - - store_test_results: - path: /tmp/artifacts - - store_artifacts: - path: /tmp/testlogs - - test-dynamo-models_export: - description: "Test the Dynamo models via torch_export path" - steps: - - run: - name: Run Dynamo models via torch_export path - command: | - cd tests/py/dynamo/models - pip3 install timm - pip3 install transformers - pytest test_models_export.py --junitxml=/tmp/artifacts/test_results/dynamo/backend/test_results.xml --ir dynamo - - - store_test_results: - path: /tmp/artifacts - - store_artifacts: - path: /tmp/testlogs - - test-dynamo-export_serde: - description: "Test the export serialize/deserialize functionality for Dynamo models" - steps: - - run: - name: Run Dynamo models and test export serde with TRT compiled modules - command: | - cd tests/py/dynamo/models - pytest test_export_serde.py --junitxml=/tmp/artifacts/test_results/dynamo/backend/test_results.xml --ir dynamo - - - store_test_results: - path: /tmp/artifacts - - store_artifacts: - path: /tmp/testlogs - - test-dynamo-converters: - description: "Test the Dynamo aten converters" - steps: - - run: - name: Run Dynamo converter tests - command: | - cd tests/py/dynamo/conversion - TESTS_TO_RUN=$(circleci tests glob "test_*.py" | circleci tests split --split-by=timings) - pytest --junitxml=/tmp/artifacts/test_results/dynamo/conversion/test_results.xml $TESTS_TO_RUN - - - store_test_results: - path: /tmp/artifacts - - store_artifacts: - path: /tmp/testlogs - -# =================== Dynamo tests end ======================== # - -# Define a job to be invoked later in a workflow. -# See: https://circleci.com/docs/2.0/configuration-reference/#jobs -jobs: - build-x86_64-linux: - parameters: - torch-build: - type: string - torchvision-build: - type: string - torch-build-index: - type: string - python-version: - type: string - cuda-version: - type: string - cxx11-abi: - type: boolean - default: false - legacy: - type: boolean - default: false - ci-image: - type: string - default: "linux-cuda-12:2023.05.1" - machine: - image: << parameters.ci-image >> - resource_class: gpu.nvidia.small - steps: - - checkout - - setup-py-version: - python-version: << parameters.python-version >> - - create-env: - os: "ubuntu2004" - platform: "x86_64" - cudnn-version: << pipeline.parameters.cudnn-version >> - trt-version-short: << pipeline.parameters.trt-version-short >> - bazel-platform: "amd64" - - install-torch-from-index: - torch-build: << parameters.torch-build >> - torchvision-build: << parameters.torchvision-build >> - torch-build-index: << parameters.torch-build-index >> - - when: - condition: << parameters.cxx11-abi >> - steps: - - build-py-cxx11-abi: - cuda-version: << parameters.cuda-version >> - torch-build-index: << parameters.torch-build-index >> - - unless: - condition: << parameters.cxx11-abi >> - steps: - - when: - condition: << parameters.legacy >> - steps: - - build-py-legacy: - cuda-version: << parameters.cuda-version >> - torch-build-index: << parameters.torch-build-index >> - - unless: - condition: << parameters.legacy >> - steps: - - build-py: - cuda-version: << parameters.cuda-version >> - torch-build-index: << parameters.torch-build-index >> - - run: - name: Move to build dir - command: | - mkdir -p /tmp/dist/x86_64-linux - cp -r /tmp/dist/builds/* /tmp/dist/x86_64-linux - - persist_to_workspace: - root: /tmp/dist - paths: - - x86_64-linux - - store_artifacts: - path: /tmp/dist/x86_64-linux - destination: x86_64-linux - - test-core-cpp-x86_64-linux: - parameters: - torch-build: - type: string - python-version: - type: string - torch-build-index: - type: string - torchvision-build: - type: string - trt-version-short: - type: string - trt-version-long: - type: string - cudnn-version: - type: string - ci-image: - type: string - default: "linux-cuda-12:2023.05.1" - platform: - type: string - default: "x86_64" - machine: - image: << parameters.ci-image >> - resource_class: gpu.nvidia.medium - parallelism: 4 - steps: - - checkout - - setup-py-version: - python-version: << parameters.python-version >> - - create-env: - os: "ubuntu2004" - platform: "x86_64" - cudnn-version: << parameters.cudnn-version >> - trt-version-short: << parameters.trt-version-short >> - bazel-platform: "amd64" - - create-py-env: - trt-version-long: << parameters.trt-version-long >> - - install-torch-from-index: - torch-build: << parameters.torch-build >> - torchvision-build: << parameters.torchvision-build >> - torch-build-index: << parameters.torch-build-index >> - - attach_workspace: - at: /tmp/dist - - run: - name: "Install torch-tensorrt" - command: pip3 install /tmp/dist/x86_64-linux/*cp310-cp310*.whl - - dump-test-env - - test-ts-core: - platform: << parameters.platform >> - - test-py-ts-x86_64-linux: - parameters: - torch-build: - type: string - torchvision-build: - type: string - torch-build-index: - type: string - trt-version-long: - type: string - python-version: - type: string - ci-image: - type: string - default: "linux-cuda-12:2023.05.1" - machine: - image: << parameters.ci-image >> - resource_class: gpu.nvidia.medium - steps: - - checkout - - setup-py-version: - python-version: << parameters.python-version >> - - create-py-env: - trt-version-long: << parameters.trt-version-long >> - - attach_workspace: - at: /tmp/dist - - install-torch-from-index: - torch-build: << parameters.torch-build >> - torchvision-build: << parameters.torchvision-build >> - torch-build-index: << parameters.torch-build-index >> - - run: - name: "Install torch-tensorrt" - command: pip3 install --pre /tmp/dist/x86_64-linux/*cp310-cp310*.whl - - dump-test-env - - test-ts-py-api - - test-py-fx-x86_64-linux: - parameters: - torch-build: - type: string - torchvision-build: - type: string - torch-build-index: - type: string - trt-version-long: - type: string - python-version: - type: string - ci-image: - type: string - default: "linux-cuda-12:2023.05.1" - parallelism: 8 - machine: - image: << parameters.ci-image >> - resource_class: gpu.nvidia.medium - steps: - - checkout - - setup-py-version: - python-version: << parameters.python-version >> - - attach_workspace: - at: /tmp/dist/ - - install-torch-from-index: - torch-build: << parameters.torch-build >> - torchvision-build: << parameters.torchvision-build >> - torch-build-index: << parameters.torch-build-index >> - - create-py-env: - trt-version-long: << parameters.trt-version-long >> - - install-cudnn - - run: - name: "Install torch-tensorrt" - command: pip3 install --pre /tmp/dist/x86_64-linux/*cp310-cp310*.whl - # We install torch after torch-trt because pip automatically enforces the version constraint otherwise - - dump-test-env - - test-fx - - test-py-fx-x86_64-linux-no-aten: - parameters: - torch-build: - type: string - torchvision-build: - type: string - torch-build-index: - type: string - trt-version-long: - type: string - python-version: - type: string - ci-image: - type: string - default: "linux-cuda-12:2023.05.1" - parallelism: 8 - machine: - image: << parameters.ci-image >> - resource_class: gpu.nvidia.medium - steps: - - checkout - - setup-py-version: - python-version: << parameters.python-version >> - - attach_workspace: - at: /tmp/dist/ - - install-torch-from-index: - torch-build: << parameters.torch-build >> - torchvision-build: << parameters.torchvision-build >> - torch-build-index: << parameters.torch-build-index >> - - create-py-env: - trt-version-long: << parameters.trt-version-long >> - - install-cudnn - # - run: - # name: "Set LD_LIBRARY_PATH path to include the installed CUDNN" - # command: export LD_LIBRARY_PATH=/usr/lib/x86_64-linux-gnu/:$LD_LIBRARY_PATH - - run: - name: "Install torch-tensorrt" - command: pip3 install --pre /tmp/dist/x86_64-linux/*cp310-cp310*.whl - # We install torch after torch-trt because pip automatically enforces the version constraint otherwise - - dump-test-env - - test-fx-no-aten - - test-py-dynamo-x86_64-linux: - parameters: - torch-build: - type: string - torchvision-build: - type: string - torch-build-index: - type: string - trt-version-long: - type: string - python-version: - type: string - ci-image: - type: string - default: "linux-cuda-12:2023.05.1" - parallelism: 4 - machine: - image: << parameters.ci-image >> - resource_class: gpu.nvidia.medium - steps: - - checkout - - setup-py-version: - python-version: << parameters.python-version >> - - attach_workspace: - at: /tmp/dist/ - - install-torch-from-index: - torch-build: << parameters.torch-build >> - torchvision-build: << parameters.torchvision-build >> - torch-build-index: << parameters.torch-build-index >> - - create-py-env: - trt-version-long: << parameters.trt-version-long >> - - install-cudnn - # - run: - # name: "Set LD_LIBRARY_PATH path to include the installed CUDNN" - # command: export LD_LIBRARY_PATH=/usr/lib/x86_64-linux-gnu/:$LD_LIBRARY_PATH - - run: - name: "Install torch-tensorrt" - command: pip3 install --pre /tmp/dist/x86_64-linux/*cp310-cp310*.whl - # We install torch after torch-trt because pip automatically enforces the version constraint otherwise - - dump-test-env - - test-dynamo-converters - - test-dynamo-backend - - test-dynamo-shared_utilities - - test-dynamo-models_torch_compile - - test-dynamo-models_export - - test-dynamo-export_serde - - package-x86_64-linux: - parameters: - enabled: - type: boolean - default: false - torch-build: - type: string - torch-build-index: - type: string - ci-image: - type: string - default: "linux-cuda-12:2023.05.1" - parallelism: 4 - machine: - image: << parameters.ci-image >> - resource_class: gpu.nvidia.small - steps: - - when: - condition: << parameters.enabled >> - steps: - - checkout - - install-nvctk - - run: - name: "Build packaging container" - command: | - cd ~/project/py/ - docker build -t torch_tensorrt_release_env --build-arg trt_version=<< pipeline.parameters.trt-version-short >> -f ci/Dockerfile.ci . - - run: - name: Build Python packages - command: | - cd ~/project/py/ - cp ~/project/toolchains/ci_workspaces/WORKSPACE.x86_64.release.rhel ~/project/WORKSPACE - docker run -it --rm --gpus all -v ~/project:/workspace/project torch_tensorrt_release_env /bin/bash -c "source /workspace/project/py/ci/build_whl.sh && $(circleci tests split --total=4 ~/project/py/ci/build_manifest.txt)" - - run: - name: Collect packages - command: | - mkdir -p /tmp/dist/x86_64-linux - cp -r ~/project/py/wheelhouse/* /tmp/dist/x86_64-linux - - persist_to_workspace: - root: /tmp/dist - paths: - - x86_64-linux - - store_artifacts: - path: /tmp/dist/x86_64-linux - destination: x86_64-linux - - unless: - condition: << parameters.enabled >> - steps: - - run: - name: Skipped packaging - command: echo -e "Packaging stage not enabled" - - # TODO: Merge this with above - package-x86_64-linux-cxx11-abi: - parameters: - enabled: - type: boolean - default: false - torch-build: - type: string - torch-build-index: - type: string - ci-image: - type: string - default: "linux-cuda-12:2023.05.1" - machine: - image: << parameters.ci-image >> - resource_class: gpu.nvidia.small - steps: - - when: - condition: << parameters.enabled >> - steps: - - checkout - - create-env: - os: "ubuntu2004" - platform: "x86_64" - cudnn-version: << pipeline.parameters.cudnn-version >> - trt-version-short: << pipeline.parameters.trt-version-short >> - bazel-platform: "amd64" - - install-torch-from-index: - torch-build: << parameters.torch-build >> - torch-build-index: << parameters.torch-build-index >> - - run: - name: Build cxx11-abi tarball - command: | - set -e - cd ~/project/ - cp ~/project/toolchains/ci_workspaces/WORKSPACE.x86_64.release.ubuntu ~/project/WORKSPACE - bazel build //:libtorchtrt -c opt --noshow_progress - sudo chown -R $(whoami) ~/project/py - cd ~/project/py - CUDA_VERSION=$(python3 -c "from versions import __cuda_version__;print(__cuda_version__)") - TORCHTRT_VERSION=$(python3 -c "from versions import __version__;print(__version__)") - TRT_VERSION=$(python3 -c "from versions import __tensorrt_version__;print(__tensorrt_version__)") - CUDNN_VERSION=$(python3 -c "from versions import __cudnn_version__;print(__cudnn_version__)") - TORCH_VERSION=$(python3 -c "from torch import __version__;print(__version__.split('+')[0])") - python3 -m pip install pip==22.0.2 - python3 -m pip install -r ~/project/py/requirements.txt - TORCH_VERSION=$(python3 -c "from torch import __version__;print(__version__.split('+')[0])") - mkdir -p ~/project/py/dist/ - cp ~/project/bazel-bin/libtorchtrt.tar.gz ~/project/py/dist/libtorchtrt-${TORCHTRT_VERSION}-cudnn${CUDNN_VERSION}-tensorrt${TRT_VERSION}-cuda${CUDA_VERSION}-libtorch${TORCH_VERSION}-x86_64-linux.tar.gz - - run: - name: Collect packages - command: | - mkdir -p /tmp/dist/x86_64-linux - cp -r ~/project/py/dist/* /tmp/dist/x86_64-linux - - store_artifacts: - path: /tmp/dist/x86_64-linux - destination: x86_64-linux-pkgs - - unless: - condition: << parameters.enabled >> - steps: - - run: - name: Skipped packaging - command: echo -e "Packaging stage not enabled" - - package-aarch64-jetson: - parameters: - enabled: - type: boolean - default: true - torch-build: - type: string - jetpack-version: - type: string - cxx11-abi: - type: boolean - default: true - python-version: - type: string - default: 3.8.10 - machine: - image: ubuntu-2004:202201-02 - resource_class: arm.xlarge - steps: - - checkout - - install-cuda: - os: "ubuntu2004" - platform: "sbsa" - architecture: "arm64" - - run: - name: Install openblas - command: sudo apt install libopenblas-dev - - create-env: - os: "ubuntu2004" - platform: "sbsa" - architecture: "arm64" - cudnn-version: << pipeline.parameters.cudnn-jetson-version >> - trt-version-short: << pipeline.parameters.trt-jetson-version-short >> - bazel-platform: "arm64" - - run: - name: Set python version - command: | - pyenv install << parameters.python-version >> - pyenv global << parameters.python-version >> - - run: - name: Install NGC Torch - environment: - TORCH_INSTALL: https://developer.download.nvidia.com/compute/redist/jp/v<< parameters.jetpack-version >>/pytorch/<< parameters.torch-build >> - command: | - set -e - python3 -m pip install --upgrade pip; python3 -m pip install setuptools wheel; python3 -m pip install expecttest xmlrunner hypothesis aiohttp numpy=='1.19.4' pyyaml scipy=='1.5.3' ninja cython typing_extensions protobuf; export "LD_LIBRARY_PATH=/usr/lib/llvm-8/lib:$LD_LIBRARY_PATH"; python3 -m pip install --upgrade protobuf; python3 -m pip install --no-cache $TORCH_INSTALL - - build-py-cxx11-abi: - platform: "sbsa" - release: true - - run: - name: Build cxx11-abi tarball - command: | - set -e - cd ~/project/py/ - bazel build //:libtorchtrt -c opt --noshow_progress - CUDA_VERSION=$(cd torch_tensorrt && python3 -c "from _version import __cuda_version__;print(__cuda_version__)") - TORCHTRT_VERSION=$(cd torch_tensorrt && python3 -c "from _version import __version__;print(__version__)") - TRT_VERSION=$(cd torch_tensorrt && python3 -c "from _version import __tensorrt_version__;print(__tensorrt_version__)") - CUDNN_VERSION=$(cd torch_tensorrt && python3 -c "from _version import __cudnn_version__;print(__cudnn_version__)") - python3 -m pip install pybind11==2.6.2 - TORCH_VERSION=$(python3 -c "from torch import __version__;print(__version__.split('+')[0])") - cp ~/project/bazel-bin/libtorchtrt.tar.gz ~/project/py/dist/libtorchtrt-${TORCHTRT_VERSION}-cudnn${CUDNN_VERSION}-tensorrt${TRT_VERSION}-cuda${CUDA_VERSION}-libtorch${TORCH_VERSION}-aarch64-linux-jp<< parameters.jetpack-version >>.tar.gz - - run: - name: Move to release dir - command: | - mkdir -p /tmp/dist/jetson - cp -r ~/project/py/dist/* /tmp/dist/jetson - - store_artifacts: - path: /tmp/dist/jetson - destination: aarch64-jetson-pkgs - - build-x86_64-linux-cmake: - parameters: - cuda-version: - type: string - cudnn-version: - type: string - trt-version-short: - type: string - torch-build: - type: string - torch-build-index: - type: string - python-version: - type: string - ci-image: - type: string - default: "linux-cuda-12:2023.05.1" - machine: - image: << parameters.ci-image >> - resource_class: gpu.nvidia.small - steps: - - checkout - - setup-py-version: - python-version: << parameters.python-version >> - - create-env: - os: "ubuntu2004" - platform: "x86_64" - cudnn-version: << parameters.cudnn-version >> - trt-version-short: << parameters.trt-version-short >> - bazel-platform: "amd64" - - install-cmake: - version: 3.24.1 - - install-torch-from-index: - torch-build: << parameters.torch-build >> - torch-build-index: << parameters.torch-build-index >> - - build-cmake: - cuda-version: << parameters.cuda-version >> - - run: - name: Move to cmake build dir - command: | - mkdir -p /tmp/dist/cmake - cp -r /tmp/dist/builds/* /tmp/dist/cmake - - persist_to_workspace: - root: /tmp/dist - paths: - - cmake - - store_artifacts: - path: /tmp/dist/cmake - destination: x86_64-cmake - - build-x86_64-linux-ngc: - parameters: - torch-base-image: - type: string - ci-image: - type: string - default: "linux-cuda-12:2023.05.1" - machine: - image: << parameters.ci-image >> - resource_class: gpu.nvidia.small - steps: - - checkout - - build-py-ngc: - release: false - torch-base-image: << parameters.torch-base-image >> - - run: - name: Move to NGC dir - command: | - mkdir -p /tmp/dist/ngc - cp -r /tmp/dist/builds/* /tmp/dist/ngc - - persist_to_workspace: - root: /tmp/dist - paths: - - ngc - - store_artifacts: - path: /tmp/dist/ngc - destination: x86_64-pyt-ngc - - package-x86_64-linux-ngc: - parameters: - enabled: - type: boolean - default: false - torch-base-image: - type: string - ci-image: - type: string - default: "linux-cuda-12:2023.05.1" - machine: - image: << parameters.ci-image >> - resource_class: gpu.nvidia.small - steps: - - when: - condition: << parameters.enabled >> - steps: - - checkout - - build-py-ngc: - release: true - torch-base-image: << parameters.torch-base-image >> - - run: - name: Collect packages - command: | - mkdir -p /tmp/dist/ngc - cp -r ~/project/py/dist/* /tmp/dist/ngc - - store_artifacts: - path: /tmp/dist/ngc - destination: x86_64-ngc-pkgs - - unless: - condition: << parameters.enabled >> - steps: - - run: - name: Skipped packaging - command: echo -e "Packaging stage not enabled" - -parameters: - python-version: - type: string - default: "3.10.9" - - # Nightly platform config - cuda-version: - type: string - default: "12.1" - torch-build: - type: string - default: "2.1.0.dev20230703+cu121" - torchvision-build: - type: string - default: "0.16.0.dev20230703+cu121" - torch-build-index: - type: string - default: "https://download.pytorch.org/whl/nightly/cu121" - cudnn-version: - type: string - default: "8.9.5.30" - trt-version-short: - type: string - default: "8.6.1" - trt-version-long: - type: string - default: "8.6.1" - - # Legacy platform config - cuda-version-legacy: - type: string - default: "11.8" - torch-build-legacy: - type: string - default: "1.13.1+cu117" - torchvision-build-legacy: - type: string - default: "0.14.1+cu117" - torch-build-index-legacy: - type: string - default: "https://download.pytorch.org/whl/cu117" - cudnn-version-legacy: - type: string - default: "8.9.5.30" - trt-version-short-legacy: - type: string - default: "8.6.1" - trt-version-long-legacy: - type: string - default: "8.6.1" - ci-image-legacy: - type: string - default: "linux-cuda-11:2023.02.1" - platform-legacy: - type: string - default: "x86_64.legacy" - - # Jetson platform config - cuda-jetson-version: - type: string - default: "11.8" - torch-jetson-build: - type: string - default: "torch-1.13.0a0+d0d6b1f2.nv22.09-cp38-cp38-linux_aarch64.whl" - jetpack-version: - type: string - default: "502" - cudnn-jetson-version: - type: string - default: "8.5.0.96" - trt-jetson-version-short: - type: string - default: "8.4.1" - trt-jetson-version-long: - type: string - default: "8.4.1.5" - - torch-ngc-base-image: - type: string - default: "nvcr.io/nvidia/pytorch:22.09-py3" - - enable-packaging: - type: boolean - default: false - - enable-ngc-packaging: - type: boolean - default: false - - enable-legacy: - type: boolean - default: true - -# Invoke jobs via workflows -# See: https://circleci.com/docs/2.0/configuration-reference/#workflows -workflows: - nightly: - triggers: - - schedule: - cron: "0 0 * * *" - filters: - branches: - only: - - main - - release/**/* - jobs: - - build-x86_64-linux: - name: build-x86_64-linux - torch-build: << pipeline.parameters.torch-build >> - torchvision-build: << pipeline.parameters.torchvision-build >> - torch-build-index: << pipeline.parameters.torch-build-index >> - python-version: << pipeline.parameters.python-version >> - cuda-version: << pipeline.parameters.cuda-version >> - - - test-core-cpp-x86_64-linux: - torch-build: << pipeline.parameters.torch-build >> - torchvision-build: << pipeline.parameters.torchvision-build >> - torch-build-index: << pipeline.parameters.torch-build-index >> - trt-version-short: << pipeline.parameters.trt-version-short >> - trt-version-long: << pipeline.parameters.trt-version-long >> - cudnn-version: << pipeline.parameters.cudnn-version >> - python-version: << pipeline.parameters.python-version >> - requires: - - build-x86_64-linux - - - test-py-ts-x86_64-linux: - torch-build: << pipeline.parameters.torch-build >> - torchvision-build: << pipeline.parameters.torchvision-build >> - torch-build-index: << pipeline.parameters.torch-build-index >> - trt-version-long: << pipeline.parameters.trt-version-long >> - python-version: << pipeline.parameters.python-version >> - requires: - - build-x86_64-linux - - - test-py-fx-x86_64-linux: - torch-build: << pipeline.parameters.torch-build >> - torchvision-build: << pipeline.parameters.torchvision-build >> - torch-build-index: << pipeline.parameters.torch-build-index >> - trt-version-long: << pipeline.parameters.trt-version-long >> - python-version: << pipeline.parameters.python-version >> - requires: - - build-x86_64-linux - - - test-py-dynamo-x86_64-linux: - torch-build: << pipeline.parameters.torch-build >> - torchvision-build: << pipeline.parameters.torchvision-build >> - torch-build-index: << pipeline.parameters.torch-build-index >> - trt-version-long: << pipeline.parameters.trt-version-long >> - python-version: << pipeline.parameters.python-version >> - requires: - - build-x86_64-linux - - - build-x86_64-linux: - name: build-x86_64-linux-legacy - cuda-version: << pipeline.parameters.cuda-version-legacy >> - torch-build: << pipeline.parameters.torch-build-legacy >> - torchvision-build: << pipeline.parameters.torchvision-build-legacy >> - torch-build-index: << pipeline.parameters.torch-build-index-legacy >> - python-version: << pipeline.parameters.python-version >> - legacy: << pipeline.parameters.enable-legacy >> - ci-image: << pipeline.parameters.ci-image-legacy >> - - - test-core-cpp-x86_64-linux: - name: test-core-cpp-x86_64-linux-legacy - torch-build: << pipeline.parameters.torch-build-legacy >> - torchvision-build: << pipeline.parameters.torchvision-build-legacy >> - torch-build-index: << pipeline.parameters.torch-build-index-legacy >> - trt-version-short: << pipeline.parameters.trt-version-short-legacy >> - trt-version-long: << pipeline.parameters.trt-version-long-legacy >> - cudnn-version: << pipeline.parameters.cudnn-version-legacy >> - python-version: << pipeline.parameters.python-version >> - ci-image: << pipeline.parameters.ci-image-legacy >> - platform: << pipeline.parameters.platform-legacy >> - requires: - - build-x86_64-linux-legacy - - - test-py-ts-x86_64-linux: - name: test-py-ts-x86_64-linux-legacy - torch-build: << pipeline.parameters.torch-build-legacy >> - torchvision-build: << pipeline.parameters.torchvision-build-legacy >> - torch-build-index: << pipeline.parameters.torch-build-index-legacy >> - trt-version-long: << pipeline.parameters.trt-version-long-legacy >> - python-version: << pipeline.parameters.python-version >> - ci-image: << pipeline.parameters.ci-image-legacy >> - requires: - - build-x86_64-linux-legacy - - - test-py-fx-x86_64-linux-no-aten: - torch-build: << pipeline.parameters.torch-build-legacy >> - torchvision-build: << pipeline.parameters.torchvision-build-legacy >> - torch-build-index: << pipeline.parameters.torch-build-index-legacy >> - trt-version-long: << pipeline.parameters.trt-version-long-legacy >> - python-version: << pipeline.parameters.python-version >> - ci-image: << pipeline.parameters.ci-image-legacy >> - requires: - - build-x86_64-linux-legacy - - release: - when: << pipeline.parameters.enable-packaging >> - jobs: - - - package-x86_64-linux: - enabled: << pipeline.parameters.enable-packaging >> - torch-build: << pipeline.parameters.torch-build >> - torch-build-index: << pipeline.parameters.torch-build-index >> - - - package-x86_64-linux-cxx11-abi: - enabled: << pipeline.parameters.enable-packaging >> - torch-build: << pipeline.parameters.torch-build >> - torch-build-index: << pipeline.parameters.torch-build-index >> - - - test-core-cpp-x86_64-linux: - torch-build: << pipeline.parameters.torch-build >> - torchvision-build: << pipeline.parameters.torchvision-build >> - torch-build-index: << pipeline.parameters.torch-build-index >> - trt-version-short: << pipeline.parameters.trt-version-short >> - trt-version-long: << pipeline.parameters.trt-version-long >> - cudnn-version: << pipeline.parameters.cudnn-version >> - python-version: << pipeline.parameters.python-version >> - requires: - - package-x86_64-linux - - - test-py-ts-x86_64-linux: - torch-build: << pipeline.parameters.torch-build >> - torchvision-build: << pipeline.parameters.torchvision-build >> - torch-build-index: << pipeline.parameters.torch-build-index >> - trt-version-long: << pipeline.parameters.trt-version-long >> - python-version: << pipeline.parameters.python-version >> - requires: - - package-x86_64-linux - - - test-py-fx-x86_64-linux: - torch-build: << pipeline.parameters.torch-build >> - torchvision-build: << pipeline.parameters.torchvision-build >> - torch-build-index: << pipeline.parameters.torch-build-index >> - trt-version-long: << pipeline.parameters.trt-version-long >> - python-version: << pipeline.parameters.python-version >> - requires: - - package-x86_64-linux - - - test-py-dynamo-x86_64-linux: - torch-build: << pipeline.parameters.torch-build >> - torchvision-build: << pipeline.parameters.torchvision-build >> - torch-build-index: << pipeline.parameters.torch-build-index >> - trt-version-long: << pipeline.parameters.trt-version-long >> - python-version: << pipeline.parameters.python-version >> - requires: - - package-x86_64-linux - - on-push: - jobs: - - build-x86_64-linux: - cuda-version: << pipeline.parameters.cuda-version >> - torch-build: << pipeline.parameters.torch-build >> - torchvision-build: << pipeline.parameters.torchvision-build >> - torch-build-index: << pipeline.parameters.torch-build-index >> - python-version: << pipeline.parameters.python-version >> - - - test-core-cpp-x86_64-linux: - torch-build: << pipeline.parameters.torch-build >> - torchvision-build: << pipeline.parameters.torchvision-build >> - torch-build-index: << pipeline.parameters.torch-build-index >> - trt-version-short: << pipeline.parameters.trt-version-short >> - trt-version-long: << pipeline.parameters.trt-version-long >> - python-version: << pipeline.parameters.python-version >> - cudnn-version: << pipeline.parameters.cudnn-version >> - requires: - - build-x86_64-linux - - - test-py-ts-x86_64-linux: - torch-build: << pipeline.parameters.torch-build >> - torchvision-build: << pipeline.parameters.torchvision-build >> - torch-build-index: << pipeline.parameters.torch-build-index >> - trt-version-long: << pipeline.parameters.trt-version-long >> - python-version: << pipeline.parameters.python-version >> - requires: - - build-x86_64-linux - - - test-py-fx-x86_64-linux: - torch-build: << pipeline.parameters.torch-build >> - torchvision-build: << pipeline.parameters.torchvision-build >> - torch-build-index: << pipeline.parameters.torch-build-index >> - trt-version-long: << pipeline.parameters.trt-version-long >> - python-version: << pipeline.parameters.python-version >> - requires: - - build-x86_64-linux - - - test-py-dynamo-x86_64-linux: - torch-build: << pipeline.parameters.torch-build >> - torchvision-build: << pipeline.parameters.torchvision-build >> - torch-build-index: << pipeline.parameters.torch-build-index >> - trt-version-long: << pipeline.parameters.trt-version-long >> - python-version: << pipeline.parameters.python-version >> - requires: - - build-x86_64-linux - - - build-x86_64-linux-cmake: - cuda-version: << pipeline.parameters.cuda-version >> - torch-build: << pipeline.parameters.torch-build >> - torch-build-index: << pipeline.parameters.torch-build-index >> - trt-version-short: << pipeline.parameters.trt-version-short >> - cudnn-version: << pipeline.parameters.cudnn-version >> - python-version: << pipeline.parameters.python-version >>