[ci] Move sm86 from trunk to pull (#93085)
Experiment on capacity
Pull Request resolved: https://github.com/pytorch/pytorch/pull/93085
Approved by: https://github.com/malfet, https://github.com/huydhn, https://github.com/ZainRizvi
diff --git a/.github/workflows/pull.yml b/.github/workflows/pull.yml
index 2ef204f..2996436 100644
--- a/.github/workflows/pull.yml
+++ b/.github/workflows/pull.yml
@@ -333,3 +333,30 @@
{ config: "default", shard: 1, num_shards: 2, runner: "linux.rocm.gpu" },
{ config: "default", shard: 2, num_shards: 2, runner: "linux.rocm.gpu" },
]}
+
+ linux-bionic-cuda11_6-py3_10-gcc7-sm86-build:
+ name: linux-bionic-cuda11.6-py3.10-gcc7-sm86
+ uses: ./.github/workflows/_linux-build.yml
+ with:
+ build-environment: linux-bionic-cuda11.6-py3.10-gcc7-sm86
+ docker-image-name: pytorch-linux-bionic-cuda11.6-cudnn8-py3-gcc7
+ cuda-arch-list: 8.6
+ test-matrix: |
+ { include: [
+ { config: "default", shard: 1, num_shards: 4, runner: "linux.g5.4xlarge.nvidia.gpu" },
+ { config: "default", shard: 2, num_shards: 4, runner: "linux.g5.4xlarge.nvidia.gpu" },
+ { config: "default", shard: 3, num_shards: 4, runner: "linux.g5.4xlarge.nvidia.gpu" },
+ { config: "default", shard: 4, num_shards: 4, runner: "linux.g5.4xlarge.nvidia.gpu" },
+ { config: "slow", shard: 1, num_shards: 2, runner: "linux.g5.4xlarge.nvidia.gpu" },
+ { config: "slow", shard: 2, num_shards: 2, runner: "linux.g5.4xlarge.nvidia.gpu" },
+ { config: "functorch", shard: 1, num_shards: 1, runner: "linux.g5.4xlarge.nvidia.gpu" },
+ ]}
+
+ linux-bionic-cuda11_6-py3_10-gcc7-sm86-test:
+ name: linux-bionic-cuda11.6-py3.10-gcc7-sm86
+ uses: ./.github/workflows/_linux-test.yml
+ needs: linux-bionic-cuda11_6-py3_10-gcc7-sm86-build
+ with:
+ build-environment: linux-bionic-cuda11.6-py3.10-gcc7-sm86
+ docker-image: ${{ needs.linux-bionic-cuda11_6-py3_10-gcc7-sm86-build.outputs.docker-image }}
+ test-matrix: ${{ needs.linux-bionic-cuda11_6-py3_10-gcc7-sm86-build.outputs.test-matrix }}
diff --git a/.github/workflows/trunk.yml b/.github/workflows/trunk.yml
index 6e2b7d1..518afdd 100644
--- a/.github/workflows/trunk.yml
+++ b/.github/workflows/trunk.yml
@@ -86,33 +86,6 @@
docker-image: ${{ needs.linux-bionic-cuda11_8-py3_10-gcc7-build.outputs.docker-image }}
test-matrix: ${{ needs.linux-bionic-cuda11_8-py3_10-gcc7-build.outputs.test-matrix }}
- linux-bionic-cuda11_6-py3_10-gcc7-sm86-build:
- name: linux-bionic-cuda11.6-py3.10-gcc7-sm86
- uses: ./.github/workflows/_linux-build.yml
- with:
- build-environment: linux-bionic-cuda11.6-py3.10-gcc7-sm86
- docker-image-name: pytorch-linux-bionic-cuda11.6-cudnn8-py3-gcc7
- cuda-arch-list: 8.6
- test-matrix: |
- { include: [
- { config: "default", shard: 1, num_shards: 4, runner: "linux.g5.4xlarge.nvidia.gpu" },
- { config: "default", shard: 2, num_shards: 4, runner: "linux.g5.4xlarge.nvidia.gpu" },
- { config: "default", shard: 3, num_shards: 4, runner: "linux.g5.4xlarge.nvidia.gpu" },
- { config: "default", shard: 4, num_shards: 4, runner: "linux.g5.4xlarge.nvidia.gpu" },
- { config: "slow", shard: 1, num_shards: 2, runner: "linux.g5.4xlarge.nvidia.gpu" },
- { config: "slow", shard: 2, num_shards: 2, runner: "linux.g5.4xlarge.nvidia.gpu" },
- { config: "functorch", shard: 1, num_shards: 1, runner: "linux.g5.4xlarge.nvidia.gpu" },
- ]}
-
- linux-bionic-cuda11_6-py3_10-gcc7-sm86-test:
- name: linux-bionic-cuda11.6-py3.10-gcc7-sm86
- uses: ./.github/workflows/_linux-test.yml
- needs: linux-bionic-cuda11_6-py3_10-gcc7-sm86-build
- with:
- build-environment: linux-bionic-cuda11.6-py3.10-gcc7-sm86
- docker-image: ${{ needs.linux-bionic-cuda11_6-py3_10-gcc7-sm86-build.outputs.docker-image }}
- test-matrix: ${{ needs.linux-bionic-cuda11_6-py3_10-gcc7-sm86-build.outputs.test-matrix }}
-
libtorch-linux-bionic-cuda11_6-py3_7-gcc7-build:
name: libtorch-linux-bionic-cuda11.6-py3.7-gcc7
uses: ./.github/workflows/_linux-build.yml