Change GPU Runners (#7840)
* Move to new GPU Runners for slow tests * Move to new GPU Runners for nightly tests
This commit is contained in:
committed by
GitHub
parent
44ba90caff
commit
ce97d7e19b
@@ -112,7 +112,7 @@ jobs:
|
|||||||
|
|
||||||
run_nightly_tests_for_other_torch_modules:
|
run_nightly_tests_for_other_torch_modules:
|
||||||
name: Torch Non-Pipelines CUDA Nightly Tests
|
name: Torch Non-Pipelines CUDA Nightly Tests
|
||||||
runs-on: docker-gpu
|
runs-on: [single-gpu, nvidia-gpu, t4, ci]
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-pytorch-cuda
|
image: diffusers/diffusers-pytorch-cuda
|
||||||
options: --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/ --gpus 0
|
options: --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/ --gpus 0
|
||||||
@@ -185,7 +185,7 @@ jobs:
|
|||||||
|
|
||||||
run_lora_nightly_tests:
|
run_lora_nightly_tests:
|
||||||
name: Nightly LoRA Tests with PEFT and TORCH
|
name: Nightly LoRA Tests with PEFT and TORCH
|
||||||
runs-on: docker-gpu
|
runs-on: [single-gpu, nvidia-gpu, t4, ci]
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-pytorch-cuda
|
image: diffusers/diffusers-pytorch-cuda
|
||||||
options: --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/ --gpus 0
|
options: --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/ --gpus 0
|
||||||
@@ -298,7 +298,7 @@ jobs:
|
|||||||
|
|
||||||
run_nightly_onnx_tests:
|
run_nightly_onnx_tests:
|
||||||
name: Nightly ONNXRuntime CUDA tests on Ubuntu
|
name: Nightly ONNXRuntime CUDA tests on Ubuntu
|
||||||
runs-on: docker-gpu
|
runs-on: [single-gpu, nvidia-gpu, t4, ci]
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-onnxruntime-cuda
|
image: diffusers/diffusers-onnxruntime-cuda
|
||||||
options: --gpus 0 --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/
|
options: --gpus 0 --shm-size "16gb" --ipc host -v /mnt/hf_cache:/mnt/cache/
|
||||||
|
|||||||
@@ -116,7 +116,7 @@ jobs:
|
|||||||
|
|
||||||
torch_cuda_tests:
|
torch_cuda_tests:
|
||||||
name: Torch CUDA Tests
|
name: Torch CUDA Tests
|
||||||
runs-on: docker-gpu
|
runs-on: [single-gpu, nvidia-gpu, t4, ci]
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-pytorch-cuda
|
image: diffusers/diffusers-pytorch-cuda
|
||||||
options: --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/ --gpus 0
|
options: --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/ --gpus 0
|
||||||
@@ -168,7 +168,7 @@ jobs:
|
|||||||
|
|
||||||
peft_cuda_tests:
|
peft_cuda_tests:
|
||||||
name: PEFT CUDA Tests
|
name: PEFT CUDA Tests
|
||||||
runs-on: docker-gpu
|
runs-on: [single-gpu, nvidia-gpu, t4, ci]
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-pytorch-cuda
|
image: diffusers/diffusers-pytorch-cuda
|
||||||
options: --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/ --gpus 0
|
options: --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/ --gpus 0
|
||||||
@@ -265,7 +265,7 @@ jobs:
|
|||||||
|
|
||||||
onnx_cuda_tests:
|
onnx_cuda_tests:
|
||||||
name: ONNX CUDA Tests
|
name: ONNX CUDA Tests
|
||||||
runs-on: docker-gpu
|
runs-on: [single-gpu, nvidia-gpu, t4, ci]
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-onnxruntime-cuda
|
image: diffusers/diffusers-onnxruntime-cuda
|
||||||
options: --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/ --gpus 0
|
options: --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/ --gpus 0
|
||||||
@@ -313,7 +313,7 @@ jobs:
|
|||||||
run_torch_compile_tests:
|
run_torch_compile_tests:
|
||||||
name: PyTorch Compile CUDA tests
|
name: PyTorch Compile CUDA tests
|
||||||
|
|
||||||
runs-on: docker-gpu
|
runs-on: [single-gpu, nvidia-gpu, t4, ci]
|
||||||
|
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-pytorch-compile-cuda
|
image: diffusers/diffusers-pytorch-compile-cuda
|
||||||
@@ -354,7 +354,7 @@ jobs:
|
|||||||
run_xformers_tests:
|
run_xformers_tests:
|
||||||
name: PyTorch xformers CUDA tests
|
name: PyTorch xformers CUDA tests
|
||||||
|
|
||||||
runs-on: docker-gpu
|
runs-on: [single-gpu, nvidia-gpu, t4, ci]
|
||||||
|
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-pytorch-xformers-cuda
|
image: diffusers/diffusers-pytorch-xformers-cuda
|
||||||
@@ -395,7 +395,7 @@ jobs:
|
|||||||
run_examples_tests:
|
run_examples_tests:
|
||||||
name: Examples PyTorch CUDA tests on Ubuntu
|
name: Examples PyTorch CUDA tests on Ubuntu
|
||||||
|
|
||||||
runs-on: docker-gpu
|
runs-on: [single-gpu, nvidia-gpu, t4, ci]
|
||||||
|
|
||||||
container:
|
container:
|
||||||
image: diffusers/diffusers-pytorch-cuda
|
image: diffusers/diffusers-pytorch-cuda
|
||||||
|
|||||||
Reference in New Issue
Block a user