name: Self-hosted runner (AMD mi250 scheduled CI caller) on: workflow_run: workflows: ["Self-hosted runner (AMD scheduled CI caller)"] branches: ["main"] types: [completed] push: branches: - run_amd_scheduled_ci_caller* jobs: model-ci: name: Model CI uses: huggingface/hf-workflows/.github/workflows/transformers_amd_ci_scheduled.yaml@main with: job: run_models_gpu slack_report_channel: "#transformers-ci-daily-amd" runner: mi250 docker: huggingface/transformers-pytorch-amd-gpu ci_event: Scheduled CI (AMD) - mi250 secrets: inherit torch-pipeline: name: Torch pipeline CI uses: huggingface/hf-workflows/.github/workflows/transformers_amd_ci_scheduled.yaml@main with: job: run_pipelines_torch_gpu slack_report_channel: "#transformers-ci-daily-amd" runner: mi250 docker: huggingface/transformers-pytorch-amd-gpu ci_event: Scheduled CI (AMD) - mi250 secrets: inherit example-ci: name: Example CI uses: huggingface/hf-workflows/.github/workflows/transformers_amd_ci_scheduled.yaml@main with: job: run_examples_gpu slack_report_channel: "#transformers-ci-daily-amd" runner: mi250 docker: huggingface/transformers-pytorch-amd-gpu ci_event: Scheduled CI (AMD) - mi250 secrets: inherit deepspeed-ci: name: DeepSpeed CI uses: huggingface/hf-workflows/.github/workflows/transformers_amd_ci_scheduled.yaml@main with: job: run_torch_cuda_extensions_gpu slack_report_channel: "#transformers-ci-daily-amd" runner: mi250 docker: huggingface/transformers-pytorch-deepspeed-amd-gpu ci_event: Scheduled CI (AMD) - mi250 secrets: inherit