|
38 | 38 | # Build and test ExecuTorch with the add model on portable backend.
|
39 | 39 | PYTHON_EXECUTABLE=python bash .ci/scripts/test_model.sh "add" "${BUILD_TOOL}" "portable"
|
40 | 40 |
|
41 |
| - test-models-linux: |
42 |
| - name: test-models-linux |
| 41 | + test-models-linux-basic: |
| 42 | + name: test-models-linux-basic |
43 | 43 | uses: pytorch/test-infra/.github/workflows/linux_job_v2.yml@main
|
44 | 44 | permissions:
|
45 | 45 | id-token: write
|
|
70 | 70 | # Build and test ExecuTorch
|
71 | 71 | PYTHON_EXECUTABLE=python bash .ci/scripts/test_model.sh "${MODEL_NAME}" "${BUILD_TOOL}" "${BACKEND}"
|
72 | 72 |
|
| 73 | + test-models-linux: |
| 74 | + name: test-models-linux |
| 75 | + uses: pytorch/test-infra/.github/workflows/linux_job_v2.yml@main |
| 76 | + permissions: |
| 77 | + id-token: write |
| 78 | + contents: read |
| 79 | + strategy: |
| 80 | + matrix: |
| 81 | + model: [linear, add, add_mul, ic3, mv2, resnet18, resnet50, mobilebert, emformer_transcribe] |
| 82 | + backend: [portable, xnnpack-quantization-delegation] |
| 83 | + runner: [linux.2xlarge] |
| 84 | + include: |
| 85 | + - model: ic4 |
| 86 | + backend: portable |
| 87 | + runner: linux.4xlarge.memory |
| 88 | + - model: ic4 |
| 89 | + backend: xnnpack-quantization-delegation |
| 90 | + runner: linux.4xlarge.memory |
| 91 | + - model: emformer_join |
| 92 | + backend: portable |
| 93 | + runner: linux.4xlarge.memory |
| 94 | + - model: emformer_join |
| 95 | + backend: xnnpack-quantization-delegation |
| 96 | + runner: linux.4xlarge.memory |
| 97 | + - model: phi-4-mini |
| 98 | + backend: portable |
| 99 | + runner: linux.4xlarge.memory |
| 100 | + - model: llama3_2_vision_encoder |
| 101 | + backend: portable |
| 102 | + runner: linux.4xlarge.memory |
| 103 | + - model: w2l |
| 104 | + backend: portable |
| 105 | + runner: linux.4xlarge.memory |
| 106 | + fail-fast: false |
| 107 | + with: |
| 108 | + runner: ${{ matrix.runner }} |
| 109 | + docker-image: executorch-ubuntu-22.04-clang12 |
| 110 | + submodules: 'true' |
| 111 | + ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }} |
| 112 | + timeout: 90 |
| 113 | + script: | |
| 114 | + # The generic Linux job chooses to use base env, not the one setup by the image |
| 115 | + CONDA_ENV=$(conda env list --json | jq -r ".envs | .[-1]") |
| 116 | + conda activate "${CONDA_ENV}" |
| 117 | +
|
| 118 | + MODEL_NAME=${{ matrix.model }} |
| 119 | + BUILD_TOOL=cmake |
| 120 | + BACKEND=${{ matrix.backend }} |
| 121 | +
|
| 122 | + PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh --build-tool "${BUILD_TOOL}" |
| 123 | + # Build and test ExecuTorch |
| 124 | + PYTHON_EXECUTABLE=python bash .ci/scripts/test_model.sh "${MODEL_NAME}" "${BUILD_TOOL}" "${BACKEND}" |
| 125 | +
|
73 | 126 | test-llama-runner-linux:
|
74 | 127 | name: test-llama-runner-linux
|
75 | 128 | uses: pytorch/test-infra/.github/workflows/linux_job_v2.yml@main
|
|
0 commit comments