Skip to content

Commit cf205db

Browse files
committed
Try to make CI more stable.
Signed-off-by: Jonas Yang <joyang@nvidia.com>
1 parent add6fe5 commit cf205db

4 files changed

Lines changed: 23 additions & 2 deletions

File tree

.github/workflows/e2e_ppo_trainer_megatron_sglang.yml

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -131,12 +131,16 @@ jobs:
131131
run: |
132132
ray stop --force
133133
OPTIM_MEMORY_EFFICIENT=True ENGINE=sglang SAVE_FREQ=1 MODEL_ID=deepseek-ai/deepseek-coder-1.3b-instruct bash tests/special_e2e/run_ppo_trainer_megatron.sh
134+
with:
135+
max_attempts: 3
134136
- name: Running GSM8K E2E training tests with 3D parallelism on 8 L20 GPUs with Megatron (DeepSeek)
135137
run: |
136138
ray stop --force
137139
export VLLM_USE_V1=1
138140
ray start --head
139141
ENGINE=sglang MODE=async RESUME_MODE=auto MODEL_ID=deepseek-ai/deepseek-coder-1.3b-instruct TOTAL_TRAIN_STEPS=2 bash tests/special_e2e/run_ppo_trainer_megatron.sh
142+
with:
143+
max_attempts: 3
140144
- name: Profiling GRPO GSM8K E2E training tests with 3D parallelism on 8 L20 GPUs with Megatron (Deepseek)
141145
run: |
142146
ray stop --force
@@ -147,6 +151,8 @@ jobs:
147151
else
148152
echo "[SUCCESS] profile success"
149153
fi
154+
with:
155+
max_attempts: 3
150156
- name: clean up
151157
run: |
152158
rm -rf checkpoints

.github/workflows/e2e_ppo_trainer_megatron_vllm.yml

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -153,11 +153,15 @@ jobs:
153153
ray stop --force
154154
ALL_OFFLOAD=True SAVE_FREQ=1 MODEL_ID=deepseek-ai/deepseek-coder-1.3b-instruct COMMON_PP=4 LORA_RANK=8 COMMON_VPP=null COMMON_CP=1 USE_MBRIDGE=True VANILLA_MBRIDGE=False VALUE_VANILLA_MBRIDGE=False USE_DIST_CKPT=False \
155155
bash tests/special_e2e/run_ppo_trainer_megatron.sh
156+
with:
157+
max_attempts: 3
156158
- name: Running GSM8K E2E training tests with 3D parallelism on 8 L20 GPUs with Megatron, use Megatron-Bridge LoRA e2e to pre-load and save (Deepseek)
157159
run: |
158160
ray stop --force
159161
RESUME_MODE=auto MODEL_ID=deepseek-ai/deepseek-coder-1.3b-instruct TOTAL_TRAIN_STEPS=2 SAVE_FREQ=1 COMMON_PP=4 LORA_RANK=8 COMMON_VPP=null COMMON_CP=1 USE_MBRIDGE=True VANILLA_MBRIDGE=False VALUE_VANILLA_MBRIDGE=False USE_DIST_CKPT=False \
160162
bash tests/special_e2e/run_ppo_trainer_megatron.sh
163+
with:
164+
max_attempts: 3
161165
- name: clean up
162166
run: |
163167
rm -rf checkpoints
@@ -189,11 +193,15 @@ jobs:
189193
run: |
190194
ray stop --force
191195
ALL_OFFLOAD=True VAL_BEFORE_TRAIN=True TEST_FREQ=1 SAVE_FREQ=1 LR_WARMUP_STEPS=1 TOTAL_TRAIN_STEPS=2 MODEL_ID=Qwen/Qwen3-0.6B bash tests/special_e2e/run_ppo_trainer_megatron.sh
196+
with:
197+
max_attempts: 3
192198
- name: Running GSM8K E2E training tests with 3D parallelism on 8 L20 GPUs with FP8 rollout
193199
run: |
194200
ray stop --force
195201
export VLLM_USE_V1=1
196202
ROLLOUT_QUANTIZATION=fp8 TOTAL_TRAIN_STEPS=2 MODEL_ID=Qwen/Qwen3-0.6B bash tests/special_e2e/run_ppo_trainer_megatron.sh
203+
with:
204+
max_attempts: 3
197205
- name: clean up
198206
run: |
199207
rm -rf checkpoints

.github/workflows/gpu_unit_tests.yml

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -88,6 +88,7 @@ jobs:
8888
mode: "create"
8989
faas-url: "${{ env.DYNAMIC_RUNNER_ENDPOINT }}"
9090
mlp-image: "${{ env.IMAGE }}"
91+
max_attempts: 3
9192

9293
gpu_unit_tests:
9394
if: github.repository_owner == 'volcengine'
@@ -105,6 +106,8 @@ jobs:
105106
fetch-depth: 0
106107
- name: Install the current repository
107108
run: |
109+
pip3 list | grep cupy || true
110+
pip3 list | grep cupy | xargs pip3 uninstall -y || true
108111
pip3 install hf_transfer
109112
pip3 install -r requirements-test.txt
110113
pip3 install --no-deps -e .
@@ -114,6 +117,8 @@ jobs:
114117
- name: Run all GPU unit tests
115118
run: |
116119
pytest -s -x --ignore-glob="*test_special_*.py" --ignore-glob='*on_cpu.py' --ignore-glob="*test_vllm*" --ignore-glob="*_sglang*" --ignore-glob="*_hf_rollout*" --ignore-glob="tests/models/" --ignore-glob='tests/special*' --ignore-glob="tests/experimental" --ignore-glob="tests/workers/reward_model" tests/
120+
with:
121+
max_attempts: 3
117122
- name: Testing LinearCrossEntropyTP Correctness, Computation Time and Memory Consumption
118123
run: |
119124
LOW_MEMORY=True torchrun --standalone --nnodes=1 --nproc-per-node=8 tests/utils/test_special_linear_cross_entropy_tp.py

.github/workflows/sgl.yml

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -113,7 +113,7 @@ jobs:
113113
fetch-depth: 0
114114
- name: Install the current repository
115115
run: |
116-
pip3 install hf_transfer fastmcp pytest-asyncio
116+
pip3 install hf_transfer fastmcp pytest-asyncio pytest-retry
117117
pip3 install -r requirements-test.txt
118118
pip3 install --no-deps -e .
119119
- name: Prepare gsm8k dataset
@@ -122,7 +122,9 @@ jobs:
122122
python3 examples/data_preprocess/gsm8k.py --local_dataset_path ${HOME}/models/hf_data/gsm8k
123123
- name: Test the latest SGLang Rollout async with agent loop
124124
run: |
125-
ROLLOUT_NAME=sglang pytest -svvv tests/experimental/agent_loop
125+
ROLLOUT_NAME=sglang pytest -svvv tests/experimental/agent_loop --retries 3 --retry-delay 5
126+
with:
127+
max_attempts: 3
126128

127129
cleanup:
128130
runs-on: ubuntu-latest

0 commit comments

Comments
 (0)