@@ -41,13 +41,9 @@ concurrency:
41
41
42
42
jobs :
43
43
lint :
44
- # Only trigger lint on pull request
45
- if : ${{ github.event_name == 'pull_request' }}
46
44
uses : ./.github/workflows/pre-commit.yml
47
45
48
46
changes :
49
- # Only trigger changes on pull request
50
- if : ${{ github.event_name == 'pull_request' }}
51
47
runs-on : ubuntu-latest
52
48
permissions :
53
49
pull-requests : read
@@ -60,20 +56,24 @@ jobs:
60
56
with :
61
57
filters : |
62
58
e2e_tracker:
59
+ - '.github/workflows/vllm_ascend_test.yaml'
63
60
- 'vllm_ascend/**'
64
61
- 'csrc/**'
65
62
- 'cmake/**'
66
63
- 'tests/e2e/**'
67
- - 'tests/conftest.py'
68
- - 'tests/model_utils.py'
69
- - 'tests/utils.py'
64
+ - 'CMakeLists.txt'
65
+ - 'setup.py'
66
+ - 'requirements.txt'
67
+ - 'requirements-dev.txt'
68
+ - 'requirements-lint.txt'
69
+ - 'packages.txt'
70
70
ut_tracker:
71
71
- 'tests/ut/**'
72
72
ut :
73
73
needs : [lint, changes]
74
74
name : unit test
75
- # only trigger unit test after lint passed and the change is e2e and ut related. Or the PR is merged.
76
- if : ${{ github.event_name == 'push' || ( needs.lint.result == 'success' && (needs.changes.outputs.e2e_tracker == 'true' || needs.changes.outputs.ut_tracker == 'true') ) }}
75
+ # only trigger unit test after lint passed and the change is e2e and ut related.
76
+ if : ${{ needs.lint.result == 'success' && (needs.changes.outputs.e2e_tracker == 'true' || needs.changes.outputs.ut_tracker == 'true') }}
77
77
runs-on : ubuntu-latest
78
78
container :
79
79
image : quay.io/ascend/cann:8.1.rc1-910b-ubuntu22.04-py3.10
@@ -112,9 +112,8 @@ jobs:
112
112
python3 -m pip install -r requirements-dev.txt --extra-index https://download.pytorch.org/whl/cpu/
113
113
python3 -m pip install -v . --extra-index https://download.pytorch.org/whl/cpu/
114
114
115
- - name : Run unit test for V1 Engine
115
+ - name : Run unit test
116
116
env :
117
- VLLM_USE_V1 : 1
118
117
VLLM_WORKER_MULTIPROC_METHOD : spawn
119
118
TORCH_DEVICE_BACKEND_AUTOLOAD : 0
120
119
run : |
@@ -189,9 +188,8 @@ jobs:
189
188
pip install -r requirements-dev.txt
190
189
pip install -v -e .
191
190
192
- - name : Run e2e test for V1 Engine
191
+ - name : Run e2e test
193
192
env :
194
- VLLM_USE_V1 : 1
195
193
VLLM_WORKER_MULTIPROC_METHOD : spawn
196
194
VLLM_USE_MODELSCOPE : True
197
195
run : |
@@ -213,26 +211,6 @@ jobs:
213
211
# TODO: revert me when test_v1_spec_decode.py::test_ngram_correctness is fixed
214
212
VLLM_USE_MODELSCOPE=True pytest -sv tests/e2e/singlecard/spec_decode_v1/test_v1_spec_decode.py
215
213
216
- - name : Run e2e test on V0 engine
217
- if : ${{ github.event_name == 'schedule' }}
218
- env :
219
- VLLM_USE_V1 : 0
220
- VLLM_USE_MODELSCOPE : True
221
- run : |
222
- pytest -sv tests/e2e/singlecard/test_offline_inference.py
223
- pytest -sv tests/e2e/singlecard/test_ilama_lora.py
224
- pytest -sv tests/e2e/singlecard/test_guided_decoding.py
225
- pytest -sv tests/e2e/singlecard/test_camem.py
226
- pytest -sv tests/e2e/singlecard/test_prompt_embedding.py
227
- pytest -sv tests/e2e/singlecard/test_embedding.py
228
- pytest -sv tests/e2e/singlecard/ \
229
- --ignore=tests/e2e/singlecard/test_offline_inference.py \
230
- --ignore=tests/e2e/singlecard/test_ilama_lora.py \
231
- --ignore=tests/e2e/singlecard/test_guided_decoding.py \
232
- --ignore=tests/e2e/singlecard/test_camem.py \
233
- --ignore=tests/e2e/singlecard/test_prompt_embedding.py \
234
- --ignore=tests/e2e/singlecard/test_embedding.py
235
-
236
214
e2e-4-cards :
237
215
needs : [e2e]
238
216
if : ${{ needs.e2e.result == 'success' }}
@@ -290,9 +268,8 @@ jobs:
290
268
pip install -r requirements-dev.txt
291
269
pip install -v -e .
292
270
293
- - name : Run vllm-project/vllm-ascend test for V1 Engine
271
+ - name : Run vllm-project/vllm-ascend test
294
272
env :
295
- VLLM_USE_V1 : 1
296
273
VLLM_WORKER_MULTIPROC_METHOD : spawn
297
274
VLLM_USE_MODELSCOPE : True
298
275
run : |
@@ -308,19 +285,3 @@ jobs:
308
285
pytest -sv tests/e2e/multicard/ --ignore=tests/e2e/multicard/test_ilama_lora_tp2.py \
309
286
--ignore=tests/e2e/multicard/test_offline_inference_distributed.py \
310
287
--ignore=tests/e2e/multicard/test_data_parallel.py
311
-
312
- - name : Run vllm-project/vllm-ascend test on V0 engine
313
- if : ${{ github.event_name == 'schedule' }}
314
- env :
315
- VLLM_USE_V1 : 0
316
- VLLM_USE_MODELSCOPE : True
317
- run : |
318
- pytest -sv tests/e2e/multicard/test_ilama_lora_tp2.py
319
- # Fixme: run VLLM_USE_MODELSCOPE=True pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py will raise error.
320
- # To avoid oom, we need to run the test in a single process.
321
- pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_QwQ
322
- pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_DeepSeek_W8A8
323
- pytest -sv tests/e2e/multicard/test_data_parallel.py
324
- pytest -sv tests/e2e/multicard/ --ignore=tests/e2e/multicard/test_ilama_lora_tp2.py \
325
- --ignore=tests/e2e/multicard/test_offline_inference_distributed.py \
326
- --ignore=tests/e2e/multicard/test_data_parallel.py
0 commit comments