5353 type : choice
5454 options :
5555 - all
56- - Qwen/Qwen2.5-7B-Instruct
5756 - Qwen/Qwen2.5-VL-7B-Instruct
5857 - Qwen/Qwen3-8B-Base
58+ - Qwen/Qwen3-30B-A3B
5959 default : ' all'
6060
6161# Bash shells do not use ~/.profile or ~/.bashrc so these shells need to be explicitly
@@ -77,58 +77,57 @@ jobs:
7777 ${{
7878 (contains(github.event.pull_request.labels.*.name, 'accuracy-test') ||
7979 contains(github.event.pull_request.labels.*.name, 'vl-accuracy-test') ||
80+ contains(github.event.pull_request.labels.*.name, 'moe-accuracy-test') ||
8081 contains(github.event.pull_request.labels.*.name, 'dense-accuracy-test')) &&
8182 contains(github.event.pull_request.labels.*.name, 'ready-for-test') ||
8283 github.event_name == 'workflow_dispatch' || github.event_name == 'schedule'
8384 }}
8485 runs-on : >-
8586 ${{
86- (matrix.model_name == 'Qwen/Qwen2.5-VL-7B-Instruct ' && 'linux-arm64-npu-4') ||
87+ (matrix.model_name == 'Qwen/Qwen3-30B-A3B ' && 'linux-arm64-npu-4') ||
8788 'linux-arm64-npu-2'
8889 }}
8990 strategy :
9091 matrix :
91- vllm_use_version : [0, 1]
92+ vllm_use_version : [1]
9293 # the accuracy test will run:
9394 # 1. workflow_dispatch with models input
94- # - all: Qwen/Qwen2.5-7B-Instruct , Qwen/Qwen2.5-VL-7B-Instruct, Qwen/Qwen3-8B-Base
95- # - specified but not all: Qwen/Qwen2.5-7B-Instruct , Qwen/Qwen2.5-VL-7B-Instruct, Qwen/Qwen3-8B-Base
95+ # - all: Qwen/Qwen3-30B-A3B , Qwen/Qwen2.5-VL-7B-Instruct, Qwen/Qwen3-8B-Base
96+ # - specified but not all: Qwen/Qwen3-30B-A3B , Qwen/Qwen2.5-VL-7B-Instruct, Qwen/Qwen3-8B-Base
9697 # 2. PR labeled with "*-accuracy-test"
97- # - accuracy-test: Qwen/Qwen2.5-7B-Instruct , Qwen/Qwen2.5-VL-7B-Instruct
98- # - dense-accuracy-test: Qwen/Qwen2.5-7B-Instruct
98+ # - accuracy-test: Qwen/Qwen3-8B-Base , Qwen/Qwen2.5-VL-7B-Instruct, Qwen/Qwen3-30B-A3B
99+ # - dense-accuracy-test: Qwen/Qwen3-8B-Base
99100 # - vl-accuracy-test: Qwen/Qwen2.5-VL-7B-Instruct
101+ # - moe-accuracy-test: Qwen/Qwen3-30B-A3B
100102 model_name : ${{ fromJSON(
101103 (github.event_name == 'schedule' &&
102- ' ["Qwen/Qwen2.5-7B-Instruct ","Qwen/Qwen2.5-VL-7B-Instruct","Qwen/Qwen3-8B-Base"]' ) ||
104+ ' ["Qwen/Qwen3-30B-A3B ","Qwen/Qwen2.5-VL-7B-Instruct","Qwen/Qwen3-8B-Base"]' ) ||
103105 (github.event.inputs.models == 'all' &&
104- ' ["Qwen/Qwen2.5-7B-Instruct ","Qwen/Qwen2.5-VL-7B-Instruct","Qwen/Qwen3-8B-Base"]' ) ||
105- (github.event.inputs.models == 'Qwen/Qwen2.5-7B-Instruct ' &&
106- ' ["Qwen/Qwen2.5-7B-Instruct "]' ) ||
106+ ' ["Qwen/Qwen3-30B-A3B ","Qwen/Qwen2.5-VL-7B-Instruct","Qwen/Qwen3-8B-Base"]' ) ||
107+ (github.event.inputs.models == 'Qwen/Qwen3-30B-A3B ' &&
108+ ' ["Qwen/Qwen3-30B-A3B "]' ) ||
107109 (github.event.inputs.models == 'Qwen/Qwen2.5-VL-7B-Instruct' &&
108110 ' ["Qwen/Qwen2.5-VL-7B-Instruct"]' ) ||
109111 (github.event.inputs.models == 'Qwen/Qwen3-8B-Base' &&
110112 ' ["Qwen/Qwen3-8B-Base"]' ) ||
111113 contains(github.event.pull_request.labels.*.name, 'accuracy-test') &&
112- ' ["Qwen/Qwen3-8B-Base","Qwen/Qwen2.5-VL-7B-Instruct"]' ||
114+ ' ["Qwen/Qwen3-8B-Base","Qwen/Qwen2.5-VL-7B-Instruct", "Qwen/Qwen3-30B-A3B" ]' ||
113115 contains(github.event.pull_request.labels.*.name, 'dense-accuracy-test') &&
114116 ' ["Qwen/Qwen3-8B-Base"]' ||
115117 contains(github.event.pull_request.labels.*.name, 'vl-accuracy-test') &&
116- ' ["Qwen/Qwen2.5-VL-7B-Instruct"]'
118+ ' ["Qwen/Qwen2.5-VL-7B-Instruct"]' ||
119+ contains(github.event.pull_request.labels.*.name, 'moe-accuracy-test') &&
120+ ' ["Qwen/Qwen3-30B-A3B"]'
117121 ) }}
118- # Remove exclude after https://github.com/vllm-project/vllm-ascend/issues/1044 resolved
119- exclude :
120- - model_name : Qwen/Qwen2.5-VL-7B-Instruct
121- vllm_use_version : 1
122122
123123 fail-fast : false
124124 name : ${{ matrix.model_name }} accuracy V${{ matrix.vllm_use_version }}
125125 container :
126126 image : m.daocloud.io/quay.io/ascend/cann:8.1.rc1-910b-ubuntu22.04-py3.10
127127 env :
128- HF_ENDPOINT : https://hf-mirror.com
129- HF_TOKEN : ${{ secrets.HF_TOKEN }}
130128 DATASET_SOURCE : ModelScope
131129 VLLM_USE_MODELSCOPE : True
130+ USE_MODELSCOPE_HUB : 1
132131 # 1. If version specified (work_dispatch), do specified branch accuracy test
133132 # 2. If no version (labeled PR), do accuracy test by default ref:
134133 # The branch, tag or SHA to checkout. When checking out the repository that
@@ -188,23 +187,19 @@ jobs:
188187 - name : Get vLLM commit hash and URL
189188 working-directory : ./vllm-empty
190189 run : |
191- VLLM_COMMIT=$(git rev-parse HEAD)
190+ VLLM_COMMIT=$(git rev-parse --short=7 HEAD)
192191 echo "VLLM_COMMIT=$VLLM_COMMIT" >> $GITHUB_ENV
193- echo "VLLM_COMMIT_URL=https://github.com/vllm-project/vllm/commit/$VLLM_COMMIT" >> $GITHUB_ENV
194192
195193 - name : Get vLLM-Ascend commit hash and URL
196194 working-directory : ./vllm-ascend
197195 run : |
198- VLLM_ASCEND_COMMIT=$(git rev-parse HEAD)
196+ VLLM_ASCEND_COMMIT=$(git rev-parse --short=7 HEAD)
199197 echo "VLLM_ASCEND_COMMIT=$VLLM_ASCEND_COMMIT" >> $GITHUB_ENV
200- echo "VLLM_ASCEND_COMMIT_URL=https://github.com/vllm-project/vllm-ascend/commit/$VLLM_ASCEND_COMMIT" >> $GITHUB_ENV
201198
202- - name : Print resolved hashes and URLs
199+ - name : Print resolved hashes
203200 run : |
204201 echo "vLLM : ${{ env.VLLM_COMMIT }}"
205- echo "vLLM link : ${{ env.VLLM_COMMIT_URL }}"
206202 echo "vLLM-Ascend: ${{ env.VLLM_ASCEND_COMMIT }}"
207- echo "Ascend link: ${{ env.VLLM_ASCEND_COMMIT_URL }}"
208203
209204 - name : Install lm-eval, ray, and datasets
210205 run : |
@@ -263,8 +258,6 @@ jobs:
263258 --vllm_version "${{ env.GHA_VLLM_VERSION }}" \
264259 --vllm_commit "${{ env.VLLM_COMMIT }}" \
265260 --vllm_ascend_commit "${{ env.VLLM_ASCEND_COMMIT }}" \
266- --vllm_commit_url "${{ env.VLLM_COMMIT_URL }}" \
267- --vllm_ascend_commit_url "${{ env.VLLM_ASCEND_COMMIT_URL }}" \
268261 --vllm_use_v1 "$VLLM_USE_V1"
269262
270263 - name : Generate step summary
@@ -373,7 +366,7 @@ jobs:
373366 git push -f origin "${{ env.BRANCH_NAME }}"
374367
375368 - name : Create PR in upstream via API
376- uses : actions/github-script@v6
369+ uses : actions/github-script@v7
377370 with :
378371 github-token : ${{ secrets.PAT_TOKEN }}
379372 script : |
@@ -386,7 +379,7 @@ jobs:
386379 body: `The accuracy results running on NPU Altlas A2 have changed, updating reports for:
387380 ${{
388381 github.event.inputs.models == 'all'
389- && 'All models (Qwen2.5-7B-Instruct , Qwen2.5-VL-7B-Instruct, Qwen3-8B-Base)'
382+ && 'All models (Qwen/Qwen3-30B-A3B , Qwen2.5-VL-7B-Instruct, Qwen3-8B-Base)'
390383 || github.event.inputs.models
391384 }}
392385
0 commit comments