44
44
env :
45
45
HOST_PIP_CACHE_DIR : /nvme/github-actions/pip-cache
46
46
HOST_LOCALTIME : /usr/share/zoneinfo/Asia/Shanghai
47
- OUTPUT_FOLDER : cuda11.8_dist_ ${{ github.run_id }}
47
+ OUTPUT_FOLDER : cuda12.4_dist_ ${{ github.run_id }}
48
48
ACTIONS_ALLOW_USE_UNSECURE_NODE_VERSION : true
49
49
REPORT_DIR : /nvme/qa_test_models/test-reports/${{ github.run_id }}
50
50
COV_PARAM : --cov /opt/py3/lib/python3.10/site-packages/lmdeploy
64
64
env :
65
65
PYTHON_VERSION : ${{ matrix.pyver }}
66
66
PLAT_NAME : manylinux2014_x86_64
67
- DOCKER_TAG : cuda11.8
67
+ DOCKER_TAG : cuda12.4
68
68
steps :
69
69
- name : Checkout repository
70
70
uses : actions/checkout@v3
96
96
runs-on : [self-hosted, linux-a100]
97
97
timeout-minutes : 50
98
98
container :
99
- image : openmmlab/lmdeploy:latest-cu12.8
99
+ image : openmmlab/lmdeploy:latest-cu12
100
100
options : " --gpus=all --ipc=host --user root -e PIP_CACHE_DIR=/root/.cache/pip -e NVIDIA_DISABLE_REQUIRE=1 --pull never"
101
101
volumes :
102
102
- /nvme/qa_test_models:/nvme/qa_test_models
@@ -136,7 +136,7 @@ jobs:
136
136
MODELSCOPE_CACHE : /nvme/qa_test_models/modelscope_hub
137
137
MODELSCOPE_MODULES_CACHE : /nvme/qa_test_models/modelscope_modules
138
138
container :
139
- image : openmmlab/lmdeploy:latest-cu12.8
139
+ image : openmmlab/lmdeploy:latest-cu12
140
140
options : " --gpus=all --ipc=host --user root -e PIP_CACHE_DIR=/root/.cache/pip -e NVIDIA_DISABLE_REQUIRE=1 --pull never"
141
141
volumes :
142
142
- /nvme/github-actions/pip-cache:/root/.cache/pip
@@ -168,7 +168,7 @@ jobs:
168
168
rm -rf allure-results
169
169
# remove tmp log in testcase
170
170
rm -rf /nvme/qa_test_models/autotest_model/log/*
171
- mkdir ${{env.REPORT_DIR}}/.pytest_cache -p
171
+ mkdir ${{env.REPORT_DIR}}/.pytest_cache -p && rm autotest/.pytest_cache -f
172
172
ln -s ${{env.REPORT_DIR}}/.pytest_cache autotest
173
173
- name : Test lmdeploy - quantization w4a16
174
174
continue-on-error : true
@@ -219,7 +219,7 @@ jobs:
219
219
MODELSCOPE_CACHE : /nvme/qa_test_models/modelscope_hub
220
220
MODELSCOPE_MODULES_CACHE : /nvme/qa_test_models/modelscope_modules
221
221
container :
222
- image : openmmlab/lmdeploy:latest-cu12.8
222
+ image : openmmlab/lmdeploy:latest-cu12
223
223
options : " --gpus=all --ipc=host --user root -e PIP_CACHE_DIR=/root/.cache/pip -e NVIDIA_DISABLE_REQUIRE=1 --pull never"
224
224
volumes :
225
225
- /nvme/github-actions/pip-cache:/root/.cache/pip
@@ -251,7 +251,7 @@ jobs:
251
251
rm -rf allure-results
252
252
# remove tmp log in testcase
253
253
rm -rf /nvme/qa_test_models/autotest_model/log/*
254
- mkdir ${{env.REPORT_DIR}}/.pytest_cache -p
254
+ mkdir ${{env.REPORT_DIR}}/.pytest_cache -p && rm autotest/.pytest_cache -f
255
255
ln -s ${{env.REPORT_DIR}}/.pytest_cache autotest
256
256
- name : Test lmdeploy - chat
257
257
continue-on-error : true
@@ -324,7 +324,7 @@ jobs:
324
324
model : Intern-S1
325
325
timeout-minutes : 60
326
326
container :
327
- image : openmmlab/lmdeploy:latest-cu12.8
327
+ image : openmmlab/lmdeploy:latest-cu12
328
328
options : " --gpus=all --ipc=host --user root -e PIP_CACHE_DIR=/root/.cache/pip -e NVIDIA_DISABLE_REQUIRE=1 --pull never"
329
329
volumes :
330
330
- /nvme/github-actions/pip-cache:/root/.cache/pip
@@ -352,7 +352,7 @@ jobs:
352
352
rm -rf allure-results
353
353
# remove tmp log in testcase
354
354
rm -rf /nvme/qa_test_models/autotest_model/log/*
355
- mkdir ${{env.REPORT_DIR}}/.pytest_cache -p
355
+ mkdir ${{env.REPORT_DIR}}/.pytest_cache -p && rm autotest/.pytest_cache -f
356
356
ln -s ${{env.REPORT_DIR}}/.pytest_cache autotest
357
357
- name : Start restful api
358
358
if : matrix.model != 'internlm2_5-20b'
@@ -408,7 +408,7 @@ jobs:
408
408
needs : test_quantization
409
409
timeout-minutes : 120
410
410
container :
411
- image : openmmlab/lmdeploy:latest-cu12.8
411
+ image : openmmlab/lmdeploy:latest-cu12
412
412
options : " --gpus=all --ipc=host --user root -e PIP_CACHE_DIR=/root/.cache/pip -e NVIDIA_DISABLE_REQUIRE=1 --pull never"
413
413
volumes :
414
414
- /nvme/github-actions/pip-cache:/root/.cache/pip
@@ -436,7 +436,7 @@ jobs:
436
436
rm -rf allure-results
437
437
# remove tmp log in testcase
438
438
rm -rf /nvme/qa_test_models/autotest_model/log/*
439
- mkdir ${{env.REPORT_DIR}}/.pytest_cache -p
439
+ mkdir ${{env.REPORT_DIR}}/.pytest_cache -p && rm autotest/.pytest_cache -f
440
440
ln -s ${{env.REPORT_DIR}}/.pytest_cache autotest
441
441
- name : Test lmdeploy - interface pipeline case
442
442
run : |
@@ -465,7 +465,7 @@ jobs:
465
465
needs : test_quantization
466
466
timeout-minutes : 120
467
467
container :
468
- image : openmmlab/lmdeploy:latest-cu12.8
468
+ image : openmmlab/lmdeploy:latest-cu12
469
469
options : " --gpus=all --ipc=host --user root -e PIP_CACHE_DIR=/root/.cache/pip -e NVIDIA_DISABLE_REQUIRE=1 --pull never"
470
470
volumes :
471
471
- /nvme/github-actions/pip-cache:/root/.cache/pip
@@ -493,7 +493,7 @@ jobs:
493
493
rm -rf allure-results
494
494
# remove tmp log in testcase
495
495
rm -rf /nvme/qa_test_models/autotest_model/log/*
496
- mkdir ${{env.REPORT_DIR}}/.pytest_cache -p
496
+ mkdir ${{env.REPORT_DIR}}/.pytest_cache -p && rm autotest/.pytest_cache -f
497
497
ln -s ${{env.REPORT_DIR}}/.pytest_cache autotest
498
498
- name : Test benchmark script
499
499
run : |
@@ -520,7 +520,7 @@ jobs:
520
520
matrix :
521
521
evaluate_type : ['chat', 'base']
522
522
container :
523
- image : openmmlab/lmdeploy:latest-cu12.8
523
+ image : openmmlab/lmdeploy:latest-cu12
524
524
options : " --gpus=all --ipc=host --user root -e PIP_CACHE_DIR=/root/.cache/pip -e NVIDIA_DISABLE_REQUIRE=1 --pull never"
525
525
volumes :
526
526
- /nvme/github-actions/pip-cache:/root/.cache/pip
@@ -550,8 +550,7 @@ jobs:
550
550
run : |
551
551
git clone --depth=1 https://github.com/open-compass/opencompass.git
552
552
cd opencompass
553
- cp /nvme/qa_test_models/offline_pkg/requirements-oc.txt requirements/runtime.txt
554
- python3 -m pip install -e .
553
+ python3 -m pip install .
555
554
echo "OPENCOMPASS_DIR=$(pwd)" >> $GITHUB_ENV
556
555
- name : Check env
557
556
run : |
@@ -560,7 +559,7 @@ jobs:
560
559
rm -rf allure-results
561
560
# remove tmp log in testcase
562
561
rm -rf /nvme/qa_test_models/autotest_model/log/*
563
- mkdir ${{env.REPORT_DIR}}/.pytest_cache -p
562
+ mkdir ${{env.REPORT_DIR}}/.pytest_cache -p && rm autotest/.pytest_cache -f
564
563
ln -s ${{env.REPORT_DIR}}/.pytest_cache autotest
565
564
- name : Setup paths for evaluation
566
565
run : |
@@ -571,7 +570,7 @@ jobs:
571
570
run : |
572
571
export LMDEPLOY_DIR=$(pwd)
573
572
574
- python3 .github/scripts/action_tools.py evaluate "[turbomind_internlm2_5_7b_chat, pytorch_internlm2_5_7b_chat, turbomind_internlm2_5_7b_chat_batch1, turbomind_internlm2_5_7b_chat_batch1_4bits, turbomind_internlm3_8b_instruct, pytorch_internlm3_8b_instruct, turbomind_internlm2_5_20b_chat, pytorch_internlm2_5_20b_chat, turbomind_qwen1_5_7b_chat, pytorch_qwen1_5_7b_chat, turbomind_llama3_8b_instruct, pytorch_llama3_8b_instruct, turbomind_llama3_1_8b_instruct, pytorch_llama3_1_8b_instruct, turbomind_qwen2_7b_instruct, pytorch_qwen2_7b_instruct, turbomind_qwen2_5_7b_instruct, pytorch_qwen2_5_7b_instruct, turbomind_llama2_7b_chat, pytorch_qwen1_5_moe_2_7b_chat, pytorch_gemma_2_9b_it, pytorch_gemma_2_27b_it]" "[*race_datasets, *gsm8k_datasets, *ifeval_datasets]" /root/evaluation-reports/${{ github.run_id }} chat true
573
+ python3 .github/scripts/action_tools.py evaluate "[turbomind_internlm2_5_7b_chat, pytorch_internlm2_5_7b_chat, turbomind_internlm2_5_7b_chat_batch1, turbomind_internlm2_5_7b_chat_batch1_4bits, turbomind_internlm3_8b_instruct, pytorch_internlm3_8b_instruct, turbomind_internlm2_5_20b_chat, pytorch_internlm2_5_20b_chat, turbomind_qwen1_5_7b_chat, pytorch_qwen1_5_7b_chat, turbomind_llama3_8b_instruct, pytorch_llama3_8b_instruct, turbomind_llama3_1_8b_instruct, pytorch_llama3_1_8b_instruct, turbomind_qwen2_5_7b_instruct, pytorch_qwen2_5_7b_instruct, turbomind_llama2_7b_chat, pytorch_qwen1_5_moe_2_7b_chat, pytorch_gemma_2_9b_it, pytorch_gemma_2_27b_it]" "[*race_datasets, *gsm8k_datasets, *ifeval_datasets]" /root/evaluation-reports/${{ github.run_id }} chat true
575
574
- name : Evaluate base models
576
575
if : matrix.evaluate_type == 'base'
577
576
run : |
@@ -594,7 +593,7 @@ jobs:
594
593
timeout-minutes : 5
595
594
runs-on : [self-hosted, linux-a100]
596
595
container :
597
- image : openmmlab/lmdeploy:latest-cu12.8
596
+ image : openmmlab/lmdeploy:latest-cu12
598
597
options : " --gpus=all --ipc=host --user root -e PIP_CACHE_DIR=/root/.cache/pip -e NVIDIA_DISABLE_REQUIRE=1 --pull never"
599
598
volumes :
600
599
- /nvme/qa_test_models:/nvme/qa_test_models
@@ -619,7 +618,7 @@ jobs:
619
618
needs : [test_tools, test_restful, test_pipeline, test_benchmark]
620
619
timeout-minutes : 5
621
620
container :
622
- image : openmmlab/lmdeploy:latest-cu12.8
621
+ image : openmmlab/lmdeploy:latest-cu12
623
622
options : " --gpus=all --ipc=host --user root -e PIP_CACHE_DIR=/root/.cache/pip -e NVIDIA_DISABLE_REQUIRE=1 --pull never"
624
623
volumes :
625
624
- /nvme/github-actions/pip-cache:/root/.cache/pip
0 commit comments