From da9270be2d5fbfb93b67fef51bc19917d88a3424 Mon Sep 17 00:00:00 2001 From: Yuwen Hu <54161268+Oscilloscope98@users.noreply.github.com> Date: Fri, 18 Oct 2024 23:20:17 +0800 Subject: [PATCH] Further update to Windows dGPU perf test (#12233) --- .github/workflows/llm_performance_tests.yml | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/.github/workflows/llm_performance_tests.yml b/.github/workflows/llm_performance_tests.yml index ba97ed38502..86fbc26d013 100644 --- a/.github/workflows/llm_performance_tests.yml +++ b/.github/workflows/llm_performance_tests.yml @@ -1133,6 +1133,9 @@ jobs: - name: Prepare igpu perf test (2048-256 int4+fp16) shell: bash run: | + if [ ${{ matrix.platform }} == "perf-dgpu" ]; then + sed -i "s/- 'baichuan-inc\/Baichuan2-13B-Chat'/# - 'baichuan-inc\/Baichuan2-13B-Chat'/" python/llm/test/benchmark/igpu-perf/2048-256_int4_fp16.yaml + fi sed -i 's/1024-128/2048-256/g' python/llm/dev/benchmark/all-in-one/run.py sed -i 's/{today}_test5/{today}_test1/g' python/llm/dev/benchmark/all-in-one/run.py sed -i "s/path to your local model hub/$MODEL_HUB_DIR/g" python/llm/test/benchmark/igpu-perf/2048-256_int4_fp16.yaml @@ -1351,7 +1354,6 @@ jobs: call conda deactivate - # Remove Qwen-VL-Chat on dGPU for 3072-384 tests - name: Prepare igpu perf test for transformers 4.36 (3072-384 int4+fp16) if: ${{ matrix.platform == 'perf-igpu' }} shell: bash