From ea5154d85e9ce4d2e9f8d7c631806a4c34752a13 Mon Sep 17 00:00:00 2001 From: Yuwen Hu <54161268+Oscilloscope98@users.noreply.github.com> Date: Mon, 21 Oct 2024 10:27:16 +0800 Subject: [PATCH] Further update to Windows dGPU perf test (#12237) --- .github/workflows/llm_performance_tests.yml | 6 ++++++ 1 file changed, 6 insertions(+) diff --git a/.github/workflows/llm_performance_tests.yml b/.github/workflows/llm_performance_tests.yml index 86fbc26d013..00d652fe48f 100644 --- a/.github/workflows/llm_performance_tests.yml +++ b/.github/workflows/llm_performance_tests.yml @@ -1323,6 +1323,9 @@ jobs: - name: Prepare igpu perf test (3072-384 int4+fp16) shell: bash run: | + if [ ${{ matrix.platform }} == "perf-dgpu" ]; then + sed -i "s/- 'meta-llama\/Llama-2-13b-chat-hf'/# - 'meta-llama\/Llama-2-13b-chat-hf'/" python/llm/test/benchmark/igpu-perf/3072-384_int4_fp16.yaml + fi sed -i 's/2048-256/3072-384/g' python/llm/dev/benchmark/all-in-one/run.py sed -i 's/{today}_test5/{today}_test1/g' python/llm/dev/benchmark/all-in-one/run.py sed -i "s/path to your local model hub/$MODEL_HUB_DIR/g" python/llm/test/benchmark/igpu-perf/3072-384_int4_fp16.yaml @@ -1542,6 +1545,9 @@ jobs: - name: Prepare igpu perf test (4096-512 int4+fp16) shell: bash run: | + if [ ${{ matrix.platform }} == "perf-dgpu" ]; then + sed -i "s/- 'meta-llama\/Llama-2-13b-chat-hf'/# - 'meta-llama\/Llama-2-13b-chat-hf'/" python/llm/test/benchmark/igpu-perf/4096-512_int4_fp16.yaml + fi sed -i 's/3072-384/4096-512/g' python/llm/dev/benchmark/all-in-one/run.py if [ ${{ matrix.platform }} == "perf-igpu" ]; then sed -i 's/{today}_test5/{today}_test1/g' python/llm/dev/benchmark/all-in-one/run.py