From a9e3221c2e81c6c837cd3bb7349d1ad020bbdcf7 Mon Sep 17 00:00:00 2001 From: "Chen, Todd" Date: Wed, 8 Apr 2026 21:38:00 -0500 Subject: [PATCH 1/5] Update Qwen3.5 benchmark configurations and scripts for improved performance - Added new config keys for Qwen3.5 BF16 and FP8 benchmarks on MI300X and MI325X. - Updated Docker image to lmsysorg/sglang:v0.5.10rc0-rocm720-mi30x for better compatibility. - Enhanced benchmark scripts with additional parameters for context length and prefill tokens. - Adjusted memory fraction settings and added new flags for server launch to optimize performance. --- .github/configs/amd-master.yaml | 8 ++++---- benchmarks/single_node/qwen3.5_bf16_mi300x.sh | 14 ++++++++++++-- benchmarks/single_node/qwen3.5_bf16_mi325x.sh | 14 ++++++++++++-- benchmarks/single_node/qwen3.5_bf16_mi355x.sh | 16 ++++++++++++++-- benchmarks/single_node/qwen3.5_fp8_mi300x.sh | 14 ++++++++++++-- benchmarks/single_node/qwen3.5_fp8_mi325x.sh | 14 ++++++++++++-- benchmarks/single_node/qwen3.5_fp8_mi355x.sh | 16 ++++++++++++++-- perf-changelog.yaml | 10 ++++++++++ 8 files changed, 90 insertions(+), 16 deletions(-) diff --git a/.github/configs/amd-master.yaml b/.github/configs/amd-master.yaml index 44cc07419..bbb6e3371 100644 --- a/.github/configs/amd-master.yaml +++ b/.github/configs/amd-master.yaml @@ -132,7 +132,7 @@ qwen3.5-bf16-mi355x-sglang: - { tp: 8, conc-start: 4, conc-end: 64 } qwen3.5-bf16-mi300x-sglang: - image: lmsysorg/sglang:v0.5.9-rocm720-mi30x + image: lmsysorg/sglang:v0.5.10rc0-rocm720-mi30x model: Qwen/Qwen3.5-397B-A17B model-prefix: qwen3.5 runner: mi300x @@ -150,7 +150,7 @@ qwen3.5-bf16-mi300x-sglang: - { tp: 8, conc-start: 4, conc-end: 64 } qwen3.5-bf16-mi325x-sglang: - image: lmsysorg/sglang:v0.5.9-rocm720-mi30x + image: lmsysorg/sglang:v0.5.10rc0-rocm720-mi30x model: Qwen/Qwen3.5-397B-A17B model-prefix: qwen3.5 runner: mi325x @@ -168,7 +168,7 @@ qwen3.5-bf16-mi325x-sglang: - { tp: 8, conc-start: 4, conc-end: 64 } qwen3.5-fp8-mi325x-sglang: - image: lmsysorg/sglang:v0.5.9-rocm720-mi30x + image: lmsysorg/sglang:v0.5.10rc0-rocm720-mi30x model: Qwen/Qwen3.5-397B-A17B-FP8 model-prefix: qwen3.5 runner: mi325x @@ -204,7 +204,7 @@ qwen3.5-fp8-mi355x-sglang: - { tp: 8, conc-start: 4, conc-end: 64 } qwen3.5-fp8-mi300x-sglang: - image: lmsysorg/sglang:v0.5.9-rocm720-mi30x + image: lmsysorg/sglang:v0.5.10rc0-rocm720-mi30x model: Qwen/Qwen3.5-397B-A17B-FP8 model-prefix: qwen3.5 runner: mi300x diff --git a/benchmarks/single_node/qwen3.5_bf16_mi300x.sh b/benchmarks/single_node/qwen3.5_bf16_mi300x.sh index 8aca9860a..f9cacdbf9 100755 --- a/benchmarks/single_node/qwen3.5_bf16_mi300x.sh +++ b/benchmarks/single_node/qwen3.5_bf16_mi300x.sh @@ -19,11 +19,14 @@ hf download "$MODEL" SERVER_LOG=/workspace/server.log PORT=${PORT:-8888} +CONTEXT_LENGTH=$((ISL + OSL + 20)) +MAX_PREFILL_TOKENS=32768 EVAL_CONTEXT_ARGS="" if [ "${EVAL_ONLY}" = "true" ]; then setup_eval_context EVAL_CONTEXT_ARGS="--context-length $EVAL_MAX_MODEL_LEN" +else EVAL_CONTEXT_ARGS="--context-length $CONTEXT_LENGTH" fi # Start GPU monitoring (power, temperature, clocks every second) start_gpu_monitor @@ -35,9 +38,16 @@ python3 -m sglang.launch_server \ --host=0.0.0.0 \ --port $PORT \ --tensor-parallel-size $TP \ + --ep-size $EP_SIZE \ + --data-parallel-size 1 \ --trust-remote-code \ - --mem-fraction-static 0.8 \ - --disable-radix-cache $EVAL_CONTEXT_ARGS > $SERVER_LOG 2>&1 & + --tokenizer-worker-num 6 \ + --enable-aiter-allreduce-fusion \ + --cuda-graph-max-bs $CONC \ + --disable-radix-cache \ + --max-prefill-tokens $MAX_PREFILL_TOKENS \ + --scheduler-recv-interval 30 \ + --mem-fraction-static 0.75 $EVAL_CONTEXT_ARGS > $SERVER_LOG 2>&1 & SERVER_PID=$! diff --git a/benchmarks/single_node/qwen3.5_bf16_mi325x.sh b/benchmarks/single_node/qwen3.5_bf16_mi325x.sh index 8aca9860a..f9cacdbf9 100644 --- a/benchmarks/single_node/qwen3.5_bf16_mi325x.sh +++ b/benchmarks/single_node/qwen3.5_bf16_mi325x.sh @@ -19,11 +19,14 @@ hf download "$MODEL" SERVER_LOG=/workspace/server.log PORT=${PORT:-8888} +CONTEXT_LENGTH=$((ISL + OSL + 20)) +MAX_PREFILL_TOKENS=32768 EVAL_CONTEXT_ARGS="" if [ "${EVAL_ONLY}" = "true" ]; then setup_eval_context EVAL_CONTEXT_ARGS="--context-length $EVAL_MAX_MODEL_LEN" +else EVAL_CONTEXT_ARGS="--context-length $CONTEXT_LENGTH" fi # Start GPU monitoring (power, temperature, clocks every second) start_gpu_monitor @@ -35,9 +38,16 @@ python3 -m sglang.launch_server \ --host=0.0.0.0 \ --port $PORT \ --tensor-parallel-size $TP \ + --ep-size $EP_SIZE \ + --data-parallel-size 1 \ --trust-remote-code \ - --mem-fraction-static 0.8 \ - --disable-radix-cache $EVAL_CONTEXT_ARGS > $SERVER_LOG 2>&1 & + --tokenizer-worker-num 6 \ + --enable-aiter-allreduce-fusion \ + --cuda-graph-max-bs $CONC \ + --disable-radix-cache \ + --max-prefill-tokens $MAX_PREFILL_TOKENS \ + --scheduler-recv-interval 30 \ + --mem-fraction-static 0.75 $EVAL_CONTEXT_ARGS > $SERVER_LOG 2>&1 & SERVER_PID=$! diff --git a/benchmarks/single_node/qwen3.5_bf16_mi355x.sh b/benchmarks/single_node/qwen3.5_bf16_mi355x.sh index 701695def..ce82b9a53 100755 --- a/benchmarks/single_node/qwen3.5_bf16_mi355x.sh +++ b/benchmarks/single_node/qwen3.5_bf16_mi355x.sh @@ -9,7 +9,8 @@ check_env_vars \ ISL \ OSL \ RANDOM_RANGE_RATIO \ - RESULT_FILENAME + RESULT_FILENAME \ + EP_SIZE if [[ -n "$SLURM_JOB_ID" ]]; then echo "JOB $SLURM_JOB_ID running on $SLURMD_NODENAME" @@ -19,11 +20,14 @@ hf download "$MODEL" SERVER_LOG=/workspace/server.log PORT=${PORT:-8888} +CONTEXT_LENGTH=$((ISL + OSL + 20)) +MAX_PREFILL_TOKENS=32768 EVAL_CONTEXT_ARGS="" if [ "${EVAL_ONLY}" = "true" ]; then setup_eval_context EVAL_CONTEXT_ARGS="--context-length $EVAL_MAX_MODEL_LEN" +else EVAL_CONTEXT_ARGS="--context-length $CONTEXT_LENGTH" fi # Start GPU monitoring (power, temperature, clocks every second) start_gpu_monitor @@ -34,8 +38,16 @@ python3 -m sglang.launch_server \ --host=0.0.0.0 \ --port $PORT \ --tensor-parallel-size $TP \ + --ep-size $EP_SIZE \ + --data-parallel-size 1 \ --trust-remote-code \ - --mem-fraction-static 0.8 $EVAL_CONTEXT_ARGS > $SERVER_LOG 2>&1 & + --tokenizer-worker-num 6 \ + --enable-aiter-allreduce-fusion \ + --cuda-graph-max-bs $CONC \ + --disable-radix-cache \ + --max-prefill-tokens $MAX_PREFILL_TOKENS \ + --scheduler-recv-interval 30 \ + --mem-fraction-static 0.75 $EVAL_CONTEXT_ARGS > $SERVER_LOG 2>&1 & SERVER_PID=$! diff --git a/benchmarks/single_node/qwen3.5_fp8_mi300x.sh b/benchmarks/single_node/qwen3.5_fp8_mi300x.sh index 00cc9cf91..cda6d11f1 100755 --- a/benchmarks/single_node/qwen3.5_fp8_mi300x.sh +++ b/benchmarks/single_node/qwen3.5_fp8_mi300x.sh @@ -19,11 +19,14 @@ hf download "$MODEL" SERVER_LOG=/workspace/server.log PORT=${PORT:-8888} +CONTEXT_LENGTH=$((ISL + OSL + 20)) +MAX_PREFILL_TOKENS=32768 EVAL_CONTEXT_ARGS="" if [ "${EVAL_ONLY}" = "true" ]; then setup_eval_context EVAL_CONTEXT_ARGS="--context-length $EVAL_MAX_MODEL_LEN" +else EVAL_CONTEXT_ARGS="--context-length $CONTEXT_LENGTH" fi # Start GPU monitoring (power, temperature, clocks every second) start_gpu_monitor @@ -36,9 +39,16 @@ python3 -m sglang.launch_server \ --host=0.0.0.0 \ --port $PORT \ --tensor-parallel-size $TP \ + --ep-size $EP_SIZE \ + --data-parallel-size 1 \ --trust-remote-code \ - --mem-fraction-static 0.8 \ - --disable-radix-cache $EVAL_CONTEXT_ARGS > $SERVER_LOG 2>&1 & + --tokenizer-worker-num 6 \ + --enable-aiter-allreduce-fusion \ + --cuda-graph-max-bs $CONC \ + --disable-radix-cache \ + --max-prefill-tokens $MAX_PREFILL_TOKENS \ + --scheduler-recv-interval 30 \ + --mem-fraction-static 0.75 $EVAL_CONTEXT_ARGS > $SERVER_LOG 2>&1 & SERVER_PID=$! diff --git a/benchmarks/single_node/qwen3.5_fp8_mi325x.sh b/benchmarks/single_node/qwen3.5_fp8_mi325x.sh index 00cc9cf91..cda6d11f1 100755 --- a/benchmarks/single_node/qwen3.5_fp8_mi325x.sh +++ b/benchmarks/single_node/qwen3.5_fp8_mi325x.sh @@ -19,11 +19,14 @@ hf download "$MODEL" SERVER_LOG=/workspace/server.log PORT=${PORT:-8888} +CONTEXT_LENGTH=$((ISL + OSL + 20)) +MAX_PREFILL_TOKENS=32768 EVAL_CONTEXT_ARGS="" if [ "${EVAL_ONLY}" = "true" ]; then setup_eval_context EVAL_CONTEXT_ARGS="--context-length $EVAL_MAX_MODEL_LEN" +else EVAL_CONTEXT_ARGS="--context-length $CONTEXT_LENGTH" fi # Start GPU monitoring (power, temperature, clocks every second) start_gpu_monitor @@ -36,9 +39,16 @@ python3 -m sglang.launch_server \ --host=0.0.0.0 \ --port $PORT \ --tensor-parallel-size $TP \ + --ep-size $EP_SIZE \ + --data-parallel-size 1 \ --trust-remote-code \ - --mem-fraction-static 0.8 \ - --disable-radix-cache $EVAL_CONTEXT_ARGS > $SERVER_LOG 2>&1 & + --tokenizer-worker-num 6 \ + --enable-aiter-allreduce-fusion \ + --cuda-graph-max-bs $CONC \ + --disable-radix-cache \ + --max-prefill-tokens $MAX_PREFILL_TOKENS \ + --scheduler-recv-interval 30 \ + --mem-fraction-static 0.75 $EVAL_CONTEXT_ARGS > $SERVER_LOG 2>&1 & SERVER_PID=$! diff --git a/benchmarks/single_node/qwen3.5_fp8_mi355x.sh b/benchmarks/single_node/qwen3.5_fp8_mi355x.sh index 701695def..ce82b9a53 100644 --- a/benchmarks/single_node/qwen3.5_fp8_mi355x.sh +++ b/benchmarks/single_node/qwen3.5_fp8_mi355x.sh @@ -9,7 +9,8 @@ check_env_vars \ ISL \ OSL \ RANDOM_RANGE_RATIO \ - RESULT_FILENAME + RESULT_FILENAME \ + EP_SIZE if [[ -n "$SLURM_JOB_ID" ]]; then echo "JOB $SLURM_JOB_ID running on $SLURMD_NODENAME" @@ -19,11 +20,14 @@ hf download "$MODEL" SERVER_LOG=/workspace/server.log PORT=${PORT:-8888} +CONTEXT_LENGTH=$((ISL + OSL + 20)) +MAX_PREFILL_TOKENS=32768 EVAL_CONTEXT_ARGS="" if [ "${EVAL_ONLY}" = "true" ]; then setup_eval_context EVAL_CONTEXT_ARGS="--context-length $EVAL_MAX_MODEL_LEN" +else EVAL_CONTEXT_ARGS="--context-length $CONTEXT_LENGTH" fi # Start GPU monitoring (power, temperature, clocks every second) start_gpu_monitor @@ -34,8 +38,16 @@ python3 -m sglang.launch_server \ --host=0.0.0.0 \ --port $PORT \ --tensor-parallel-size $TP \ + --ep-size $EP_SIZE \ + --data-parallel-size 1 \ --trust-remote-code \ - --mem-fraction-static 0.8 $EVAL_CONTEXT_ARGS > $SERVER_LOG 2>&1 & + --tokenizer-worker-num 6 \ + --enable-aiter-allreduce-fusion \ + --cuda-graph-max-bs $CONC \ + --disable-radix-cache \ + --max-prefill-tokens $MAX_PREFILL_TOKENS \ + --scheduler-recv-interval 30 \ + --mem-fraction-static 0.75 $EVAL_CONTEXT_ARGS > $SERVER_LOG 2>&1 & SERVER_PID=$! diff --git a/perf-changelog.yaml b/perf-changelog.yaml index 2d904c002..23a88309e 100644 --- a/perf-changelog.yaml +++ b/perf-changelog.yaml @@ -1307,3 +1307,13 @@ - "Model: nvidia/Qwen3.5-397B-A17B-NVFP4" - "Configs: 1k1k (TP4 conc 4-128), 8k1k (TP4 conc 4-128)" pr-link: https://github.com/SemiAnalysisAI/InferenceX/pull/820 + +- config-keys: + - qwen3.5-bf16-mi300x-sglang + - qwen3.5-bf16-mi325x-sglang + - qwen3.5-fp8-mi300x-sglang + - qwen3.5-fp8-mi325x-sglang + description: + - "Update cli args of Qwen3.5 FP8 and BF16 SGLang benchmarks for MI300X and MI325X to achieve better performance" + - "Use lmsysorg/sglang:v0.5.10rc0-rocm720-mi30x" + pr-link: https://github.com/SemiAnalysisAI/InferenceX/pull/xx From f05c9164905680c7e45aa56231290ef84f062b94 Mon Sep 17 00:00:00 2001 From: "Chen, Todd" Date: Wed, 8 Apr 2026 21:38:15 -0500 Subject: [PATCH 2/5] Update Qwen3.5 benchmark PR link in perf-changelog.yaml --- perf-changelog.yaml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/perf-changelog.yaml b/perf-changelog.yaml index 23a88309e..dc1022e03 100644 --- a/perf-changelog.yaml +++ b/perf-changelog.yaml @@ -1316,4 +1316,4 @@ description: - "Update cli args of Qwen3.5 FP8 and BF16 SGLang benchmarks for MI300X and MI325X to achieve better performance" - "Use lmsysorg/sglang:v0.5.10rc0-rocm720-mi30x" - pr-link: https://github.com/SemiAnalysisAI/InferenceX/pull/xx + pr-link: https://github.com/SemiAnalysisAI/InferenceX/pull/986 From c647a86b9f46cbc5518d9589980e3e010c0cd62b Mon Sep 17 00:00:00 2001 From: "Chen, Todd" Date: Wed, 8 Apr 2026 21:38:15 -0500 Subject: [PATCH 3/5] Remove --ep-size parameter from Qwen3.5 benchmark scripts for BF16 and FP8 configurations on MI300X and MI325X to streamline server launch commands. --- benchmarks/single_node/qwen3.5_bf16_mi300x.sh | 1 - benchmarks/single_node/qwen3.5_bf16_mi325x.sh | 1 - benchmarks/single_node/qwen3.5_fp8_mi300x.sh | 1 - benchmarks/single_node/qwen3.5_fp8_mi325x.sh | 1 - 4 files changed, 4 deletions(-) diff --git a/benchmarks/single_node/qwen3.5_bf16_mi300x.sh b/benchmarks/single_node/qwen3.5_bf16_mi300x.sh index f9cacdbf9..c4d971a41 100755 --- a/benchmarks/single_node/qwen3.5_bf16_mi300x.sh +++ b/benchmarks/single_node/qwen3.5_bf16_mi300x.sh @@ -38,7 +38,6 @@ python3 -m sglang.launch_server \ --host=0.0.0.0 \ --port $PORT \ --tensor-parallel-size $TP \ - --ep-size $EP_SIZE \ --data-parallel-size 1 \ --trust-remote-code \ --tokenizer-worker-num 6 \ diff --git a/benchmarks/single_node/qwen3.5_bf16_mi325x.sh b/benchmarks/single_node/qwen3.5_bf16_mi325x.sh index f9cacdbf9..c4d971a41 100644 --- a/benchmarks/single_node/qwen3.5_bf16_mi325x.sh +++ b/benchmarks/single_node/qwen3.5_bf16_mi325x.sh @@ -38,7 +38,6 @@ python3 -m sglang.launch_server \ --host=0.0.0.0 \ --port $PORT \ --tensor-parallel-size $TP \ - --ep-size $EP_SIZE \ --data-parallel-size 1 \ --trust-remote-code \ --tokenizer-worker-num 6 \ diff --git a/benchmarks/single_node/qwen3.5_fp8_mi300x.sh b/benchmarks/single_node/qwen3.5_fp8_mi300x.sh index cda6d11f1..406631bbb 100755 --- a/benchmarks/single_node/qwen3.5_fp8_mi300x.sh +++ b/benchmarks/single_node/qwen3.5_fp8_mi300x.sh @@ -39,7 +39,6 @@ python3 -m sglang.launch_server \ --host=0.0.0.0 \ --port $PORT \ --tensor-parallel-size $TP \ - --ep-size $EP_SIZE \ --data-parallel-size 1 \ --trust-remote-code \ --tokenizer-worker-num 6 \ diff --git a/benchmarks/single_node/qwen3.5_fp8_mi325x.sh b/benchmarks/single_node/qwen3.5_fp8_mi325x.sh index cda6d11f1..406631bbb 100755 --- a/benchmarks/single_node/qwen3.5_fp8_mi325x.sh +++ b/benchmarks/single_node/qwen3.5_fp8_mi325x.sh @@ -39,7 +39,6 @@ python3 -m sglang.launch_server \ --host=0.0.0.0 \ --port $PORT \ --tensor-parallel-size $TP \ - --ep-size $EP_SIZE \ --data-parallel-size 1 \ --trust-remote-code \ --tokenizer-worker-num 6 \ From 2c1719120152c00aa62812d8129c2e0714f092f6 Mon Sep 17 00:00:00 2001 From: "Chen, Todd" Date: Wed, 8 Apr 2026 21:39:33 -0500 Subject: [PATCH 4/5] Update Qwen3.5 benchmark scripts to use 'aiter' as the attention backend instead of 'triton' for BF16 and FP8 configurations on MI300X and MI325X. --- benchmarks/single_node/qwen3.5_bf16_mi300x.sh | 2 +- benchmarks/single_node/qwen3.5_bf16_mi325x.sh | 2 +- benchmarks/single_node/qwen3.5_fp8_mi300x.sh | 2 +- benchmarks/single_node/qwen3.5_fp8_mi325x.sh | 2 +- 4 files changed, 4 insertions(+), 4 deletions(-) diff --git a/benchmarks/single_node/qwen3.5_bf16_mi300x.sh b/benchmarks/single_node/qwen3.5_bf16_mi300x.sh index c4d971a41..f7c71963d 100755 --- a/benchmarks/single_node/qwen3.5_bf16_mi300x.sh +++ b/benchmarks/single_node/qwen3.5_bf16_mi300x.sh @@ -33,7 +33,7 @@ start_gpu_monitor # following Andy Luo linkedin's recipe https://www.linkedin.com/feed/update/urn:li:activity:7429203734389280768/ python3 -m sglang.launch_server \ - --attention-backend triton \ + --attention-backend aiter \ --model-path $MODEL \ --host=0.0.0.0 \ --port $PORT \ diff --git a/benchmarks/single_node/qwen3.5_bf16_mi325x.sh b/benchmarks/single_node/qwen3.5_bf16_mi325x.sh index c4d971a41..f7c71963d 100644 --- a/benchmarks/single_node/qwen3.5_bf16_mi325x.sh +++ b/benchmarks/single_node/qwen3.5_bf16_mi325x.sh @@ -33,7 +33,7 @@ start_gpu_monitor # following Andy Luo linkedin's recipe https://www.linkedin.com/feed/update/urn:li:activity:7429203734389280768/ python3 -m sglang.launch_server \ - --attention-backend triton \ + --attention-backend aiter \ --model-path $MODEL \ --host=0.0.0.0 \ --port $PORT \ diff --git a/benchmarks/single_node/qwen3.5_fp8_mi300x.sh b/benchmarks/single_node/qwen3.5_fp8_mi300x.sh index 406631bbb..fe761d88d 100755 --- a/benchmarks/single_node/qwen3.5_fp8_mi300x.sh +++ b/benchmarks/single_node/qwen3.5_fp8_mi300x.sh @@ -34,7 +34,7 @@ start_gpu_monitor # following AMD Andy linkedin's recipe # https://www.linkedin.com/feed/update/urn:li:activity:7429203734389280768/ python3 -m sglang.launch_server \ - --attention-backend triton \ + --attention-backend aiter \ --model-path $MODEL \ --host=0.0.0.0 \ --port $PORT \ diff --git a/benchmarks/single_node/qwen3.5_fp8_mi325x.sh b/benchmarks/single_node/qwen3.5_fp8_mi325x.sh index 406631bbb..fe761d88d 100755 --- a/benchmarks/single_node/qwen3.5_fp8_mi325x.sh +++ b/benchmarks/single_node/qwen3.5_fp8_mi325x.sh @@ -34,7 +34,7 @@ start_gpu_monitor # following AMD Andy linkedin's recipe # https://www.linkedin.com/feed/update/urn:li:activity:7429203734389280768/ python3 -m sglang.launch_server \ - --attention-backend triton \ + --attention-backend aiter \ --model-path $MODEL \ --host=0.0.0.0 \ --port $PORT \ From 970742b192f08be93a4b865d010b623787ed4090 Mon Sep 17 00:00:00 2001 From: "Chen, Todd" Date: Wed, 8 Apr 2026 22:01:04 -0500 Subject: [PATCH 5/5] Update Qwen3.5 benchmark configurations to use stable Docker image version for MI300X and MI325X setups. Changed image tag from 'v0.5.10rc0-rocm720-mi30x' to 'v0.5.10-rocm720-mi30x' for consistency and reliability. --- .github/configs/amd-master.yaml | 8 ++++---- perf-changelog.yaml | 2 +- 2 files changed, 5 insertions(+), 5 deletions(-) diff --git a/.github/configs/amd-master.yaml b/.github/configs/amd-master.yaml index bbb6e3371..c20003c64 100644 --- a/.github/configs/amd-master.yaml +++ b/.github/configs/amd-master.yaml @@ -132,7 +132,7 @@ qwen3.5-bf16-mi355x-sglang: - { tp: 8, conc-start: 4, conc-end: 64 } qwen3.5-bf16-mi300x-sglang: - image: lmsysorg/sglang:v0.5.10rc0-rocm720-mi30x + image: lmsysorg/sglang:v0.5.10-rocm720-mi30x model: Qwen/Qwen3.5-397B-A17B model-prefix: qwen3.5 runner: mi300x @@ -150,7 +150,7 @@ qwen3.5-bf16-mi300x-sglang: - { tp: 8, conc-start: 4, conc-end: 64 } qwen3.5-bf16-mi325x-sglang: - image: lmsysorg/sglang:v0.5.10rc0-rocm720-mi30x + image: lmsysorg/sglang:v0.5.10-rocm720-mi30x model: Qwen/Qwen3.5-397B-A17B model-prefix: qwen3.5 runner: mi325x @@ -168,7 +168,7 @@ qwen3.5-bf16-mi325x-sglang: - { tp: 8, conc-start: 4, conc-end: 64 } qwen3.5-fp8-mi325x-sglang: - image: lmsysorg/sglang:v0.5.10rc0-rocm720-mi30x + image: lmsysorg/sglang:v0.5.10-rocm720-mi30x model: Qwen/Qwen3.5-397B-A17B-FP8 model-prefix: qwen3.5 runner: mi325x @@ -204,7 +204,7 @@ qwen3.5-fp8-mi355x-sglang: - { tp: 8, conc-start: 4, conc-end: 64 } qwen3.5-fp8-mi300x-sglang: - image: lmsysorg/sglang:v0.5.10rc0-rocm720-mi30x + image: lmsysorg/sglang:v0.5.10-rocm720-mi30x model: Qwen/Qwen3.5-397B-A17B-FP8 model-prefix: qwen3.5 runner: mi300x diff --git a/perf-changelog.yaml b/perf-changelog.yaml index dc1022e03..c773f5b97 100644 --- a/perf-changelog.yaml +++ b/perf-changelog.yaml @@ -1315,5 +1315,5 @@ - qwen3.5-fp8-mi325x-sglang description: - "Update cli args of Qwen3.5 FP8 and BF16 SGLang benchmarks for MI300X and MI325X to achieve better performance" - - "Use lmsysorg/sglang:v0.5.10rc0-rocm720-mi30x" + - "Use lmsysorg/sglang:v0.5.10-rocm720-mi30x" pr-link: https://github.com/SemiAnalysisAI/InferenceX/pull/986