diff --git a/.github/configs/nvidia-master.yaml b/.github/configs/nvidia-master.yaml index e876b8af6..fbcf05444 100644 --- a/.github/configs/nvidia-master.yaml +++ b/.github/configs/nvidia-master.yaml @@ -1571,6 +1571,448 @@ dsr1-fp8-h200-dynamo-trt: ep: 8 dp-attn: true +dsr1-fp8-h100-dynamo-trt: + image: nvcr.io/nvidia/ai-dynamo/tensorrtllm-runtime:0.8.1.post3 + model: deepseek-ai/DeepSeek-R1-0528 + model-prefix: DeepSeek-R1-0528 + runner: h100-multinode-slurm + precision: fp8 + framework: dynamo-trt + multinode: true + disagg: true + seq-len-configs: + - isl: 1024 + osl: 1024 + search-space: + # MTP configurations + - spec-decoding: "mtp" + conc-list: [6] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/1k1k/mtp/ctx1_gen3_tep16_batch1_eplb0_mtp3.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/1k1k/mtp/ctx1_gen3_tep16_batch1_eplb0_mtp3.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: false + - spec-decoding: "mtp" + conc-list: [9] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/1k1k/mtp/ctx1_gen3_tep16_batch2_eplb0_mtp3.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/1k1k/mtp/ctx1_gen3_tep16_batch2_eplb0_mtp3.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: false + - spec-decoding: "mtp" + conc-list: [30] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/1k1k/mtp/ctx1_gen3_tep16_batch8_eplb0_mtp3.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/1k1k/mtp/ctx1_gen3_tep16_batch8_eplb0_mtp3.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: false + - spec-decoding: "mtp" + conc-list: [60] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/1k1k/mtp/ctx1_gen3_tep16_batch16_eplb0_mtp3.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/1k1k/mtp/ctx1_gen3_tep16_batch16_eplb0_mtp3.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: false + - spec-decoding: "mtp" + conc-list: [117] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/1k1k/mtp/ctx1_gen3_tep16_batch32_eplb0_mtp3.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/1k1k/mtp/ctx1_gen3_tep16_batch32_eplb0_mtp3.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: false + - spec-decoding: "mtp" + conc-list: [231] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/1k1k/mtp/ctx1_gen3_dep16_batch4_eplb0_mtp3.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/1k1k/mtp/ctx1_gen3_dep16_batch4_eplb0_mtp3.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: true + - spec-decoding: "mtp" + conc-list: [462] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/1k1k/mtp/ctx1_gen3_tep16_batch128_eplb0_mtp3.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/1k1k/mtp/ctx1_gen3_tep16_batch128_eplb0_mtp3.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: false + - spec-decoding: "mtp" + conc-list: [615] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/1k1k/mtp/ctx1_gen1_dep16_batch32_eplb0_mtp2.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/1k1k/mtp/ctx1_gen1_dep16_batch32_eplb0_mtp2.yaml" + decode: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + - spec-decoding: "mtp" + conc-list: [1229] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/1k1k/mtp/ctx1_gen1_dep16_batch64_eplb0_mtp1.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/1k1k/mtp/ctx1_gen1_dep16_batch64_eplb0_mtp1.yaml" + decode: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + # Non-MTP configurations (STP) + - conc-list: [6] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/1k1k/stp/ctx1_gen3_tep16_batch1_eplb0_mtp0.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/1k1k/stp/ctx1_gen3_tep16_batch1_eplb0_mtp0.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: false + - conc-list: [9] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/1k1k/stp/ctx1_gen3_tep16_batch2_eplb0_mtp0.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/1k1k/stp/ctx1_gen3_tep16_batch2_eplb0_mtp0.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: false + - conc-list: [30] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/1k1k/stp/ctx1_gen3_tep16_batch8_eplb0_mtp0.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/1k1k/stp/ctx1_gen3_tep16_batch8_eplb0_mtp0.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: false + - conc-list: [60] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/1k1k/stp/ctx1_gen3_tep16_batch16_eplb0_mtp0.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/1k1k/stp/ctx1_gen3_tep16_batch16_eplb0_mtp0.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: false + - conc-list: [231] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/1k1k/stp/ctx1_gen3_dep16_batch4_eplb0_mtp0.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/1k1k/stp/ctx1_gen3_dep16_batch4_eplb0_mtp0.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: true + - conc-list: [462] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/1k1k/stp/ctx1_gen3_dep16_batch8_eplb0_mtp0.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/1k1k/stp/ctx1_gen3_dep16_batch8_eplb0_mtp0.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: true + - conc-list: [924] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/1k1k/stp/ctx1_gen3_dep16_batch16_eplb0_mtp0.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/1k1k/stp/ctx1_gen3_dep16_batch16_eplb0_mtp0.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: true + - conc-list: [1845] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/1k1k/stp/ctx1_gen3_dep16_batch32_eplb0_mtp0.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/1k1k/stp/ctx1_gen3_dep16_batch32_eplb0_mtp0.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: true + - conc-list: [4916] + prefill: + num-worker: 2 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/1k1k/stp/ctx2_gen1_dep16_batch256_eplb0_mtp0.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/1k1k/stp/ctx2_gen1_dep16_batch256_eplb0_mtp0.yaml" + decode: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + - isl: 8192 + osl: 1024 + search-space: + # MTP configurations (6 points) + - spec-decoding: "mtp" + conc-list: [6] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/8k1k/mtp/ctx1_gen3_tep16_batch1_eplb0_mtp3.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/8k1k/mtp/ctx1_gen3_tep16_batch1_eplb0_mtp3.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: false + - spec-decoding: "mtp" + conc-list: [9] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/8k1k/mtp/ctx1_gen3_tep16_batch2_eplb0_mtp3.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/8k1k/mtp/ctx1_gen3_tep16_batch2_eplb0_mtp3.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: false + - spec-decoding: "mtp" + conc-list: [30] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/8k1k/mtp/ctx1_gen3_tep16_batch8_eplb0_mtp3.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/8k1k/mtp/ctx1_gen3_tep16_batch8_eplb0_mtp3.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: false + - spec-decoding: "mtp" + conc-list: [77] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/8k1k/mtp/ctx1_gen1_dep16_batch4_eplb0_mtp3.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/8k1k/mtp/ctx1_gen1_dep16_batch4_eplb0_mtp3.yaml" + decode: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + - spec-decoding: "mtp" + conc-list: [78] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/8k1k/mtp/ctx1_gen2_tep16_batch32_eplb0_mtp3.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/8k1k/mtp/ctx1_gen2_tep16_batch32_eplb0_mtp3.yaml" + decode: + num-worker: 2 + tp: 16 + ep: 16 + dp-attn: false + - spec-decoding: "mtp" + conc-list: [154] + prefill: + num-worker: 2 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/8k1k/mtp/ctx2_gen1_dep16_batch8_eplb0_mtp3.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/8k1k/mtp/ctx2_gen1_dep16_batch8_eplb0_mtp3.yaml" + decode: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + # STP configurations (5 points) + - conc-list: [6] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/8k1k/stp/ctx1_gen3_tep16_batch1_eplb0_mtp0.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/8k1k/stp/ctx1_gen3_tep16_batch1_eplb0_mtp0.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: false + - conc-list: [9] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/8k1k/stp/ctx1_gen3_tep16_batch2_eplb0_mtp0.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/8k1k/stp/ctx1_gen3_tep16_batch2_eplb0_mtp0.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: false + - conc-list: [30] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/8k1k/stp/ctx1_gen3_tep16_batch8_eplb0_mtp0.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/8k1k/stp/ctx1_gen3_tep16_batch8_eplb0_mtp0.yaml" + decode: + num-worker: 3 + tp: 16 + ep: 16 + dp-attn: false + - conc-list: [154] + prefill: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/8k1k/stp/ctx1_gen2_tep16_batch64_eplb0_mtp0.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/8k1k/stp/ctx1_gen2_tep16_batch64_eplb0_mtp0.yaml" + decode: + num-worker: 2 + tp: 16 + ep: 16 + dp-attn: false + - conc-list: [308] + prefill: + num-worker: 2 + tp: 16 + ep: 16 + dp-attn: true + additional-settings: + # https://github.com/ishandhanani/srt-slurm/blob/sa-submission-q1-2026/recipes/trtllm/h100-fp8/8k1k/stp/ctx2_gen1_dep16_batch16_eplb0_mtp0.yaml + - "CONFIG_FILE=recipes/trtllm/h100-fp8/8k1k/stp/ctx2_gen1_dep16_batch16_eplb0_mtp0.yaml" + decode: + num-worker: 1 + tp: 16 + ep: 16 + dp-attn: true + gptoss-fp4-b200-trt: image: nvcr.io#nvidia/tensorrt-llm/release:1.2.0rc2.post2 model: openai/gpt-oss-120b diff --git a/.github/configs/runners.yaml b/.github/configs/runners.yaml index 86ef27315..eb9602ecb 100644 --- a/.github/configs/runners.yaml +++ b/.github/configs/runners.yaml @@ -66,3 +66,5 @@ b300: - 'b300-nv_0' gb300: - 'gb300-nv_0' +h100-multinode-slurm: +- 'h100-dgxc-slurm_0' diff --git a/perf-changelog.yaml b/perf-changelog.yaml index f41d685d3..34f1091c6 100644 --- a/perf-changelog.yaml +++ b/perf-changelog.yaml @@ -400,14 +400,14 @@ - "Update GPT-OSS FP4 B200 TRT pareto configurations and new container image" - "Extend maximum concurrency to 256 across all sequence lengths" pr-link: https://github.com/InferenceMAX/InferenceMAX/pull/639 - + - config-keys: - dsr1-fp8-mi355x-sglang-disagg description: - "Add --use-chat-template argument to benchmark_serving script" - "Without this arg, MTP acceptance rates are artificially high for DeepSeek with MTP" pr-link: https://github.com/InferenceMAX/InferenceMAX/pull/647 - + - config-keys: - dsr1-fp8-b200-sglang-mtp description: @@ -417,7 +417,7 @@ - "Update launch_b200-dgxc.sh to support SPEC_SUFFIX for MTP script selection" - "Configurations: TP=8, EP=1, concurrency 4-64 for 1k1k, 1k8k, and 8k1k sequence lengths" pr-link: https://github.com/InferenceMAX/InferenceMAX/pull/626 - + - config-keys: - dsr1-fp4-b200-trt-mtp description: @@ -439,3 +439,9 @@ description: - "Add DeepSeek R1 FP8 GB300 Dynamo TRT-LLM disaggregated multinode configurations for 1k8k" pr-link: https://github.com/InferenceMAX/InferenceMAX/pull/654 + +- config-keys: + - dsr1-fp8-h100-dynamo-trt + description: + - "Add DeepSeek R1 FP8 H100 Dynamo TRT-LLM disaggregated multinode configurations" + pr-link: https://github.com/InferenceMAX/InferenceMAX/pull/651 diff --git a/runners/launch_h100-dgxc-slurm.sh b/runners/launch_h100-dgxc-slurm.sh new file mode 100644 index 000000000..fb0ae69f7 --- /dev/null +++ b/runners/launch_h100-dgxc-slurm.sh @@ -0,0 +1,170 @@ +#!/usr/bin/bash + +set -x + +echo "Cloning srt-slurm repository..." +SRT_REPO_DIR="srt-slurm" +if [ -d "$SRT_REPO_DIR" ]; then + echo "Removing existing $SRT_REPO_DIR..." + rm -rf "$SRT_REPO_DIR" +fi + +git clone https://github.com/ishandhanani/srt-slurm.git "$SRT_REPO_DIR" +cd "$SRT_REPO_DIR" +git checkout sa-submission-q1-2026 + +echo "Installing srtctl..." +curl -LsSf https://astral.sh/uv/install.sh | sh +source $HOME/.local/bin/env + +uv venv +source .venv/bin/activate +uv pip install -e . + +if ! command -v srtctl &> /dev/null; then + echo "Error: Failed to install srtctl" + exit 1 +fi + +echo "Configs available at: $SRT_REPO_DIR/" + +export SLURM_PARTITION="hpc-gpu-1" +export SLURM_ACCOUNT="customer" + +# Convert IMAGE to srt-slurm format (nvcr.io/ -> nvcr.io#) +CONTAINER_KEY=$(echo "$IMAGE" | sed 's|nvcr.io/|nvcr.io#|') + +# Map container image to local squash file +SQUASH_FILE="/mnt/nfs/sa-shared/containers/$(echo "$IMAGE" | sed 's|nvcr.io/||' | sed 's/[\/:@#]/+/g').sqsh" + +if [[ $MODEL_PREFIX == "DeepSeek-R1-0528" ]]; then + export MODEL_PATH="/mnt/numa1/shared/models/dsr1-fp8" + export SERVED_MODEL_NAME="DeepSeek-R1-0528" +else + echo "Unsupported model prefix: $MODEL_PREFIX. Supported prefixes are: DeepSeek-R1-0528" + exit 1 +fi + +export ISL="$ISL" +export OSL="$OSL" + +# Create srtslurm.yaml for srtctl +echo "Creating srtslurm.yaml configuration..." +cat > srtslurm.yaml <&1) +echo "$SRTCTL_OUTPUT" + +# Extract JOB_ID from srtctl output +JOB_ID=$(echo "$SRTCTL_OUTPUT" | grep -oP '✅ Job \K[0-9]+' || echo "$SRTCTL_OUTPUT" | grep -oP 'Job \K[0-9]+') + +if [ -z "$JOB_ID" ]; then + echo "Error: Failed to extract JOB_ID from srtctl output" + exit 1 +fi + +echo "Extracted JOB_ID: $JOB_ID" + +# Wait for this specific job to complete +echo "Waiting for job $JOB_ID to complete..." +while [ -n "$(squeue -j $JOB_ID --noheader 2>/dev/null)" ]; do + echo "Job $JOB_ID still running..." + squeue -j $JOB_ID + sleep 30 +done +echo "Job $JOB_ID completed!" + +echo "Collecting results..." + +# Use the JOB_ID to find the logs directory +# srtctl creates logs in outputs/JOB_ID/logs/ +LOGS_DIR="outputs/$JOB_ID/logs" + +if [ ! -d "$LOGS_DIR" ]; then + echo "Warning: Logs directory not found at $LOGS_DIR" + exit 1 +fi + +echo "Found logs directory: $LOGS_DIR" + +cat $LOGS_DIR/sweep_${JOB_ID}.log + +for file in $LOGS_DIR/*; do + if [ -f "$file" ]; then + tail -n 500 $file + fi +done + +# Find all result subdirectories +RESULT_SUBDIRS=$(find "$LOGS_DIR" -maxdepth 1 -type d -name "*isl*osl*" 2>/dev/null) + +if [ -z "$RESULT_SUBDIRS" ]; then + echo "Warning: No result subdirectories found in $LOGS_DIR" +else + # Process results from all configurations + for result_subdir in $RESULT_SUBDIRS; do + echo "Processing result subdirectory: $result_subdir" + + # Extract configuration info from directory name + CONFIG_NAME=$(basename "$result_subdir") + + # Find all result JSON files + RESULT_FILES=$(find "$result_subdir" -name "results_concurrency_*.json" 2>/dev/null) + + for result_file in $RESULT_FILES; do + if [ -f "$result_file" ]; then + # Extract metadata from filename + # Files are of the format "results_concurrency_gpus_{num gpus}_ctx_{num ctx}_gen_{num gen}.json" + filename=$(basename "$result_file") + concurrency=$(echo "$filename" | sed -n 's/results_concurrency_\([0-9]*\)_gpus_.*/\1/p') + gpus=$(echo "$filename" | sed -n 's/results_concurrency_[0-9]*_gpus_\([0-9]*\)_ctx_.*/\1/p') + ctx=$(echo "$filename" | sed -n 's/.*_ctx_\([0-9]*\)_gen_.*/\1/p') + gen=$(echo "$filename" | sed -n 's/.*_gen_\([0-9]*\)\.json/\1/p') + + echo "Processing concurrency $concurrency with $gpus GPUs (ctx: $ctx, gen: $gen): $result_file" + + WORKSPACE_RESULT_FILE="$GITHUB_WORKSPACE/${RESULT_FILENAME}_${CONFIG_NAME}_conc${concurrency}_gpus_${gpus}_ctx_${ctx}_gen_${gen}.json" + cp "$result_file" "$WORKSPACE_RESULT_FILE" + + echo "Copied result file to: $WORKSPACE_RESULT_FILE" + fi + done + done +fi + +echo "All result files processed" + +# Cleanup +echo "Cleaning up..." +deactivate 2>/dev/null || true +rm -rf .venv +echo "Cleanup complete"