From 9f1a329f90ad5b844fbf6e3c08db46684a4a2e1b Mon Sep 17 00:00:00 2001 From: ashors1 Date: Thu, 26 Jun 2025 17:22:51 -0700 Subject: [PATCH] add dynamic_batching key to sft openmathinstruct config Signed-off-by: ashors1 --- examples/configs/sft_openmathinstruct2.yaml | 3 +++ 1 file changed, 3 insertions(+) diff --git a/examples/configs/sft_openmathinstruct2.yaml b/examples/configs/sft_openmathinstruct2.yaml index e934f7aa29..2040bdd5ff 100644 --- a/examples/configs/sft_openmathinstruct2.yaml +++ b/examples/configs/sft_openmathinstruct2.yaml @@ -37,6 +37,9 @@ policy: context_parallel_size: 1 custom_parallel_plan: null + dynamic_batching: + enabled: false + # makes the training sequence length divisible by the tensor parallel size # this is useful for sequence parallel training make_sequence_length_divisible_by: ${policy.dtensor_cfg.tensor_parallel_size}