From 085d429ddb13d00a05bc3d0ddb77a6332db6aa16 Mon Sep 17 00:00:00 2001 From: flybird11111 <1829166702@qq.com> Date: Tue, 4 Jun 2024 07:09:22 +0000 Subject: [PATCH] fix --- colossalai/shardformer/modeling/llama.py | 1 - 1 file changed, 1 deletion(-) diff --git a/colossalai/shardformer/modeling/llama.py b/colossalai/shardformer/modeling/llama.py index 8a6a7cf17e08..528419654784 100644 --- a/colossalai/shardformer/modeling/llama.py +++ b/colossalai/shardformer/modeling/llama.py @@ -493,7 +493,6 @@ def forward( if sp_mode in ["split_gather", "ring"]: q_len *= sp_size - assert q_len % 4 == 0, "Flash Attention Error: The sequence length should be a multiple of 4." query_states = self.q_proj(hidden_states) key_states = self.k_proj(hidden_states)