From 44599221346476a2d393e9bf23e964514a6c3892 Mon Sep 17 00:00:00 2001 From: "Peter St. John" Date: Fri, 19 Sep 2025 06:58:07 -0700 Subject: [PATCH] add accepts_loss_kwargs=False to EsmPreTrainedModel Signed-off-by: Peter St. John --- src/transformers/models/esm/modeling_esm.py | 1 + 1 file changed, 1 insertion(+) diff --git a/src/transformers/models/esm/modeling_esm.py b/src/transformers/models/esm/modeling_esm.py index 5db366aa6197..9c88548508c4 100755 --- a/src/transformers/models/esm/modeling_esm.py +++ b/src/transformers/models/esm/modeling_esm.py @@ -591,6 +591,7 @@ class EsmPreTrainedModel(PreTrainedModel): config: EsmConfig base_model_prefix = "esm" supports_gradient_checkpointing = True + accepts_loss_kwargs = False _no_split_modules = ["EsmLayer", "EsmFoldTriangularSelfAttentionBlock", "EsmEmbeddings"] _keys_to_ignore_on_load_unexpected = ["position_embeddings.weight"] _supports_flash_attn = True