From 977d4b4dd243e5b799a29e0a79996245a3686434 Mon Sep 17 00:00:00 2001 From: Winnie Chow Date: Tue, 30 Sep 2025 21:12:39 +0000 Subject: [PATCH 1/2] Add num_hidden_layers to top level config --- src/transformers/models/t5gemma/modular_t5gemma.py | 1 + 1 file changed, 1 insertion(+) diff --git a/src/transformers/models/t5gemma/modular_t5gemma.py b/src/transformers/models/t5gemma/modular_t5gemma.py index d41c8796e5ee..10b746b76ef6 100644 --- a/src/transformers/models/t5gemma/modular_t5gemma.py +++ b/src/transformers/models/t5gemma/modular_t5gemma.py @@ -314,6 +314,7 @@ def __init__( super().__init__(**kwargs) self.is_encoder_decoder = is_encoder_decoder + self.num_hidden_layers = kwargs.get("num_hidden_layers", decoder.num_hidden_layers) self.use_cache = kwargs.get("use_cache", decoder.use_cache) self.initializer_range = kwargs.get("initializer_range", decoder.initializer_range) self.dropout_rate = dropout_rate From db2c61a0ebbc19ea789ead9e9b730d889b50bdd2 Mon Sep 17 00:00:00 2001 From: Winnie Chow Date: Tue, 30 Sep 2025 21:56:29 +0000 Subject: [PATCH 2/2] Update configuration_t5_gemma to match modular --- src/transformers/models/t5gemma/configuration_t5gemma.py | 1 + 1 file changed, 1 insertion(+) diff --git a/src/transformers/models/t5gemma/configuration_t5gemma.py b/src/transformers/models/t5gemma/configuration_t5gemma.py index 76ad99132056..f2820789afd6 100644 --- a/src/transformers/models/t5gemma/configuration_t5gemma.py +++ b/src/transformers/models/t5gemma/configuration_t5gemma.py @@ -299,6 +299,7 @@ def __init__( super().__init__(**kwargs) self.is_encoder_decoder = is_encoder_decoder + self.num_hidden_layers = kwargs.get("num_hidden_layers", decoder.num_hidden_layers) self.use_cache = kwargs.get("use_cache", decoder.use_cache) self.initializer_range = kwargs.get("initializer_range", decoder.initializer_range) self.dropout_rate = dropout_rate