System Info
transformers version: 5.5.0
- Platform: Linux-6.6.113+-x86_64-with-glibc2.35
- Python version: 3.12.13
- Huggingface_hub version: 1.8.0
- Safetensors version: 0.7.0
- Accelerate version: 1.13.0
- Accelerate config: not found
- DeepSpeed version: not installed
- PyTorch version (accelerator?): 2.10.0+cu128 (CUDA)
- Using distributed or parallel set-up in script?:
- Using GPU in script?:
- GPU type: Tesla T4
(Google Colaboratory GPU)
Who can help?
@zucchini-nlp @Cyrilvallez
Information
Tasks
Reproduction
Reproduce the behavior:
from transformers import Gemma4ForConditionalGeneration
mdl = Gemma4ForConditionalGeneration.from_pretrained("google/gemma-4-E2B-it")
e = mdl.get_input_embeddings()
f = mdl.model.language_model.embed_tokens_per_layer
g = mdl.get_output_embeddings()
print(e.num_embeddings, f.num_embeddings, g.out_features)
assert e.num_embeddings == f.num_embeddings == g.out_features
e = mdl.resize_token_embeddings(e.num_embeddings + 1)
f = mdl.model.language_model.embed_tokens_per_layer
g = mdl.get_output_embeddings()
print(e.num_embeddings, f.num_embeddings, g.out_features)
assert e.num_embeddings == f.num_embeddings == g.out_features
Expected behavior
All e.num_embeddings, f.num_embeddings and g.out_features should be increased to 262145.
System Info
transformersversion: 5.5.0(Google Colaboratory GPU)
Who can help?
@zucchini-nlp @Cyrilvallez
Information
Tasks
examplesfolder (such as GLUE/SQuAD, ...)Reproduction
Reproduce the behavior:
Expected behavior
All
e.num_embeddings,f.num_embeddingsandg.out_featuresshould be increased to 262145.