Skip to content

Fix deepcopy in ProcessorMixin.to_dict for GemmaTokenizerFast#41851

Open
aijadugar wants to merge 17 commits intohuggingface:mainfrom
aijadugar:fix-multiple-tokenizers-saved-v2
Open

Fix deepcopy in ProcessorMixin.to_dict for GemmaTokenizerFast#41851
aijadugar wants to merge 17 commits intohuggingface:mainfrom
aijadugar:fix-multiple-tokenizers-saved-v2

Conversation

@aijadugar
Copy link
Copy Markdown

Description:

Replaced deepcopy with a shallow copy to avoid AttributeError with fast tokenizers.

Added a test (tests/utils/test_processor_utils.py) to verify multiple tokenizers save/load correctly.

Fix #41837

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant