diff --git a/src/transformers/trainer.py b/src/transformers/trainer.py index 394a7109d8ac..8d98d629efdb 100755 --- a/src/transformers/trainer.py +++ b/src/transformers/trainer.py @@ -917,7 +917,7 @@ def _align_special_tokens(self): if isinstance(self.processing_class, ProcessorMixin): tokenizer = self.processing_class.tokenizer else: - tokenizer = self.tokenizer + tokenizer = self.processing_class model_has_generation_config = ( hasattr(self.model, "generation_config") and self.model.generation_config is not None )