From 0060f9d9ddd316081c7bf9f4b687bee24cbb1148 Mon Sep 17 00:00:00 2001 From: ethansmith2000 <98723285+ethansmith2000@users.noreply.github.com> Date: Wed, 3 May 2023 00:37:28 -0400 Subject: [PATCH] Remove second instance of scale_lr lr gets scaled twice --- examples/dreambooth/train_dreambooth_lora.py | 5 ----- 1 file changed, 5 deletions(-) diff --git a/examples/dreambooth/train_dreambooth_lora.py b/examples/dreambooth/train_dreambooth_lora.py index 5cefc57c614d..9af81aa5a95d 100644 --- a/examples/dreambooth/train_dreambooth_lora.py +++ b/examples/dreambooth/train_dreambooth_lora.py @@ -746,11 +746,6 @@ def main(args): accelerator.register_for_checkpointing(text_encoder_lora_layers) del temp_pipeline - if args.scale_lr: - args.learning_rate = ( - args.learning_rate * args.gradient_accumulation_steps * args.train_batch_size * accelerator.num_processes - ) - # Enable TF32 for faster training on Ampere GPUs, # cf https://pytorch.org/docs/stable/notes/cuda.html#tensorfloat-32-tf32-on-ampere-devices if args.allow_tf32: