removed restrictions for custom optimizer#161
removed restrictions for custom optimizer#161ShadenSmith merged 3 commits intodeepspeedai:masterfrom CalogeroZarbo:custom_optimizer
Conversation
|
Looks good |
|
Thanks for your contribution to DeepSpeed! |
|
Thank you too for this amazing lib! |
|
I was thinking this over some more and I'm worried that the lone warning is easy to miss in a typically-long logfile. That could lead to some "silent" divergence errors. What if we keep the warning, but require an additional config in the JSON such as |
|
Yep @ShadenSmith that seems to me as a very good approach to avoid problems difficult do debug in a complex training system. I'll fork again and make another pull-request if that's ok for you. Let me know. Cheers, |
|
@CalogeroZarbo this sounds good. Thanks! |
|
Perfect, thanks @CalogeroZarbo ! |
* Add optimizer swapping * Swap fp16 params to nvme * Formatting * Address review feedback * License file
|
what are the possible errors if i use custom optimizer? or how to ensure its correctness, should i do some tests? |
|
@FWkey, this is a tricky question. You could compare training loss for a smaller model of pytorch DDP vs ZeRO |
Removed the restriction to be able to use any optimizer of choice. A warning message has been added.