Add two features which supports training PPO in one graphic card for large model and ChatGLM-6B model support#3567
Add two features which supports training PPO in one graphic card for large model and ChatGLM-6B model support#3567yynil wants to merge 3 commits intohpcaitech:mainfrom
Conversation
|
@yynil hello, if you want to support more models, can you add all models class in https://github.com/hpcaitech/ColossalAI/tree/main/applications/Chat/coati/models? |
|
I think this PR changes too many things, I think you can split it into three PR:
|
I'll create another branch to separate these changes. |
|
Since the ChatGLM is not willing to release a smaller model to public to train a reward model, I'm suspending the support to ChatGLM. My Branch will then move to bloom because bloom has a very good small model to allow us train reward model much easier. |
Thanks, you are welcome to share and update your PR. |
📌 Checklist before creating the PR
[doc/gemini/tensor/...]: A concise description🚨 Issue number
📝 What does this PR do?
💥 Checklist before requesting a review
⭐️ Do you enjoy contributing to Colossal-AI?
Tell us more if you don't enjoy contributing to Colossal-AI.