[chat] fix train_prompts.py gemini strategy bug#3666
Conversation
|
What is the problem? I test https://github.com/hpcaitech/ColossalAI/blob/main/applications/Chat/benchmarks/benchmark_opt_lora_dummy.py with gemini strategy and no error occurs. A naive torch module should be able to receive ColoTensor as well. |
run train_prompts.sh with colossal_gemini strategy cause following error initial_model and reward_model are not ZeroDDP module so their weights are not ColoTensor. |
How do you run this script? |
I installed colossal by Something like this under applications/Chat folder |
|
This issue can be simply resolved by move |
|
ColoTensor will be removed in the future. So we'd better reduce the dependency on ColoTensor. |
That's a better solution. I submitted what you proposed. |
📌 Checklist before creating the PR
[doc/gemini/tensor/...]: A concise description🚨 Issue number
📝 What does this PR do?
Intial model and reward model are not wrapped by ZeroDDP wrapper,so they cannot accept ColoTensor as model input. Here we use .data as model input.
💥 Checklist before requesting a review
⭐️ Do you enjoy contributing to Colossal-AI?
Tell us more if you don't enjoy contributing to Colossal-AI.