🐛 Describe the bug
when training GPT2-S using a single card on colab, !torchrun --standalone --nproc_per_node 1 benchmark_gpt_dummy.py --model s --strategy colossalai_gemini_cpu --experience_batch_size 1 --train_batch_size 1
meetting a bug "AssertionError: You should use zero_ddp_wrapper first
ERROR:torch.distributed.elastic.multiprocessing.api:failed (exitcode: 1) local_rank: 0 (pid: 21342) of binary: /usr/bin/python3"

Environment
No response
🐛 Describe the bug
when training GPT2-S using a single card on colab,
!torchrun --standalone --nproc_per_node 1 benchmark_gpt_dummy.py --model s --strategy colossalai_gemini_cpu --experience_batch_size 1 --train_batch_size 1meetting a bug "AssertionError: You should use
zero_ddp_wrapperfirstERROR:torch.distributed.elastic.multiprocessing.api:failed (exitcode: 1) local_rank: 0 (pid: 21342) of binary: /usr/bin/python3"
Environment
No response