Skip to content

[zero] improving compatibility of low level zero #3954

@Gy-Lu

Description

@Gy-Lu

Overview

For now, the way of low level zero in ColossalAI sharding the OS when init may lead to an unbalanced load on each rank. For instance, if a model had 5 params(each contains some parameters) and we got 8 GPUs with Zero-DP, with the current implementation, 3 GPUs would have no OS on them. Thus, a sharding method splitting all parameters evenly is needed.

Goal

With a new sharding method, each rank would get a similar number of parameters.

Metadata

Metadata

Assignees

Labels

No labels
No labels

Type

No type

Projects

Status

✅ Done

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions