Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
3 changes: 2 additions & 1 deletion colossalai/communication/utils.py
Original file line number Diff line number Diff line change
@@ -1,10 +1,11 @@
from typing import List, Tuple, Union

import torch
import torch.distributed as dist

from colossalai.context.parallel_mode import ParallelMode
from colossalai.core import global_context as gpc
from colossalai.utils import get_current_device
from typing import Union, List, Tuple

TensorShape = Union[torch.Size, List[int], Tuple[int]]

Expand Down
Original file line number Diff line number Diff line change
@@ -1,5 +1,7 @@
from typing import Tuple

import torch

from ..registry import meta_profiler_module


Expand All @@ -8,4 +10,4 @@ def torch_nn_embedding(self: torch.nn.Embedding, input: torch.Tensor) -> Tuple[i
# nn.Embedding is a dictionary lookup, so technically it has 0 FLOPs. (https://discuss.pytorch.org/t/correct-way-to-calculate-flops-in-model/67198/6)
flops = 0
macs = 0
return flops, macs
return flops, macs