From cd560a4598cad0b750cafba493ef035c6ca155e5 Mon Sep 17 00:00:00 2001 From: Minghao Huang Date: Mon, 27 Mar 2023 17:16:59 +0800 Subject: [PATCH] [NFC] polish colossalai/fx/profiler/experimental/profiler_module/embedding.py code style --- .../fx/profiler/experimental/profiler_module/embedding.py | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/colossalai/fx/profiler/experimental/profiler_module/embedding.py b/colossalai/fx/profiler/experimental/profiler_module/embedding.py index dca6f9453af3..a1ade5d3ad93 100644 --- a/colossalai/fx/profiler/experimental/profiler_module/embedding.py +++ b/colossalai/fx/profiler/experimental/profiler_module/embedding.py @@ -1,5 +1,7 @@ from typing import Tuple + import torch + from ..registry import meta_profiler_module @@ -8,4 +10,4 @@ def torch_nn_embedding(self: torch.nn.Embedding, input: torch.Tensor) -> Tuple[i # nn.Embedding is a dictionary lookup, so technically it has 0 FLOPs. (https://discuss.pytorch.org/t/correct-way-to-calculate-flops-in-model/67198/6) flops = 0 macs = 0 - return flops, macs \ No newline at end of file + return flops, macs