Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion tests/test_moe/test_moe_checkpoint.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@
from colossalai.zero import ColoInitContext
from tests.test_moe.test_moe_zero_init import MoeModel
from tests.test_tensor.common_utils import debug_print
from tests.test_zero.common import CONFIG
from tests.test_zero.test_legacy.common import CONFIG


def exam_moe_checkpoint():
Expand Down
2 changes: 1 addition & 1 deletion tests/test_moe/test_moe_colo_init.py
Original file line number Diff line number Diff line change
Expand Up @@ -13,7 +13,7 @@
from colossalai.zero import ColoInitContext
from tests.test_moe.test_moe_zero_init import MoeModel
from tests.test_tensor.common_utils import debug_print
from tests.test_zero.common import CONFIG
from tests.test_zero.test_legacy.common import CONFIG


@parameterize("init_device_type", ['cpu', 'cuda'])
Expand Down
2 changes: 1 addition & 1 deletion tests/test_moe/test_moe_zero_init.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@
from colossalai.utils import free_port, get_current_device
from colossalai.zero.legacy.init_ctx import ZeroInitContext
from colossalai.zero.legacy.shard_utils import BucketTensorShardStrategy, TensorShardStrategy
from tests.test_zero.common import CONFIG
from tests.test_zero.test_legacy.common import CONFIG


class MoeModel(nn.Module):
Expand Down
2 changes: 1 addition & 1 deletion tests/test_moe/test_moe_zero_model.py
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@
from colossalai.zero.legacy.sharded_model.utils import col_model_deepcopy
from tests.components_to_test.registry import non_distributed_component_funcs
from tests.test_moe.test_moe_zero_init import MoeModel
from tests.test_zero.common import CONFIG, check_grads_padding, run_fwd_bwd
from tests.test_zero.test_legacy.common import CONFIG, check_grads_padding, run_fwd_bwd


@parameterize("enable_autocast", [False])
Expand Down
2 changes: 1 addition & 1 deletion tests/test_moe/test_moe_zero_optim.py
Original file line number Diff line number Diff line change
Expand Up @@ -20,7 +20,7 @@
from colossalai.zero.low_level._utils import has_inf_or_nan
from tests.components_to_test.registry import non_distributed_component_funcs
from tests.test_moe.test_moe_zero_init import MoeModel
from tests.test_zero.common import CONFIG, check_sharded_model_params
from tests.test_zero.test_legacy.common import CONFIG, check_sharded_model_params


def _run_step(model, optimizer, data, label, criterion, grad_handler):
Expand Down
File renamed without changes.
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@
import torch
import torch.multiprocessing as mp
from common import CONFIG
from test_sharded_optim_v2 import _run_step

import colossalai
from colossalai.nn.optimizer import HybridAdam
Expand All @@ -16,7 +17,6 @@
from colossalai.zero.legacy.sharded_optim import ShardedOptimizerV2
from colossalai.zero.low_level._utils import has_inf_or_nan
from tests.components_to_test.registry import non_distributed_component_funcs
from tests.test_zero.test_sharded_optim_v2 import _run_step


@parameterize("cpu_offload", [True, False])
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@
import pytest
import torch
import torch.multiprocessing as mp
from common import CONFIG, allclose

import colossalai
from colossalai.testing import parameterize, rerun_if_address_is_in_use
Expand All @@ -12,7 +13,6 @@
from colossalai.zero.legacy.shard_utils import BucketTensorShardStrategy, TensorShardStrategy
from colossalai.zero.legacy.sharded_param import ShardedTensor
from colossalai.zero.legacy.sharded_param.sharded_param import ShardedParamV2
from tests.test_zero.common import CONFIG, allclose


@parameterize("shard_strategy_class", [TensorShardStrategy, BucketTensorShardStrategy])
Expand Down
Original file line number Diff line number Diff line change
@@ -1,7 +1,6 @@
#!/usr/bin/env python
# -*- encoding: utf-8 -*-

from copy import deepcopy
from functools import partial

import pytest
Expand Down