diff --git a/tensorrt_llm/_torch/modules/fused_moe/quantization.py b/tensorrt_llm/_torch/modules/fused_moe/quantization.py index cfc54c028551..3086469ca887 100644 --- a/tensorrt_llm/_torch/modules/fused_moe/quantization.py +++ b/tensorrt_llm/_torch/modules/fused_moe/quantization.py @@ -1614,7 +1614,7 @@ def load_all_fp4_weight_scales_and_alphas( expert_idx = local_slot_id if not torch.allclose(w1_weight_scale_2, w3_weight_scale_2): - logger.warning( + trtllm_logger.logger.warning( f"w1_weight_scale_2 != w3_weight_scale_2 ({w1_weight_scale_2} != {w3_weight_scale_2}), selecting the larger value. Accuracy may be affected." ) w1_weight_scale_2 = torch.max(w1_weight_scale_2,