diff --git a/op_builder/builder.py b/op_builder/builder.py index 5a45f59f4b43..807cbee7aa70 100644 --- a/op_builder/builder.py +++ b/op_builder/builder.py @@ -296,7 +296,7 @@ def cxx_args(self): ''' return [] - def is_compatible(self, verbose=True): + def is_compatible(self, verbose=False): ''' Check if all non-python dependencies are satisfied to build this op ''' @@ -679,7 +679,7 @@ def version_dependent_macros(self): version_ge_1_5 = ['-DVERSION_GE_1_5'] return version_ge_1_1 + version_ge_1_3 + version_ge_1_5 - def is_compatible(self, verbose=True): + def is_compatible(self, verbose=False): return super().is_compatible(verbose) def builder(self): diff --git a/op_builder/cpu/comm.py b/op_builder/cpu/comm.py index 69b9226fd806..fec960b63b2e 100644 --- a/op_builder/cpu/comm.py +++ b/op_builder/cpu/comm.py @@ -28,7 +28,7 @@ def include_paths(self): def cxx_args(self): return ['-O2', '-fopenmp'] - def is_compatible(self, verbose=True): + def is_compatible(self, verbose=False): # TODO: add soft compatibility check for private binary release. # a soft check, as in we know it can be trivially changed. return super().is_compatible(verbose) @@ -65,7 +65,7 @@ def include_paths(self): def cxx_args(self): return ['-O2', '-fopenmp'] - def is_compatible(self, verbose=True): + def is_compatible(self, verbose=False): # TODO: add soft compatibility check for private binary release. # a soft check, as in we know it can be trivially changed. return super().is_compatible(verbose) diff --git a/op_builder/inference_core_ops.py b/op_builder/inference_core_ops.py index 45e8628e669f..f7c0b47f92c6 100755 --- a/op_builder/inference_core_ops.py +++ b/op_builder/inference_core_ops.py @@ -19,7 +19,7 @@ def __init__(self, name=None): def absolute_name(self): return f'deepspeed.inference.v2.kernels{self.NAME}' - def is_compatible(self, verbose=True): + def is_compatible(self, verbose=False): try: import torch except ImportError: diff --git a/op_builder/inference_cutlass_builder.py b/op_builder/inference_cutlass_builder.py index fda6e74bbf6a..aa5294b1cbda 100644 --- a/op_builder/inference_cutlass_builder.py +++ b/op_builder/inference_cutlass_builder.py @@ -18,7 +18,7 @@ def __init__(self, name=None): def absolute_name(self): return f'deepspeed.inference.v2.kernels.cutlass_ops.{self.NAME}' - def is_compatible(self, verbose=True): + def is_compatible(self, verbose=False): try: import torch except ImportError: diff --git a/op_builder/npu/async_io.py b/op_builder/npu/async_io.py index 3eb898b75693..76d495b8106f 100644 --- a/op_builder/npu/async_io.py +++ b/op_builder/npu/async_io.py @@ -83,7 +83,7 @@ def check_for_libaio_pkg(self): break return found - def is_compatible(self, verbose=True): + def is_compatible(self, verbose=False): # Check for the existence of libaio by using distutils # to compile and link a test program that calls io_submit, # which is a function provided by libaio that is used in the async_io op. diff --git a/op_builder/ragged_ops.py b/op_builder/ragged_ops.py index a4e365786a2b..10afb193c738 100644 --- a/op_builder/ragged_ops.py +++ b/op_builder/ragged_ops.py @@ -19,7 +19,7 @@ def __init__(self, name=None): def absolute_name(self): return f'deepspeed.inference.v2.kernels.ragged_ops.{self.NAME}' - def is_compatible(self, verbose=True): + def is_compatible(self, verbose=False): try: import torch except ImportError: diff --git a/op_builder/ragged_utils.py b/op_builder/ragged_utils.py index a855f072af8c..654ba07e0879 100755 --- a/op_builder/ragged_utils.py +++ b/op_builder/ragged_utils.py @@ -19,7 +19,7 @@ def __init__(self, name=None): def absolute_name(self): return f'deepspeed.inference.v2.{self.NAME}' - def is_compatible(self, verbose=True): + def is_compatible(self, verbose=False): try: import torch except ImportError: diff --git a/op_builder/sparse_attn.py b/op_builder/sparse_attn.py index 2385adc8fe9c..66f748ee8f7a 100644 --- a/op_builder/sparse_attn.py +++ b/op_builder/sparse_attn.py @@ -64,7 +64,6 @@ def is_compatible(self, verbose=False): if verbose: self.warning( f'{self.NAME} requires a torch version >= 1.5 and < 2.0 but detected {TORCH_MAJOR}.{TORCH_MINOR}') - try: import triton except ImportError: diff --git a/op_builder/spatial_inference.py b/op_builder/spatial_inference.py index 8a6b36cce0b0..d6c5fa661156 100644 --- a/op_builder/spatial_inference.py +++ b/op_builder/spatial_inference.py @@ -17,7 +17,7 @@ def __init__(self, name=None): def absolute_name(self): return f'deepspeed.ops.spatial.{self.NAME}_op' - def is_compatible(self, verbose=True): + def is_compatible(self, verbose=False): try: import torch except ImportError: diff --git a/op_builder/transformer_inference.py b/op_builder/transformer_inference.py index 88b77499cc0e..1b056ecef3ed 100755 --- a/op_builder/transformer_inference.py +++ b/op_builder/transformer_inference.py @@ -17,7 +17,7 @@ def __init__(self, name=None): def absolute_name(self): return f'deepspeed.ops.transformer.inference.{self.NAME}_op' - def is_compatible(self, verbose=True): + def is_compatible(self, verbose=False): try: import torch except ImportError: diff --git a/op_builder/xpu/async_io.py b/op_builder/xpu/async_io.py index 62c3298fc72c..7ed527e016fa 100644 --- a/op_builder/xpu/async_io.py +++ b/op_builder/xpu/async_io.py @@ -79,7 +79,7 @@ def check_for_libaio_pkg(self): break return found - def is_compatible(self, verbose=True): + def is_compatible(self, verbose=False): # Check for the existence of libaio by using distutils # to compile and link a test program that calls io_submit, # which is a function provided by libaio that is used in the async_io op.