From 5defd901891603f0057ada317e62c1adefd38daf Mon Sep 17 00:00:00 2001 From: Patrick von Platen Date: Tue, 4 Jul 2023 11:55:18 +0200 Subject: [PATCH 1/3] Fix failing np tests --- tests/pipelines/test_pipelines_common.py | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/tests/pipelines/test_pipelines_common.py b/tests/pipelines/test_pipelines_common.py index 52dd4afd6b21..676dfe6e0322 100644 --- a/tests/pipelines/test_pipelines_common.py +++ b/tests/pipelines/test_pipelines_common.py @@ -698,11 +698,13 @@ def _test_xformers_attention_forwardGenerator_pass( pipe.set_progress_bar_config(disable=None) inputs = self.get_dummy_inputs(torch_device) - output_without_offload = pipe(**inputs)[0].cpu() + output_without_offload = pipe(**inputs)[0] + output_without_offload.cpu() if torch.is_tensor(output_without_offload) pipe.enable_xformers_memory_efficient_attention() inputs = self.get_dummy_inputs(torch_device) output_with_offload = pipe(**inputs)[0].cpu() + output_with_offload.cpu() if torch.is_tensor(output_with_offload) if test_max_difference: max_diff = np.abs(output_with_offload - output_without_offload).max() From 629ec263bc3b525a08371b8b3c2fa3ebde80f546 Mon Sep 17 00:00:00 2001 From: Patrick von Platen Date: Tue, 4 Jul 2023 13:42:56 +0200 Subject: [PATCH 2/3] Apply suggestions from code review --- tests/pipelines/test_pipelines_common.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/tests/pipelines/test_pipelines_common.py b/tests/pipelines/test_pipelines_common.py index 676dfe6e0322..256bdc35cd8d 100644 --- a/tests/pipelines/test_pipelines_common.py +++ b/tests/pipelines/test_pipelines_common.py @@ -699,12 +699,12 @@ def _test_xformers_attention_forwardGenerator_pass( inputs = self.get_dummy_inputs(torch_device) output_without_offload = pipe(**inputs)[0] - output_without_offload.cpu() if torch.is_tensor(output_without_offload) + output_without_offload.cpu() if torch.is_tensor(output_without_offload) else output_without_offload pipe.enable_xformers_memory_efficient_attention() inputs = self.get_dummy_inputs(torch_device) output_with_offload = pipe(**inputs)[0].cpu() - output_with_offload.cpu() if torch.is_tensor(output_with_offload) + output_with_offload.cpu() if torch.is_tensor(output_with_offload) else output_without_offload if test_max_difference: max_diff = np.abs(output_with_offload - output_without_offload).max() From 9d21bb984c8f01d7e9805622caf5eb9bb524ab99 Mon Sep 17 00:00:00 2001 From: Patrick von Platen Date: Tue, 4 Jul 2023 13:43:16 +0200 Subject: [PATCH 3/3] Update tests/pipelines/test_pipelines_common.py --- tests/pipelines/test_pipelines_common.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tests/pipelines/test_pipelines_common.py b/tests/pipelines/test_pipelines_common.py index 256bdc35cd8d..e97bdb352b22 100644 --- a/tests/pipelines/test_pipelines_common.py +++ b/tests/pipelines/test_pipelines_common.py @@ -703,7 +703,7 @@ def _test_xformers_attention_forwardGenerator_pass( pipe.enable_xformers_memory_efficient_attention() inputs = self.get_dummy_inputs(torch_device) - output_with_offload = pipe(**inputs)[0].cpu() + output_with_offload = pipe(**inputs)[0] output_with_offload.cpu() if torch.is_tensor(output_with_offload) else output_without_offload if test_max_difference: