From 83531070fd145d8258c8bee54ece8e05a6ec124d Mon Sep 17 00:00:00 2001 From: Kent Keirsey <31807370+hipsterusername@users.noreply.github.com> Date: Tue, 13 Dec 2022 21:11:29 -0500 Subject: [PATCH] Commented out WIP attention map saver --- ldm/models/diffusion/shared_invokeai_diffusion.py | 8 +++++--- 1 file changed, 5 insertions(+), 3 deletions(-) diff --git a/ldm/models/diffusion/shared_invokeai_diffusion.py b/ldm/models/diffusion/shared_invokeai_diffusion.py index 9261d3abbc9..da5be7d6006 100644 --- a/ldm/models/diffusion/shared_invokeai_diffusion.py +++ b/ldm/models/diffusion/shared_invokeai_diffusion.py @@ -55,10 +55,12 @@ def remove_cross_attention_control(self): def setup_attention_map_saving(self, saver: AttentionMapSaver): def callback(slice, dim, offset, slice_size, key): - if dim is not None: + #if dim is not None: # sliced tokens attention map saving is not implemented - return - saver.add_attention_maps(slice, key) + #return + #saver.add_attention_maps(slice, key) + # NOTE: Commented out above and added a return to fix regression in Unified Canvas + return tokens_cross_attention_modules = get_cross_attention_modules(self.model, CrossAttentionType.TOKENS) for identifier, module in tokens_cross_attention_modules: