1
0
mirror of https://github.com/huggingface/diffusers.git synced 2026-01-27 17:22:53 +03:00

Support attn2==None for xformers (#1759)

This commit is contained in:
Anton Lozhkov
2022-12-19 12:43:30 +01:00
committed by GitHub
parent 429e5449c1
commit 32a5d70c42

View File

@@ -473,7 +473,8 @@ class BasicTransformerBlock(nn.Module):
except Exception as e:
raise e
self.attn1._use_memory_efficient_attention_xformers = use_memory_efficient_attention_xformers
self.attn2._use_memory_efficient_attention_xformers = use_memory_efficient_attention_xformers
if self.attn2 is not None:
self.attn2._use_memory_efficient_attention_xformers = use_memory_efficient_attention_xformers
def forward(self, hidden_states, encoder_hidden_states=None, timestep=None, attention_mask=None):
# 1. Self-Attention