1
0
mirror of https://github.com/huggingface/diffusers.git synced 2026-01-27 17:22:53 +03:00

[Versatile Diffusion] Fix cross_attention_kwargs (#1849)

fix versatile
This commit is contained in:
Patrick von Platen
2022-12-28 18:49:04 +01:00
committed by GitHub
parent 2ba42aa9b1
commit debc74f442

View File

@@ -703,7 +703,13 @@ class DualTransformer2DModel(nn.Module):
self.transformer_index_for_condition = [1, 0]
def forward(
self, hidden_states, encoder_hidden_states, timestep=None, attention_mask=None, return_dict: bool = True
self,
hidden_states,
encoder_hidden_states,
timestep=None,
attention_mask=None,
cross_attention_kwargs=None,
return_dict: bool = True,
):
"""
Args:
@@ -738,6 +744,7 @@ class DualTransformer2DModel(nn.Module):
input_states,
encoder_hidden_states=condition_state,
timestep=timestep,
cross_attention_kwargs=cross_attention_kwargs,
return_dict=False,
)[0]
encoded_states.append(encoded_state - input_states)