diff --git a/src/diffusers/models/attention_processor.py b/src/diffusers/models/attention_processor.py index da2920fa67..5206ae7a4b 100644 --- a/src/diffusers/models/attention_processor.py +++ b/src/diffusers/models/attention_processor.py @@ -1096,7 +1096,6 @@ class AttnProcessor2_0: batch_size, sequence_length, _ = ( hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape ) - inner_dim = hidden_states.shape[-1] if attention_mask is not None: attention_mask = attn.prepare_attention_mask(attention_mask, sequence_length, batch_size) @@ -1117,6 +1116,7 @@ class AttnProcessor2_0: key = attn.to_k(encoder_hidden_states) value = attn.to_v(encoder_hidden_states) + inner_dim = key.shape[-1] head_dim = inner_dim // attn.heads query = query.view(batch_size, -1, attn.heads, head_dim).transpose(1, 2)