From ada3bb941bb45c1f9c3c4ea6b8feca4a3031768c Mon Sep 17 00:00:00 2001 From: Rinne Date: Thu, 7 Mar 2024 15:55:22 +0800 Subject: [PATCH] fix: remove duplicated code in TemporalBasicTransformerBlock. (#7212) fix: remove duplicate code in TemporalBasicTransformerBlock. Co-authored-by: Sayak Paul --- src/diffusers/models/attention.py | 1 - 1 file changed, 1 deletion(-) diff --git a/src/diffusers/models/attention.py b/src/diffusers/models/attention.py index a4b3ee58a8..509c201a45 100644 --- a/src/diffusers/models/attention.py +++ b/src/diffusers/models/attention.py @@ -440,7 +440,6 @@ class TemporalBasicTransformerBlock(nn.Module): # Define 3 blocks. Each block has its own normalization layer. # 1. Self-Attn - self.norm_in = nn.LayerNorm(dim) self.ff_in = FeedForward( dim, dim_out=time_mix_inner_dim,