From a57a7af45cbef004c38e2a294a6457f7f3574e5d Mon Sep 17 00:00:00 2001 From: Aryan Date: Tue, 6 Aug 2024 07:23:48 +0530 Subject: [PATCH] [bug] remove unreachable norm_type=ada_norm_continuous from norm3 initialization conditions (#9006) remove ada_norm_continuous from norm3 list Co-authored-by: Sayak Paul --- src/diffusers/models/attention.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/diffusers/models/attention.py b/src/diffusers/models/attention.py index b204770e6d..74a9d9efc6 100644 --- a/src/diffusers/models/attention.py +++ b/src/diffusers/models/attention.py @@ -376,7 +376,7 @@ class BasicTransformerBlock(nn.Module): "layer_norm", ) - elif norm_type in ["ada_norm_zero", "ada_norm", "layer_norm", "ada_norm_continuous"]: + elif norm_type in ["ada_norm_zero", "ada_norm", "layer_norm"]: self.norm3 = nn.LayerNorm(dim, norm_eps, norm_elementwise_affine) elif norm_type == "layer_norm_i2vgen": self.norm3 = None