From e76fc948b0df0cc0357a4ab3c1cf4b2991f3377d Mon Sep 17 00:00:00 2001 From: Aryan Date: Wed, 16 Jul 2025 21:41:09 +0200 Subject: [PATCH] fix --- src/diffusers/models/attention_dispatch.py | 1 + 1 file changed, 1 insertion(+) diff --git a/src/diffusers/models/attention_dispatch.py b/src/diffusers/models/attention_dispatch.py index 6357bdad0f..8d149cb7a2 100644 --- a/src/diffusers/models/attention_dispatch.py +++ b/src/diffusers/models/attention_dispatch.py @@ -899,6 +899,7 @@ def _templated_context_parallel_attention( @_AttentionBackendRegistry.register( AttentionBackendName.FLASH, constraints=[_check_device, _check_qkv_dtype_bf16_or_fp16, _check_shape], + supports_context_parallel=True, ) def _flash_attention( query: torch.Tensor,