galbria
5b8ec94fc0
Merge branch 'main' into bria-fibo
2025-10-28 09:47:41 +02:00
galbria
d0a6cb6ed1
Remove unused BriaFibo imports and update CPU offload method in BriaFiboPipeline
2025-10-28 07:47:06 +00:00
Meatfucker
40528e9ae7
Fix typos in kandinsky5 docs ( #12552 )
...
Update kandinsky5.md
Fix typos
2025-10-28 02:54:24 -03:00
Wang, Yi
dc622a95d0
fix crash if tiling mode is enabled ( #12521 )
...
* fix crash in tiling mode is enabled
Signed-off-by: Wang, Yi A <yi.a.wang@intel.com >
* fmt
Signed-off-by: Wang, Yi A <yi.a.wang@intel.com >
---------
Signed-off-by: Wang, Yi A <yi.a.wang@intel.com >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-10-27 17:59:20 -10:00
Dhruv Nair
ecfbc8f952
[Pipelines] Enable Wan VACE to run since single transformer ( #12428 )
...
* update
* update
* update
* update
* update
2025-10-28 09:21:31 +05:30
Sayak Paul
df0e2a4f2c
support latest few-step wan LoRA. ( #12541 )
...
* support latest few-step wan LoRA.
* up
* up
2025-10-28 08:55:24 +05:30
G.O.D
303efd2b8d
Improve pos embed for Flux.1 inference on Ascend NPU ( #12534 )
...
improve pos embed for ascend npu
Co-authored-by: felix01.yu <felix01.yu@vipshop.com >
2025-10-27 16:55:36 -10:00
Lev Novitskiy
5afbcce176
Kandinsky 5 10 sec (NABLA suport) ( #12520 )
...
* add transformer pipeline first version
* updates
* fix 5sec generation
* rewrite Kandinsky5T2VPipeline to diffusers style
* add multiprompt support
* remove prints in pipeline
* add nabla attention
* Wrap Transformer in Diffusers style
* fix license
* fix prompt type
* add gradient checkpointing and peft support
* add usage example
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: Álvaro Somoza <asomoza@users.noreply.github.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: Álvaro Somoza <asomoza@users.noreply.github.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: Álvaro Somoza <asomoza@users.noreply.github.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: Álvaro Somoza <asomoza@users.noreply.github.com >
* Update src/diffusers/models/transformers/transformer_kandinsky.py
Co-authored-by: Álvaro Somoza <asomoza@users.noreply.github.com >
* remove unused imports
* add 10 second models support
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* remove no_grad and simplified prompt paddings
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* moved template to __init__
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/models/transformers/transformer_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* moved sdps inside processor
* remove oneline function
* remove reset_dtype methods
* Transformer: move all methods to forward
* separated prompt encoding
* Update src/diffusers/models/transformers/transformer_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* refactoring
* Update src/diffusers/models/transformers/transformer_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* refactoring acording to acabbc0033
* Update src/diffusers/models/transformers/transformer_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/models/transformers/transformer_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/models/transformers/transformer_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/models/transformers/transformer_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/models/transformers/transformer_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/models/transformers/transformer_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/models/transformers/transformer_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/models/transformers/transformer_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/models/transformers/transformer_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/pipelines/kandinsky5/pipeline_kandinsky.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* fixed
* style +copies
* Update src/diffusers/models/transformers/transformer_kandinsky.py
Co-authored-by: Charles <charles@huggingface.co >
* more
* Apply suggestions from code review
* add lora loader doc
* add compiled Nabla Attention
* all needed changes for 10 sec models are added!
* add docs
* Apply style fixes
* update docs
* add kandinsky5 to toctree
* add tests
* fix tests
* Apply style fixes
* update tests
---------
Co-authored-by: Álvaro Somoza <asomoza@users.noreply.github.com >
Co-authored-by: YiYi Xu <yixu310@gmail.com >
Co-authored-by: Charles <charles@huggingface.co >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-10-28 07:47:18 +05:30
alirezafarashah
6d1a648602
Fix small inconsistency in output dimension of "_get_t5_prompt_embeds" function in sd3 pipeline ( #12531 )
...
* Fix small inconsistency in output dimension of t5 embeds when text_encoder_3 is None
* first commit
---------
Co-authored-by: Alireza Farashah <alireza.farashah@cn-g017.server.mila.quebec >
Co-authored-by: Alireza Farashah <alireza.farashah@login-2.server.mila.quebec >
2025-10-27 07:16:43 -10:00
galbria
0d3fc837fd
Merge branch 'bria-fibo' of github.com:galbria/diffusers into bria-fibo
2025-10-27 16:01:37 +00:00
galbria
57e6315f4d
edit the docs of FIBO
2025-10-27 16:01:31 +00:00
galbria
526004c74e
Merge branch 'main' into bria-fibo
2025-10-27 17:57:34 +02:00
galbria
66aecc438e
Merge branch 'bria-fibo' of github.com:galbria/diffusers into bria-fibo
2025-10-27 15:54:34 +00:00
galbria
f1b5232730
Refactor BriaFibo classes and update pipeline parameters
...
- Updated BriaFiboAttnProcessor and BriaFiboAttention classes to reflect changes from Flux equivalents.
- Modified the _unpack_latents method in BriaFiboPipeline to improve clarity.
- Increased the default max_sequence_length to 3000 and added a new optional parameter do_patching.
- Cleaned up test_pipeline_bria_fibo.py by removing unused imports and skipping unsupported tests.
2025-10-27 15:54:29 +00:00
Mikko Lauri
250f5cb53d
Add AITER attention backend ( #12549 )
...
* add aiter attention backend
* Apply style fixes
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-10-27 20:25:02 +05:30
galbria
84a0beecbe
Merge branch 'main' into bria-fibo
2025-10-27 15:47:39 +02:00
galbria
a617433ace
fix CR
2025-10-27 13:04:57 +00:00
josephrocca
dc6bd1511a
Fix Chroma attention padding order and update docs to use lodestones/Chroma1-HD ( #12508 )
...
* [Fix] Move attention mask padding after T5 embedding
* [Fix] Move attention mask padding after T5 embedding
* Clean up whitespace in pipeline_chroma.py
Removed unnecessary blank lines for cleaner code.
* Fix
* Fix
* Update model to final Chroma1-HD checkpoint
* Update to Chroma1-HD
* Update model to Chroma1-HD
* Update model to Chroma1-HD
* Update Chroma model links to Chroma1-HD
* Add comment about padding/masking
* Fix checkpoint/repo references
* Apply style fixes
---------
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2025-10-27 16:25:20 +05:30
Sayak Paul
500b9cf184
[chore] Move guiders experimental warning ( #12543 )
...
* move guiders experimental warning to init.
* up
2025-10-26 07:41:23 -10:00
Gal Davidi
371e5f511e
style fixs
2025-10-26 16:46:42 +00:00
Gal Davidi
9e253a7bb7
Bria FIBO pipeline
2025-10-26 16:41:39 +00:00
Dhruv Nair
d34b18c783
Deprecate Stable Cascade ( #12537 )
...
* update
* update
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-10-24 22:06:31 +05:30
kaixuanliu
7536f647e4
Loose the criteria tolerance appropriately for Intel XPU devices ( #12460 )
...
* Loose the criteria tolerance appropriately for Intel XPU devices
Signed-off-by: Liu, Kaixuan <kaixuan.liu@intel.com >
* change back the atol value
Signed-off-by: Liu, Kaixuan <kaixuan.liu@intel.com >
* use expectations
Signed-off-by: Liu, Kaixuan <kaixuan.liu@intel.com >
* Update tests/pipelines/kandinsky2_2/test_kandinsky_controlnet.py
---------
Signed-off-by: Liu, Kaixuan <kaixuan.liu@intel.com >
Co-authored-by: Ilyas Moutawwakil <57442720+IlyasMoutawwakil@users.noreply.github.com >
2025-10-24 12:18:15 +02:00
YiYi Xu
a138d71ec1
HunyuanImage21 ( #12333 )
...
* add hunyuanimage2.1
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-10-23 22:31:12 -10:00
Sayak Paul
bc4039886d
fix constants.py to user upper() ( #12479 )
2025-10-24 12:00:02 +05:30
Dhruv Nair
9c3b58dcf1
Handle deprecated transformer classes ( #12517 )
...
* update
* update
* update
2025-10-23 16:22:07 +05:30
Aishwarya Badlani
74b5fed434
Fix MPS compatibility in get_1d_sincos_pos_embed_from_grid #12432 ( #12449 )
...
* Fix MPS compatibility in get_1d_sincos_pos_embed_from_grid #12432
* Fix trailing whitespace in docstring
* Apply style fixes
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-10-23 16:18:07 +05:30
kaixuanliu
85eb505672
fix CI bug for kandinsky3_img2img case ( #12474 )
...
* fix CI bug for kandinsky3_img2img case
Signed-off-by: Liu, Kaixuan <kaixuan.liu@intel.com >
* update code
Signed-off-by: Liu, Kaixuan <kaixuan.liu@intel.com >
---------
Signed-off-by: Liu, Kaixuan <kaixuan.liu@intel.com >
2025-10-23 16:17:22 +05:30
Sayak Paul
ccdd96ca52
[tests] Test attention backends ( #12388 )
...
* add a lightweight test suite for attention backends.
* up
* up
* Apply suggestions from code review
* formatting
2025-10-23 15:09:41 +05:30
Sayak Paul
4c723d8ec3
[CI] xfail the test_wuerstchen_prior test ( #12530 )
...
xfail the test_wuerstchen_prior test
2025-10-22 08:45:47 -10:00
YiYi Xu
bec2d8eaea
Fix: Add _skip_keys for AutoencoderKLWan ( #12523 )
...
add
2025-10-22 07:53:13 -10:00
Álvaro Somoza
a0a51eb098
Kandinsky5 No cfg fix ( #12527 )
...
fix
2025-10-22 22:02:47 +05:30
Sayak Paul
a5a0ccf86a
[core] AutoencoderMixin to abstract common methods ( #12473 )
...
* up
* correct wording.
* up
* up
* up
2025-10-22 08:52:06 +05:30
David Bertoin
dd07b19e27
Prx ( #12525 )
...
* rename photon to prx
* rename photon into prx
* Revert .gitignore to state before commit b7fb0fe9d6
* rename photon to prx
* rename photon into prx
* Revert .gitignore to state before commit b7fb0fe9d6
* make fix-copies
2025-10-21 17:09:22 -07:00
vb
57636ad4f4
purge HF_HUB_ENABLE_HF_TRANSFER; promote Xet ( #12497 )
...
* purge HF_HUB_ENABLE_HF_TRANSFER; promote Xet
* purge HF_HUB_ENABLE_HF_TRANSFER; promote Xet x2
* restrict docker build test to the ones we actually use in CI.
---------
Co-authored-by: YiYi Xu <yixu310@gmail.com >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-10-22 00:59:20 +05:30
David Bertoin
cefc2cf82d
Add Photon model and pipeline support ( #12456 )
...
* Add Photon model and pipeline support
This commit adds support for the Photon image generation model:
- PhotonTransformer2DModel: Core transformer architecture
- PhotonPipeline: Text-to-image generation pipeline
- Attention processor updates for Photon-specific attention mechanism
- Conversion script for loading Photon checkpoints
- Documentation and tests
* just store the T5Gemma encoder
* enhance_vae_properties if vae is provided only
* remove autocast for text encoder forwad
* BF16 example
* conditioned CFG
* remove enhance vae and use vae.config directly when possible
* move PhotonAttnProcessor2_0 in transformer_photon
* remove einops dependency and now inherits from AttentionMixin
* unify the structure of the forward block
* update doc
* update doc
* fix T5Gemma loading from hub
* fix timestep shift
* remove lora support from doc
* Rename EmbedND for PhotoEmbedND
* remove modulation dataclass
* put _attn_forward and _ffn_forward logic in PhotonBlock's forward
* renam LastLayer for FinalLayer
* remove lora related code
* rename vae_spatial_compression_ratio for vae_scale_factor
* support prompt_embeds in call
* move xattention conditionning out computation out of the denoising loop
* add negative prompts
* Use _import_structure for lazy loading
* make quality + style
* add pipeline test + corresponding fixes
* utility function that determines the default resolution given the VAE
* Refactor PhotonAttention to match Flux pattern
* built-in RMSNorm
* Revert accidental .gitignore change
* parameter names match the standard diffusers conventions
* renaming and remove unecessary attributes setting
* Update docs/source/en/api/pipelines/photon.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* quantization example
* added doc to toctree
* Update docs/source/en/api/pipelines/photon.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* Update docs/source/en/api/pipelines/photon.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* Update docs/source/en/api/pipelines/photon.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* use dispatch_attention_fn for multiple attention backend support
* naming changes
* make fix copy
* Update docs/source/en/api/pipelines/photon.md
Co-authored-by: dg845 <58458699+dg845@users.noreply.github.com >
* Add PhotonTransformer2DModel to TYPE_CHECKING imports
* make fix-copies
* Use Tuple instead of tuple
Co-authored-by: dg845 <58458699+dg845@users.noreply.github.com >
* restrict the version of transformers
Co-authored-by: dg845 <58458699+dg845@users.noreply.github.com >
* Update tests/pipelines/photon/test_pipeline_photon.py
Co-authored-by: dg845 <58458699+dg845@users.noreply.github.com >
* Update tests/pipelines/photon/test_pipeline_photon.py
Co-authored-by: dg845 <58458699+dg845@users.noreply.github.com >
* change | for Optional
* fix nits.
* use typing Dict
---------
Co-authored-by: davidb <davidb@worker-10.soperator-worker-svc.soperator.svc.cluster.local >
Co-authored-by: David Briand <david@photoroom.com >
Co-authored-by: davidb <davidb@worker-8.soperator-worker-svc.soperator.svc.cluster.local >
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
Co-authored-by: dg845 <58458699+dg845@users.noreply.github.com >
Co-authored-by: sayakpaul <spsayakpaul@gmail.com >
2025-10-21 20:55:55 +05:30
Sayak Paul
b3e56e71fb
styling issues. ( #12522 )
2025-10-21 20:04:54 +05:30
Steven Liu
5b5fa49a89
[docs] Organize toctree by modality ( #12514 )
...
* reorganize
* fix
---------
Co-authored-by: Álvaro Somoza <asomoza@users.noreply.github.com >
2025-10-21 10:18:54 +05:30
Fei Xie
decfa3c9e1
Fix: Use incorrect temporary variable key when replacing adapter name… ( #12502 )
...
Fix: Use incorrect temporary variable key when replacing adapter name in state dict within load_lora_adapter function
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-10-20 15:45:37 -10:00
Dhruv Nair
48305755bf
Raise warning instead of error when imports are missing for custom code ( #12513 )
...
update
2025-10-20 07:02:23 -10:00
dg845
7853bfbed7
Remove Qwen Image Redundant RoPE Cache ( #12452 )
...
Refactor QwenEmbedRope to only use the LRU cache for RoPE caching
2025-10-19 18:41:58 -07:00
Lev Novitskiy
23ebbb4bc8
Kandinsky 5 is finally in Diffusers! ( #12478 )
...
* add kandinsky5 transformer pipeline first version
---------
Co-authored-by: Álvaro Somoza <asomoza@users.noreply.github.com >
Co-authored-by: YiYi Xu <yixu310@gmail.com >
Co-authored-by: Charles <charles@huggingface.co >
2025-10-17 18:34:30 -10:00
Ali Imran
1b456bd5d5
docs: cleanup of runway model ( #12503 )
...
* cleanup of runway model
* quality fixes
2025-10-17 14:10:50 -07:00
Sayak Paul
af769881d3
[tests] introduce VAETesterMixin to consolidate tests for slicing and tiling ( #12374 )
...
* up
* up
* up
* up
* up
* u[
* up
* up
* up
2025-10-17 12:02:29 +05:30
Sayak Paul
4715c5c769
[ci] xfail more incorrect transformer imports. ( #12455 )
...
* xfail more incorrect transformer imports.
* xfail more.
* up
* up
* up
2025-10-17 10:35:19 +05:30
Steven Liu
dbe413668d
[CI] Check links ( #12491 )
...
* check links
* update
* feedback
* remove
2025-10-16 10:38:16 -07:00
Steven Liu
26475082cb
[docs] Attention checks ( #12486 )
...
* checks
* feedback
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-10-16 09:19:30 -07:00
YiYi Xu
f072c64bf2
ltx0.9.8 (without IC lora, autoregressive sampling) ( #12493 )
...
update
Co-authored-by: Aryan <aryan@huggingface.co >
2025-10-15 07:41:17 -10:00
Sayak Paul
aed636f5f0
[tests] fix clapconfig for text backbone in audioldm2 ( #12490 )
...
fix clapconfig for text backbone in audioldm2
2025-10-15 10:57:09 +05:30
Sayak Paul
53a10518b9
remove unneeded checkpoint imports. ( #12488 )
2025-10-15 09:51:18 +05:30