Steven Liu
cc5b31ffc9
[docs] Migrate syntax ( #12390 )
...
* change syntax
* make style
2025-09-30 10:11:19 -07:00
Steven Liu
d7a1a0363f
[docs] CP ( #12331 )
...
* init
* feedback
* feedback
* feedback
* feedback
* feedback
* feedback
2025-09-30 09:33:41 -07:00
Lucain
b59654544b
Install latest prerelease from huggingface_hub when installing transformers from main ( #12395 )
...
* Allow prerelease when installing transformers from main
* maybe better
* maybe better
* and now?
* just bored
* should be better
* works now
2025-09-30 17:02:33 +05:30
Yao Matrix
0e12ba7454
fix 3 xpu failures uts w/ latest pytorch ( #12408 )
...
fix xpu ut failures w/ latest pytorch
Signed-off-by: Yao, Matrix <matrix.yao@intel.com >
2025-09-30 14:07:48 +05:30
Dhruv Nair
20fd00b14b
[Tests] Add single file tester mixin for Models and remove unittest dependency ( #12352 )
...
* update
* update
* update
* update
* update
2025-09-30 13:28:34 +05:30
YiYi Xu
76d4e416bc
[modular]some small fix ( #12307 )
...
* fix
* add mellon node registry
* style
* update docstring to include more info!
* support custom node mellon
* HTTPErrpr -> HfHubHTTPErrpr
* up
* Update src/diffusers/modular_pipelines/qwenimage/node_utils.py
2025-09-29 11:42:34 -10:00
Steven Liu
c07fcf780a
[docs] Model formats ( #12256 )
...
* init
* config
* lora metadata
* feedback
* fix
* cache allocator warmup for from_single_file
* feedback
* feedback
2025-09-29 11:36:14 -07:00
Steven Liu
ccedeca96e
[docs] Distributed inference ( #12285 )
...
* init
* feedback
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-09-29 11:24:26 -07:00
Sayak Paul
64a5187d96
[quantization] feat: support aobaseconfig classes in TorchAOConfig ( #12275 )
...
* feat: support aobaseconfig classes.
* [docs] AOBaseConfig (#12302 )
init
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* up
* replace with is_torchao_version
* up
* up
---------
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
2025-09-29 18:04:18 +05:30
Akshay Babbar
0a151115bb
Fix #12116 : preserve boolean dtype for attention masks in ChromaPipeline ( #12263 )
...
* fix: preserve boolean dtype for attention masks in ChromaPipeline
- Convert attention masks to bool and prevent dtype corruption
- Fix both positive and negative mask handling in _get_t5_prompt_embeds
- Remove float conversion in _prepare_attention_mask method
Fixes #12116
* test: add ChromaPipeline attention mask dtype tests
* test: add slow ChromaPipeline attention mask tests
* chore: removed comments
* refactor: removing redundant type conversion
* Remove dedicated dtype tests as per feedback
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2025-09-29 14:20:05 +05:30
Sayak Paul
19085ac8f4
Don't skip Qwen model tests for group offloading with disk ( #12382 )
...
u[
2025-09-29 13:08:05 +05:30
Sayak Paul
041501aea9
[docs] remove docstrings from repeated methods in lora_pipeline.py ( #12393 )
...
* start unbloating docstrings (save_lora_weights).
* load_lora_weights()
* lora_state_dict
* fuse_lora
* unfuse_lora
* load_lora_into_transformer
2025-09-26 22:38:43 +05:30
Sayak Paul
9c0944581a
[docs] slight edits to the attention backends docs. ( #12394 )
...
* slight edits to the attention backends docs.
* Update docs/source/en/optimization/attention_backends.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
---------
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
2025-09-26 21:50:16 +05:30
Sayak Paul
4588bbeb42
[CI] disable installing transformers from main in ci for now. ( #12397 )
...
* disable installing transformers from main in ci for now.
* up
* u[p
2025-09-26 18:41:17 +05:30
Lucain
ec5449f3a1
Support both huggingface_hub v0.x and v1.x ( #12389 )
...
* Support huggingface_hub 0.x and 1.x
* httpx
2025-09-25 18:28:54 +02:00
DefTruth
310fdaf556
Introduce cache-dit to community optimization ( #12366 )
...
* docs: introduce cache-dit to diffusers
* docs: introduce cache-dit to diffusers
* docs: introduce cache-dit to diffusers
* docs: introduce cache-dit to diffusers
* docs: introduce cache-dit to diffusers
* docs: introduce cache-dit to diffusers
* docs: introduce cache-dit to diffusers
* misc: update examples link
* misc: update examples link
* docs: introduce cache-dit to diffusers
* docs: introduce cache-dit to diffusers
* docs: introduce cache-dit to diffusers
* docs: introduce cache-dit to diffusers
* docs: introduce cache-dit to diffusers
* Refine documentation for CacheDiT features
Updated the wording for clarity and consistency in the documentation. Adjusted sections on cache acceleration, automatic block adapter, patch functor, and hybrid cache configuration.
2025-09-24 10:50:57 -07:00
Aryan
dcb6dd9b7a
Context Parallel w/ Ring & Ulysses & Unified Attention ( #11941 )
...
* update
* update
* add coauthor
Co-Authored-By: Dhruv Nair <dhruv.nair@gmail.com >
* improve test
* handle ip adapter params correctly
* fix chroma qkv fusion test
* fix fastercache implementation
* fix more tests
* fight more tests
* add back set_attention_backend
* update
* update
* make style
* make fix-copies
* make ip adapter processor compatible with attention dispatcher
* refactor chroma as well
* remove rmsnorm assert
* minify and deprecate npu/xla processors
* update
* refactor
* refactor; support flash attention 2 with cp
* fix
* support sage attention with cp
* make torch compile compatible
* update
* refactor
* update
* refactor
* refactor
* add ulysses backward
* try to make dreambooth script work; accelerator backward not playing well
* Revert "try to make dreambooth script work; accelerator backward not playing well"
This reverts commit 768d0ea6fa .
* workaround compilation problems with triton when doing all-to-all
* support wan
* handle backward correctly
* support qwen
* support ltx
* make fix-copies
* Update src/diffusers/models/modeling_utils.py
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
* apply review suggestions
* update docs
* add explanation
* make fix-copies
* add docstrings
* support passing parallel_config to from_pretrained
* apply review suggestions
* make style
* update
* Update docs/source/en/api/parallel.md
Co-authored-by: Aryan <aryan@huggingface.co >
* up
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
Co-authored-by: sayakpaul <spsayakpaul@gmail.com >
2025-09-24 19:03:25 +05:30
Alberto Chimenti
043ab2520f
Fix WanVACEPipeline to allow prompt to be None and skip encoding step ( #12251 )
...
Fixed WanVACEPipeline to allow prompt to be None and skip encoding step
2025-09-24 15:15:04 +05:30
Yao Matrix
08c29020dd
fix marigold ut case fail on xpu ( #12350 )
...
Signed-off-by: Yao, Matrix <matrix.yao@intel.com >
2025-09-24 09:32:06 +05:30
Yao Matrix
7a58734994
xpu enabling for 4 cases ( #12345 )
...
Signed-off-by: Yao, Matrix <matrix.yao@intel.com >
2025-09-24 09:31:45 +05:30
Sayak Paul
9ef118509e
[tests] disable xformer tests for pipelines it isn't popular. ( #12277 )
...
disable xformer tests for pipelines it isn't popular.
2025-09-24 09:02:25 +05:30
Dhruv Nair
7c54a7b38a
Fix Custom Code loading ( #12378 )
...
* update
* update
* update
2025-09-24 08:53:41 +05:30
Sayak Paul
09e777a3e1
[tests] Single scheduler in lora tests ( #12315 )
...
* single scheduler please.
* up
* up
* up
2025-09-24 08:36:50 +05:30
Steven Liu
a72bc0c4bb
[docs] Attention backends ( #12320 )
...
* init
* feedback
* update
* feedback
* fixes
2025-09-23 10:59:46 -07:00
Dhruv Nair
80de641c1c
Allow Automodel to support custom model code ( #12353 )
...
* update
* update
2025-09-23 07:31:42 -10:00
Steven Liu
76810eca2b
[docs] Schedulers ( #12246 )
...
* init
* toctree
* scheduler suggestions
* toctree
2025-09-23 10:29:16 -07:00
SahilCarterr
1448b03585
[Fix] chroma docs ( #12360 )
...
* Fixes chroma docs
* fix docs
fixed docs are now consistent
2025-09-22 13:04:13 -07:00
Sayak Paul
5796735015
add test and doc for QwenImageEdit Plus ( #12363 )
...
* up
* xfail some tests
* up
* up
2025-09-22 21:57:30 +05:30
Sayak Paul
d8310a8fca
[lora] factor out the overlaps in save_lora_weights(). ( #12027 )
...
* factor out the overlaps in save_lora_weights().
* remove comment.
* remove comment.
* up
* fix-copies
2025-09-22 15:14:39 +05:30
SahilCarterr
78031c2938
[Fix] enable_xformers_memory_efficient_attention() in Flux Pipeline ( #12337 )
...
* FIxes enable_xformers_memory_efficient_attention()
* Update attention.py
2025-09-22 12:37:41 +05:30
Chen Mingyi
d83d35c1bb
Fix bug with VAE slicing in autoencoder_dc.py ( #12343 )
2025-09-22 12:25:34 +05:30
Sayak Paul
843355f89f
[tests] xfail some kandinsky tests. ( #12364 )
...
xfail some kandinsky tests.
2025-09-22 11:17:47 +05:30
Jason Cox
c006a95df1
Fix example server install instructions ( #12362 )
...
* Upgrade huggingface-hub to version 0.35.0
Updated huggingface-hub version from 0.26.1 to 0.35.0.
* Add uvicorn and accelerate to requirements
* Fix install instructions for server
2025-09-22 08:37:17 +05:30
naykun
df267ee4e8
feat: Add QwenImageEditPlus to support future feature upgrades ( #12357 )
...
* feat: add support of qwenimageeditplus
* add copies statement
* fix copies statement
* remove vl_processor reference
2025-09-21 06:10:52 -10:00
Dhruv Nair
edd614ea38
[CI] Fix TRANSFORMERS_FLAX_WEIGHTS_NAME import issue ( #12354 )
...
update
2025-09-20 09:01:40 +05:30
Dave Lage
7e7e62c6ff
Convert alphas for embedders for sd-scripts to ai toolkit conversion ( #12332 )
...
* Convert alphas for embedders for sd-scripts to ai toolkit conversion
* Add kohya embedders conversion test
* Apply style fixes
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-09-18 12:41:22 +05:30
Fredy
eda9ff8300
Add RequestScopedPipeline for safe concurrent inference, tokenizer lock and non-mutating retrieve_timesteps ( #12328 )
...
* Basic implementation of request scheduling
* Basic editing in SD and Flux Pipelines
* Small Fix
* Fix
* Update for more pipelines
* Add examples/server-async
* Add examples/server-async
* Updated RequestScopedPipeline to handle a single tokenizer lock to avoid race conditions
* Fix
* Fix _TokenizerLockWrapper
* Fix _TokenizerLockWrapper
* Delete _TokenizerLockWrapper
* Fix tokenizer
* Update examples/server-async
* Fix server-async
* Optimizations in examples/server-async
* We keep the implementation simple in examples/server-async
* Update examples/server-async/README.md
* Update examples/server-async/README.md for changes to tokenizer locks and backward-compatible retrieve_timesteps
* The changes to the diffusers core have been undone and all logic is being moved to exmaples/server-async
* Update examples/server-async/utils/*
* Fix BaseAsyncScheduler
* Rollback in the core of the diffusers
* Update examples/server-async/README.md
* Complete rollback of diffusers core files
* Simple implementation of an asynchronous server compatible with SD3-3.5 and Flux Pipelines
* Update examples/server-async/README.md
* Fixed import errors in 'examples/server-async/serverasync.py'
* Flux Pipeline Discard
* Update examples/server-async/README.md
* Apply style fixes
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-09-18 11:33:43 +05:30
DefTruth
efb7a299af
Fix many type hint errors ( #12289 )
...
* fix hidream type hint
* fix hunyuan-video type hint
* fix many type hint
* fix many type hint errors
* fix many type hint errors
* fix many type hint errors
* make stype & make quality
2025-09-16 18:52:15 -10:00
Zijian Zhou
d06750a5fd
Fix autoencoder_kl_wan.py bugs for Wan2.2 VAE ( #12335 )
...
* Update autoencoder_kl_wan.py
When using the Wan2.2 VAE, the spatial compression ratio calculated here is incorrect. It should be 16 instead of 8. Pass it in directly via the config to ensure it’s correct here.
* Update autoencoder_kl_wan.py
2025-09-16 13:43:15 -10:00
Sari Hleihil
8c72cd12ee
Added LucyEditPipeline ( #12340 )
...
* Added LucyEditPipeline
* add import & stype
missing copied from
* Fix example doc string
---------
Co-authored-by: yiyixuxu <yixu310@gmail.com >
2025-09-16 13:41:05 -10:00
Samarth Agrawal
751e250f70
fixed bug in defining embed dim for UNet1D ( #12111 )
...
* fixed bug in defining embed dim
* matched 1d temb process to 2d
* Update src/diffusers/models/unets/unet_1d.py
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2025-09-16 12:18:48 +05:30
Linoy Tsaban
b50014067d
Add Wan2.2 VACE - Fun ( #12324 )
...
* support Wan2.2-VACE-Fun-A14B
* support Wan2.2-VACE-Fun-A14B
* support Wan2.2-VACE-Fun-A14B
* Apply style fixes
* test
---------
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-09-15 21:31:26 +05:30
Daniel Socek
f5c113e439
Use SDP on BF16 in GPU/HPU migration ( #12310 )
...
* Use SDP on BF16 in GPU/HPU migration
Signed-off-by: Daniel Socek <daniel.socek@intel.com >
* Formatting fix for enabling SDP with BF16 precision on HPU
Signed-off-by: Daniel Socek <daniel.socek@intel.com >
---------
Signed-off-by: Daniel Socek <daniel.socek@intel.com >
2025-09-12 08:00:36 -10:00
Sayak Paul
5e181eddfe
Deprecate slicing and tiling methods from DiffusionPipeline ( #12271 )
...
* deprecate slicing from flux pipeline.
* propagate.
* tiling
* up
* up
2025-09-11 10:04:35 +05:30
Justin Ruan
55f0b3d758
Fix AttributeError of VisualClozeProcessor ( #12121 )
...
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2025-09-11 04:17:34 +05:30
Sayak Paul
eb7ef26736
[quant] allow components_to_quantize to be a non-list for single components ( #12234 )
...
* allow non list components_to_quantize.
* up
* Apply suggestions from code review
* Apply suggestions from code review
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* [docs] components_to_quantize (#12287 )
init
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
---------
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
2025-09-10 09:47:08 -10:00
ttio2tech
e1b7f1f240
fix for the qwen controlnet pipeline - wrong device can be used ( #12309 )
...
fix the device for textencoder
2025-09-10 08:59:08 -10:00
Sayak Paul
9e7ae568d6
[feat] cache allocator warmup for from_single_model ( #12305 )
...
* add
* add a test
2025-09-10 12:55:32 +05:30
Sayak Paul
f7b79452b4
[modular] fix flux modular pipelines for t2i and i2i ( #12272 )
...
fix flux modular pipelines for t2i and i2i
2025-09-10 12:39:55 +05:30
Sayak Paul
43459079ab
[core] feat: support group offloading at the pipeline level ( #12283 )
...
* feat: support group offloading at the pipeline level.
* add tests
* up
* [docs] Pipeline group offloading (#12286 )
init
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
---------
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
2025-09-10 09:09:57 +05:30