Aryan
f859fdf7ba
refactor; support flash attention 2 with cp
2025-07-16 21:32:27 +02:00
Aryan
79736265c5
refactor
2025-07-16 21:24:24 +02:00
Aryan
51fed50837
update
2025-07-16 17:19:30 +02:00
Aryan
84d2c84ea4
Merge branch 'main' into to-single-file/flux
2025-07-16 16:22:18 +05:30
Aryan
251bb61925
minify and deprecate npu/xla processors
2025-07-16 12:51:27 +02:00
Aryan
1e6b1c51a8
remove rmsnorm assert
2025-07-16 12:51:21 +02:00
Sayak Paul
9c13f86579
[training] add an offload utility that can be used as a context manager. ( #11775 )
...
* add an offload utility that can be used as a context manager.
* update
---------
Co-authored-by: Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com >
2025-07-16 09:09:13 +01:00
G.O.D
5c5209720e
enable flux pipeline compatible with unipc and dpm-solver ( #11908 )
...
* Update pipeline_flux.py
have flux pipeline work with unipc/dpm schedulers
* clean code
* Update scheduling_dpmsolver_multistep.py
* Update scheduling_unipc_multistep.py
* Update pipeline_flux.py
* Update scheduling_deis_multistep.py
* Update scheduling_dpmsolver_singlestep.py
* Apply style fixes
---------
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
Co-authored-by: YiYi Xu <yixu310@gmail.com >
Co-authored-by: Álvaro Somoza <asomoza@users.noreply.github.com >
2025-07-15 17:49:57 -10:00
Álvaro Somoza
aa14f090f8
[ControlnetUnion] Propagate #11888 to img2img ( #11929 )
...
img2img fixes
2025-07-15 21:41:35 -04:00
Guoqing Zhu
c5d6e0b537
Fixed bug: Uncontrolled recursive calls that caused an infinite loop when loading certain pipelines containing Transformer2DModel ( #11923 )
...
* fix a bug about loop call
* fix a bug about loop call
* ruff format
---------
Co-authored-by: Álvaro Somoza <asomoza@users.noreply.github.com >
2025-07-15 14:58:37 -10:00
lostdisc
39831599f1
Remove forced float64 from onnx stable diffusion pipelines ( #11054 )
...
* Update pipeline_onnx_stable_diffusion.py to remove float64
init_noise_sigma was being set as float64 before multiplying with latents, which changed latents into float64 too, which caused errors with onnxruntime since the latter wanted float16.
* Update pipeline_onnx_stable_diffusion_inpaint.py to remove float64
init_noise_sigma was being set as float64 before multiplying with latents, which changed latents into float64 too, which caused errors with onnxruntime since the latter wanted float16.
* Update pipeline_onnx_stable_diffusion_upscale.py to remove float64
init_noise_sigma was being set as float64 before multiplying with latents, which changed latents into float64 too, which caused errors with onnxruntime since the latter wanted float16.
* Update pipeline_onnx_stable_diffusion.py with comment for previous commit
Added comment on purpose of init_noise_sigma. This comment exists in related scripts that use the same line of code, but it was missing here.
---------
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2025-07-15 14:57:28 -10:00
Aryan
b73c738392
Remove device synchronization when loading weights ( #11927 )
...
* update
* make style
2025-07-15 21:40:57 +05:30
Aryan
a73cb396ba
refactor chroma as well
2025-07-15 13:53:27 +02:00
Aryan
d9c1683b07
make ip adapter processor compatible with attention dispatcher
2025-07-15 13:30:52 +02:00
Aryan
4f52e3499c
make fix-copies
2025-07-15 12:30:45 +02:00
Aryan
1e7217f82d
make style
2025-07-15 12:30:34 +02:00
Aryan
e909b7355f
update
2025-07-15 12:30:07 +02:00
Aryan
576da52f45
update
2025-07-15 12:29:57 +02:00
Aryan
4dcd672907
add back set_attention_backend
2025-07-15 12:29:28 +02:00
Aryan
c141520714
fight more tests
2025-07-15 09:49:17 +02:00
Aryan
a0b276da53
fix more tests
2025-07-15 08:30:26 +02:00
Aryan
bc64f12c98
fix fastercache implementation
2025-07-15 08:01:42 +02:00
Aryan
0cda91d467
fix chroma qkv fusion test
2025-07-15 07:51:58 +02:00
Aryan
17b678fc6f
Merge branch 'main' into to-single-file/flux
2025-07-15 10:47:58 +05:30
Aryan
06fd427797
[tests] Improve Flux tests ( #11919 )
...
update
2025-07-15 10:47:41 +05:30
dependabot[bot]
48a551251d
Bump aiohttp from 3.10.10 to 3.12.14 in /examples/server ( #11924 )
...
Bumps [aiohttp](https://github.com/aio-libs/aiohttp ) from 3.10.10 to 3.12.14.
- [Release notes](https://github.com/aio-libs/aiohttp/releases )
- [Changelog](https://github.com/aio-libs/aiohttp/blob/master/CHANGES.rst )
- [Commits](https://github.com/aio-libs/aiohttp/compare/v3.10.10...v3.12.14 )
---
updated-dependencies:
- dependency-name: aiohttp
dependency-version: 3.12.14
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2025-07-15 09:15:57 +05:30
Hengyue-Bi
6398fbc391
Fix: Align VAE processing in ControlNet SD3 training with inference ( #11909 )
...
Fix: Apply vae_shift_factor in ControlNet SD3 training
2025-07-14 14:54:38 -04:00
Aryan
b8f7fe61e1
handle ip adapter params correctly
2025-07-14 08:21:47 +02:00
Aryan
ff21b7fe8b
improve test
2025-07-14 07:46:32 +02:00
Aryan
ecabd2a46e
add coauthor
...
Co-Authored-By: Dhruv Nair <dhruv.nair@gmail.com >
2025-07-14 07:38:41 +02:00
Aryan
7e97e43efc
update
2025-07-14 04:56:55 +02:00
Aryan
d7b9e42958
update
2025-07-14 04:47:53 +02:00
Colle
3c8b67b371
Flux: pass joint_attention_kwargs when using gradient_checkpointing ( #11814 )
...
Flux: pass joint_attention_kwargs when gradient_checkpointing
2025-07-11 08:35:18 -10:00
Steven Liu
9feb946432
[docs] torch.compile blog post ( #11837 )
...
* add blog post
* feedback
* feedback
2025-07-11 10:29:40 -07:00
Aryan
c90352754a
Speedup model loading by 4-5x ⚡ ( #11904 )
...
* update
* update
* update
* pin accelerate version
* add comment explanations
* update docstring
* make style
* non_blocking does not matter for dtype cast
* _empty_cache -> clear_cache
* update
* Update src/diffusers/models/model_loading_utils.py
Co-authored-by: Marc Sun <57196510+SunMarc@users.noreply.github.com >
* Update src/diffusers/models/model_loading_utils.py
---------
Co-authored-by: Marc Sun <57196510+SunMarc@users.noreply.github.com >
2025-07-11 21:43:53 +05:30
Sayak Paul
7a935a0bbe
[tests] Unify compilation + offloading tests in quantization ( #11910 )
...
* unify the quant compile + offloading tests.
* fix
* update
2025-07-11 17:02:29 +05:30
chenxiao
941b7fc084
Avoid creating tensor in CosmosAttnProcessor2_0 ( #11761 ) ( #11763 )
...
* Avoid creating tensor in CosmosAttnProcessor2_0 (#11761 )
* up
---------
Co-authored-by: yiyixuxu <yixu310@gmail.com >
2025-07-10 11:51:05 -10:00
Álvaro Somoza
76a62ac9cc
[ControlnetUnion] Multiple Fixes ( #11888 )
...
fixes
---------
Co-authored-by: hlky <hlky@hlky.ac >
2025-07-10 14:35:28 -04:00
Sayak Paul
1c6ab9e900
[utils] account for MPS when available in get_device(). ( #11905 )
...
* account for MPS when available in get_device().
* fix
2025-07-10 13:30:54 +05:30
Sayak Paul
265840a098
[LoRA] fix: disabling hooks when loading loras. ( #11896 )
...
fix: disabling hooks when loading loras.
2025-07-10 10:30:10 +05:30
dependabot[bot]
9f4d997d8f
Bump torch from 2.4.1 to 2.7.0 in /examples/server ( #11429 )
...
Bumps [torch](https://github.com/pytorch/pytorch ) from 2.4.1 to 2.7.0.
- [Release notes](https://github.com/pytorch/pytorch/releases )
- [Changelog](https://github.com/pytorch/pytorch/blob/main/RELEASE.md )
- [Commits](https://github.com/pytorch/pytorch/compare/v2.4.1...v2.7.0 )
---
updated-dependencies:
- dependency-name: torch
dependency-version: 2.7.0
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2025-07-10 09:24:10 +05:30
Sayak Paul
b41abb2230
[quant] QoL improvements for pipeline-level quant config ( #11876 )
...
* add repr for pipelinequantconfig.
* update
2025-07-10 08:53:01 +05:30
YiYi Xu
f33b89bafb
The Modular Diffusers ( #9672 )
...
adding modular diffusers as experimental feature
---------
Co-authored-by: hlky <hlky@hlky.ac >
Co-authored-by: Álvaro Somoza <asomoza@users.noreply.github.com >
Co-authored-by: Aryan <aryan@huggingface.co >
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-07-09 16:00:28 -10:00
Álvaro Somoza
48a6d29550
[SD3] CFG Cutoff fix and official callback ( #11890 )
...
fix and official callback
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2025-07-09 14:31:11 -04:00
Sayak Paul
2d3d376bc0
Fix unique memory address when doing group-offloading with disk ( #11767 )
...
* fix memory address problem
* add more tests
* updates
* updates
* update
* _group_id = group_id
* update
* Apply suggestions from code review
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
* update
* update
* update
* fix
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2025-07-09 21:29:34 +05:30
Sébastien Iooss
db715e2c8c
feat: add multiple input image support in Flux Kontext ( #11880 )
...
* feat: add multiple input image support in Flux Kontext
* move model to community
* fix linter
2025-07-09 11:09:59 -04:00
Sayak Paul
754fe85cac
[tests] add compile + offload tests for GGUF. ( #11740 )
...
* add compile + offload tests for GGUF.
* quality
* add init.
* prop.
* change to flux.
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2025-07-09 13:42:13 +05:30
Sayak Paul
cc1f9a2ce3
[tests] mark the wanvace lora tester flaky ( #11883 )
...
* mark wanvace lora tests as flaky
* ability to apply is_flaky at a class-level
* update
* increase max_attempt.
* increase attemtp.
2025-07-09 13:27:15 +05:30
Sayak Paul
737d7fc3b0
[tests] Remove more deprecated tests ( #11895 )
...
* remove k diffusion tests
* remove script
2025-07-09 13:10:44 +05:30
Sayak Paul
be23f7df00
[Docker] update doc builder dockerfile to include quant libs. ( #11728 )
...
update doc builder dockerfile to include quant libs.
2025-07-09 12:27:22 +05:30