yiyixuxu
|
462429b687
|
remove modular reelated change from pipelines folder
|
2025-05-10 03:50:10 +02:00 |
|
yiyixuxu
|
cf01aaeb49
|
update imports on guiders
|
2025-05-10 03:49:30 +02:00 |
|
yiyixuxu
|
2017ae5624
|
fix auto denoise so all tests pass
|
2025-05-09 08:19:24 +02:00 |
|
yiyixuxu
|
2b361a2413
|
fix get_execusion blocks with loopsequential
|
2025-05-09 08:17:10 +02:00 |
|
yiyixuxu
|
c677d528e4
|
change warning to debug
|
2025-05-09 08:16:24 +02:00 |
|
yiyixuxu
|
0f0618ff2b
|
refactor the denoiseestep using LoopSequential! also add a new file for denoise step
|
2025-05-08 11:28:52 +02:00 |
|
yiyixuxu
|
d89631fc50
|
update input formating, consider kwarggs_type inputs with no name, e/g *_controlnet_kwargs
|
2025-05-08 11:27:17 +02:00 |
|
yiyixuxu
|
16b6583fa8
|
allow input_fields as input & update message
|
2025-05-08 11:25:31 +02:00 |
|
yiyixuxu
|
f552773572
|
remove controlnet union denoise step, refactor & reuse controlnet denoisee step to accept aditional contrlnet kwargs
|
2025-05-06 10:00:14 +02:00 |
|
yiyixuxu
|
dc4dbfe107
|
reefactor pipeline/block states so that it can dynamically accept kwargs
|
2025-05-06 09:58:44 +02:00 |
|
yiyixuxu
|
43ac1ff7e7
|
refactor controlnet union
|
2025-05-04 22:17:25 +02:00 |
|
yiyixuxu
|
efd70b7838
|
seperate controlnet step into input + denoise
|
2025-05-03 20:22:05 +02:00 |
|
yiyixuxu
|
7ca860c24b
|
rename pipeline -> components, data -> block_state
|
2025-05-03 01:32:59 +02:00 |
|
yiyixuxu
|
7b86fcea31
|
remove lora step and ip-adapter step -> no longer needed
|
2025-05-02 11:31:25 +02:00 |
|
yiyixuxu
|
c8b5d56412
|
make loader optional
|
2025-05-02 00:46:31 +02:00 |
|
YiYi Xu
|
ce642e92da
|
Merge branch 'modular-diffusers' into modular-refactor
|
2025-04-30 17:56:51 -10:00 |
|
YiYi Xu
|
6a509ba862
|
Merge branch 'main' into modular-diffusers
|
2025-04-30 17:56:25 -10:00 |
|
YiYi Xu
|
6d5beefe29
|
[modular diffusers] introducing ModularLoader (#11462)
* cfg; slg; pag; sdxl without controlnet
---------
Co-authored-by: Aryan <aryan@huggingface.co>
|
2025-04-30 11:17:20 -10:00 |
|
Yao Matrix
|
06beecafc5
|
make autoencoders. controlnet_flux and wan_transformer3d_single_file pass on xpu (#11461)
* make autoencoders. controlnet_flux and wan_transformer3d_single_file
pass on XPU
Signed-off-by: Yao Matrix <matrix.yao@intel.com>
* Apply style fixes
---------
Signed-off-by: Yao Matrix <matrix.yao@intel.com>
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
Co-authored-by: Aryan <aryan@huggingface.co>
|
2025-05-01 02:43:31 +05:30 |
|
Daniel Socek
|
fbce7aeb32
|
Add generic support for Intel Gaudi accelerator (hpu device) (#11328)
* Add generic support for Intel Gaudi accelerator (hpu device)
Signed-off-by: Daniel Socek <daniel.socek@intel.com>
Co-authored-by: Libin Tang <libin.tang@intel.com>
* Add loggers for generic HPU support
Signed-off-by: Daniel Socek <daniel.socek@intel.com>
* Refactor hpu support with is_hpu_available() logic
Signed-off-by: Daniel Socek <daniel.socek@intel.com>
* Fix style for hpu support update
Signed-off-by: Daniel Socek <daniel.socek@intel.com>
* Decouple soft HPU check from hard device validation to support HPU migration
Signed-off-by: Daniel Socek <daniel.socek@intel.com>
---------
Signed-off-by: Daniel Socek <daniel.socek@intel.com>
Co-authored-by: Libin Tang <libin.tang@intel.com>
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
|
2025-04-30 14:45:02 +05:30 |
|
Aryan
|
c86511586f
|
torch.compile fullgraph compatibility for Hunyuan Video (#11457)
udpate
|
2025-04-30 11:21:17 +05:30 |
|
Aryan
|
8fe5a14d9b
|
Raise warning instead of error for block offloading with streams (#11425)
raise warning instead of error
|
2025-04-30 08:26:16 +05:30 |
|
urpetkov-amd
|
4a9ab650aa
|
Fixing missing provider options argument (#11397)
* Fixing missing provider options argument
* Adding if else for provider options
* Apply suggestions from code review
Co-authored-by: YiYi Xu <yixu310@gmail.com>
* Apply style fixes
* Update src/diffusers/pipelines/onnx_utils.py
Co-authored-by: YiYi Xu <yixu310@gmail.com>
* Update src/diffusers/pipelines/onnx_utils.py
Co-authored-by: YiYi Xu <yixu310@gmail.com>
---------
Co-authored-by: Uros Petkovic <urpektov@amd.com>
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
Co-authored-by: YiYi Xu <yixu310@gmail.com>
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
|
2025-04-28 10:23:05 -10:00 |
|
Yao Matrix
|
7567adfc45
|
enable 28 GGUF test cases on XPU (#11404)
* enable gguf test cases on XPU
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
* make SD35LargeGGUFSingleFileTests::test_pipeline_inference pas
Signed-off-by: root <root@a4bf01945cfe.jf.intel.com>
* make FluxControlLoRAGGUFTests::test_lora_loading pass
Signed-off-by: Yao Matrix <matrix.yao@intel.com>
* polish code
Signed-off-by: Yao Matrix <matrix.yao@intel.com>
* Apply style fixes
---------
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
Signed-off-by: root <root@a4bf01945cfe.jf.intel.com>
Signed-off-by: Yao Matrix <matrix.yao@intel.com>
Co-authored-by: root <root@a4bf01945cfe.jf.intel.com>
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
|
2025-04-28 21:32:04 +05:30 |
|
Yao Matrix
|
a7e9f85e21
|
enable test_layerwise_casting_memory cases on XPU (#11406)
* enable test_layerwise_casting_memory cases on XPU
Signed-off-by: Yao Matrix <matrix.yao@intel.com>
* fix style
Signed-off-by: Yao Matrix <matrix.yao@intel.com>
---------
Signed-off-by: Yao Matrix <matrix.yao@intel.com>
|
2025-04-28 06:38:39 +05:30 |
|
Aryan
|
b863bdd6ca
|
Modular Diffusers Guiders (#11311)
* cfg; slg; pag; sdxl without controlnet
* support sdxl controlnet
* support controlnet union
* update
* update
* cfg zero*
* use unwrap_module for torch compiled modules
* remove guider kwargs
* remove commented code
* remove old guider
* fix slg bug
* remove debug print
* autoguidance
* smoothed energy guidance
* add note about seg
* tangential cfg
* cfg plus plus
* support cfgpp in ddim
* apply review suggestions
* refactor
* rename enable/disable
* remove cfg++ for now
* rename do_classifier_free_guidance->prepare_unconditional_embeds
* remove unused
|
2025-04-26 03:42:42 +05:30 |
|
co63oc
|
f00a995753
|
Fix typos in strings and comments (#11407)
|
2025-04-24 08:53:47 -10:00 |
|
Ishan Modi
|
e8312e7ca9
|
[BUG] fixed WAN docstring (#11226)
update
|
2025-04-24 08:49:37 -10:00 |
|
Teriks
|
b4be42282d
|
Kolors additional pipelines, community contrib (#11372)
* Kolors additional pipelines, community contrib
---------
Co-authored-by: Teriks <Teriks@users.noreply.github.com>
Co-authored-by: Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com>
|
2025-04-23 11:07:27 -10:00 |
|
Ishan Modi
|
a4f9c3cbc3
|
[Feature] Added Xlab Controlnet support (#11249)
update
|
2025-04-23 10:43:50 -10:00 |
|
Aryan
|
6cef71de3a
|
Fix group offloading with block_level and use_stream=True (#11375)
* fix
* add tests
* add message check
|
2025-04-23 18:17:53 +05:30 |
|
YiYi Xu
|
448c72a230
|
[HiDream] move deprecation to 0.35.0 (#11384)
up
|
2025-04-22 08:08:08 -10:00 |
|
Aryan
|
f108ad8888
|
Update modeling imports (#11129)
update
|
2025-04-22 06:59:25 -10:00 |
|
Linoy Tsaban
|
e30d3bf544
|
[LoRA] add LoRA support to HiDream and fine-tuning script (#11281)
* initial commit
* initial commit
* initial commit
* initial commit
* initial commit
* initial commit
* Update examples/dreambooth/train_dreambooth_lora_hidream.py
Co-authored-by: Bagheera <59658056+bghira@users.noreply.github.com>
* move prompt embeds, pooled embeds outside
* Update examples/dreambooth/train_dreambooth_lora_hidream.py
Co-authored-by: hlky <hlky@hlky.ac>
* Update examples/dreambooth/train_dreambooth_lora_hidream.py
Co-authored-by: hlky <hlky@hlky.ac>
* fix import
* fix import and tokenizer 4, text encoder 4 loading
* te
* prompt embeds
* fix naming
* shapes
* initial commit to add HiDreamImageLoraLoaderMixin
* fix init
* add tests
* loader
* fix model input
* add code example to readme
* fix default max length of text encoders
* prints
* nullify training cond in unpatchify for temp fix to incompatible shaping of transformer output during training
* smol fix
* unpatchify
* unpatchify
* fix validation
* flip pred and loss
* fix shift!!!
* revert unpatchify changes (for now)
* smol fix
* Apply style fixes
* workaround moe training
* workaround moe training
* remove prints
* to reduce some memory, keep vae in `weight_dtype` same as we have for flux (as it's the same vae)
bbd0c161b5/examples/dreambooth/train_dreambooth_lora_flux.py (L1207)
* refactor to align with HiDream refactor
* refactor to align with HiDream refactor
* refactor to align with HiDream refactor
* add support for cpu offloading of text encoders
* Apply style fixes
* adjust lr and rank for train example
* fix copies
* Apply style fixes
* update README
* update README
* update README
* fix license
* keep prompt2,3,4 as None in validation
* remove reverse ode comment
* Update examples/dreambooth/train_dreambooth_lora_hidream.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
* Update examples/dreambooth/train_dreambooth_lora_hidream.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
* vae offload change
* fix text encoder offloading
* Apply style fixes
* cleaner to_kwargs
* fix module name in copied from
* add requirements
* fix offloading
* fix offloading
* fix offloading
* update transformers version in reqs
* try AutoTokenizer
* try AutoTokenizer
* Apply style fixes
* empty commit
* Delete tests/lora/test_lora_layers_hidream.py
* change tokenizer_4 to load with AutoTokenizer as well
* make text_encoder_four and tokenizer_four configurable
* save model card
* save model card
* revert T5
* fix test
* remove non diffusers lumina2 conversion
---------
Co-authored-by: Bagheera <59658056+bghira@users.noreply.github.com>
Co-authored-by: hlky <hlky@hlky.ac>
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
|
2025-04-22 11:44:02 +03:00 |
|
apolinário
|
6ab62c7431
|
Add stochastic sampling to FlowMatchEulerDiscreteScheduler (#11369)
* Add stochastic sampling to FlowMatchEulerDiscreteScheduler
This PR adds stochastic sampling to FlowMatchEulerDiscreteScheduler based on b1aeddd7cc ltx_video/schedulers/rf.py
* Apply style fixes
* Use config value directly
* Apply style fixes
* Swap order
* Update src/diffusers/schedulers/scheduling_flow_match_euler_discrete.py
Co-authored-by: YiYi Xu <yixu310@gmail.com>
* Update src/diffusers/schedulers/scheduling_flow_match_euler_discrete.py
Co-authored-by: YiYi Xu <yixu310@gmail.com>
---------
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
Co-authored-by: YiYi Xu <yixu310@gmail.com>
|
2025-04-21 17:18:30 -10:00 |
|
Ishan Modi
|
f59df3bb8b
|
[Refactor] Minor Improvement for import utils (#11161)
* update
* update
* addressed PR comments
* update
---------
Co-authored-by: YiYi Xu <yixu310@gmail.com>
|
2025-04-21 09:56:55 -10:00 |
|
josephrocca
|
a00c73a5e1
|
Support different-length pos/neg prompts for FLUX.1-schnell variants like Chroma (#11120)
Co-authored-by: YiYi Xu <yixu310@gmail.com>
|
2025-04-21 09:28:19 -10:00 |
|
OleehyO
|
0434db9a99
|
[cogview4][feat] Support attention mechanism with variable-length support and batch packing (#11349)
* [cogview4] Enhance attention mechanism with variable-length support and batch packing
---------
Co-authored-by: YiYi Xu <yixu310@gmail.com>
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
|
2025-04-21 09:27:55 -10:00 |
|
Aamir Nazir
|
aff574fb29
|
Add Serialized Type Name kwarg in Model Output (#10502)
* Update outputs.py
|
2025-04-21 08:45:28 -10:00 |
|
Ishan Modi
|
79ea8eb258
|
[BUG] fixes in kadinsky pipeline (#11080)
* bug fix kadinsky pipeline
|
2025-04-21 08:41:09 -10:00 |
|
Aryan
|
e7f3a73786
|
Fix Wan I2V prepare_latents dtype (#11371)
update
|
2025-04-21 08:18:50 -10:00 |
|
YiYi Xu
|
5a2e0f715c
|
update output for Hidream transformer (#11366)
up
|
2025-04-18 14:07:21 -10:00 |
|
YiYi Xu
|
0021bfa1e1
|
support Wan-FLF2V (#11353)
* update transformer
---------
Co-authored-by: Aryan <aryan@huggingface.co>
|
2025-04-18 10:27:50 -10:00 |
|
Marc Sun
|
bbd0c161b5
|
[BNB] Fix test_moving_to_cpu_throws_warning (#11356)
fix
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
|
2025-04-18 09:44:51 +05:30 |
|
Frank (Haofan) Wang
|
ee6ad51d96
|
Update controlnet_flux.py (#11350)
|
2025-04-17 10:05:01 -10:00 |
|
Sayak Paul
|
4397f59a37
|
[bitsandbytes] improve dtype mismatch handling for bnb + lora. (#11270)
* improve dtype mismatch handling for bnb + lora.
* add a test
* fix and updates
* update
|
2025-04-17 19:51:49 +05:30 |
|
YiYi Xu
|
056793295c
|
[Hi Dream] follow-up (#11296)
* add
|
2025-04-17 01:17:44 -10:00 |
|
Sayak Paul
|
29d2afbfe2
|
[LoRA] Propagate hotswap better (#11333)
* propagate hotswap to other load_lora_weights() methods.
* simplify documentations.
* updates
* propagate to load_lora_into_text_encoder.
* empty commit
|
2025-04-17 10:35:38 +05:30 |
|
nPeppon
|
3e59d531d1
|
Fix wrong dtype argument name as torch_dtype (#11346)
|
2025-04-16 16:00:25 -04:00 |
|
Dhruv Nair
|
59f1b7b1c8
|
Hunyuan I2V fast tests fix (#11341)
* update
* update
|
2025-04-16 18:40:33 +05:30 |
|