Thien Tran
3c2e2aa8a9
.from_single_file() - Add missing .shape (#10332 )
...
Add missing `.shape`
2024-12-23 08:57:25 +05:30
Junsong Chen
b58868e6f4
[Sana bug] bug fix for 2K model config ( #10340 )
...
* fix the Positinoal Embedding bug in 2K model;
* Change the default model to the BF16 one for more stable training and output
* make style
* substract buffer size
* add compute_module_persistent_sizes
---------
Co-authored-by: yiyixuxu <yixu310@gmail.com >
2024-12-23 08:56:25 +05:30
Dhruv Nair
da21d590b5
[Single File] Add Single File support for HunYuan video ( #10320 )
...
* update
* Update src/diffusers/loaders/single_file_utils.py
Co-authored-by: Aryan <aryan@huggingface.co >
---------
Co-authored-by: Aryan <aryan@huggingface.co >
2024-12-23 08:44:58 +05:30
YiYi Xu
7c2f0afb1c
update get_parameter_dtype ( #10342 )
...
add:
q
2024-12-23 08:14:13 +05:30
hlky
f615f00f58
Fix enable_sequential_cpu_offload in test_kandinsky_combined ( #10324 )
...
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-12-22 15:28:28 -10:00
Aryan
6aaa0518e3
Community hosted weights for diffusers format HunyuanVideo weights ( #10344 )
...
update docs and example to use community weights
2024-12-22 15:26:28 -10:00
Mehmet Yiğit Özgenç
233dffdc3f
flux controlnet inpaint config bug ( #10291 )
...
* flux controlnet inpaint config bug
* Update src/diffusers/pipelines/flux/pipeline_flux_controlnet_inpainting.py
---------
Co-authored-by: yigitozgenc <yigit@quantuslabs.ai >
Co-authored-by: hlky <hlky@hlky.ac >
2024-12-21 18:44:43 +00:00
hlky
be2070991f
Support Flux IP Adapter ( #10261 )
...
* Flux IP-Adapter
* test cfg
* make style
* temp remove copied from
* fix test
* fix test
* v2
* fix
* make style
* temp remove copied from
* Apply suggestions from code review
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Move encoder_hid_proj to inside FluxTransformer2DModel
* merge
* separate encode_prompt, add copied from, image_encoder offload
* make
* fix test
* fix
* Update src/diffusers/pipelines/flux/pipeline_flux.py
* test_flux_prompt_embeds change not needed
* true_cfg -> true_cfg_scale
* fix merge conflict
* test_flux_ip_adapter_inference
* add fast test
* FluxIPAdapterMixin not test mixin
* Update pipeline_flux.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
---------
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2024-12-21 17:49:58 +00:00
hlky
bf9a641f1a
Fix EMAModel test_from_pretrained ( #10325 )
2024-12-21 14:10:44 +00:00
hlky
a756694bf0
Fix push_tests_mps.yml ( #10326 )
2024-12-21 14:10:32 +00:00
Sayak Paul
d41388145e
[Docs] Update gguf.md to remove generator from the pipeline from_pretrained ( #10299 )
...
Update gguf.md to remove generator from the pipeline from_pretrained
2024-12-21 07:15:03 +05:30
Junsong Chen
a6288a5571
[Sana]add 2K related model for Sana ( #10322 )
...
add 2K related model for Sana
2024-12-20 07:21:34 -10:00
Steven Liu
7d4db57037
[docs] Fix quantization links ( #10323 )
...
Update overview.md
2024-12-20 08:30:21 -08:00
Aditya Raj
902008608a
[BUG FIX] [Stable Audio Pipeline] Resolve torch.Tensor.new_zeros() TypeError in function prepare_latents caused by audio_vae_length ( #10306 )
...
[BUG FIX] [Stable Audio Pipeline] TypeError: new_zeros(): argument 'size' failed to unpack the object at pos 3 with error "type must be tuple of ints,but got float"
torch.Tensor.new_zeros() takes a single argument size (int...) – a list, tuple, or torch.Size of integers defining the shape of the output tensor.
in function prepare_latents:
audio_vae_length = self.transformer.config.sample_size * self.vae.hop_length
audio_shape = (batch_size // num_waveforms_per_prompt, audio_channels, audio_vae_length)
...
audio = initial_audio_waveforms.new_zeros(audio_shape)
audio_vae_length evaluates to float because self.transformer.config.sample_size returns a float
Co-authored-by: hlky <hlky@hlky.ac >
2024-12-20 15:29:58 +00:00
Leojc
c8ee4af228
docs: fix a mistake in docstring ( #10319 )
...
Update pipeline_hunyuan_video.py
docs: fix a mistake
2024-12-20 15:22:32 +00:00
Sayak Paul
b64ca6c11c
[Docs] Update ltx_video.md to remove generator from from_pretrained() ( #10316 )
...
Update ltx_video.md to remove generator from `from_pretrained()`
2024-12-20 18:32:22 +05:30
Dhruv Nair
e12d610faa
Mochi docs ( #9934 )
...
* update
* update
* update
* update
* update
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-12-20 16:27:38 +05:30
Sayak Paul
bf6eaa8aec
[Tests] add integration tests for lora expansion stuff in Flux. ( #10318 )
...
add integration tests for lora expansion stuff in Flux.
2024-12-20 16:14:58 +05:30
Sayak Paul
17128c42a4
[LoRA] feat: support loading regular Flux LoRAs into Flux Control, and Fill ( #10259 )
...
* lora expansion with dummy zeros.
* updates
* fix working 🥳
* working.
* use torch.device meta for state dict expansion.
* tests
Co-authored-by: a-r-r-o-w <contact.aryanvs@gmail.com >
* fixes
* fixes
* switch to debug
* fix
* Apply suggestions from code review
Co-authored-by: Aryan <aryan@huggingface.co >
* fix stuff
* docs
---------
Co-authored-by: a-r-r-o-w <contact.aryanvs@gmail.com >
Co-authored-by: Aryan <aryan@huggingface.co >
2024-12-20 14:30:32 +05:30
Dhruv Nair
dbc1d505f0
[Single File] Add GGUF support for LTX ( #10298 )
...
* update
* add docs.
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-12-20 11:52:29 +05:30
Aryan
151b74cd77
Make tensors in ResNet contiguous for Hunyuan VAE ( #10309 )
...
contiguous tensors in resnet
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2024-12-20 11:45:37 +05:30
Aryan
41ba8c0bf6
Add support for sharded models when TorchAO quantization is enabled ( #10256 )
...
* add sharded + device_map check
2024-12-19 15:42:20 -10:00
Daniel Regado
3191248472
[WIP] SD3.5 IP-Adapter Pipeline Integration ( #9987 )
...
* Added support for single IPAdapter on SD3.5 pipeline
---------
Co-authored-by: hlky <hlky@hlky.ac >
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2024-12-19 14:48:18 -10:00
dg845
648d968cfc
Enable Gradient Checkpointing for UNet2DModel (New) ( #7201 )
...
* Port UNet2DModel gradient checkpointing code from #6718 .
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: Vincent Neemie <92559302+VincentNeemie@users.noreply.github.com >
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com >
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
Co-authored-by: hlky <hlky@hlky.ac >
2024-12-19 14:45:45 -10:00
djm
b756ec6e80
unet's sample_size attribute is to accept tuple(h, w) in StableDiffusionPipeline ( #10181 )
2024-12-19 22:24:18 +00:00
Aryan
d8825e7697
Fix failing lora tests after HunyuanVideo lora ( #10307 )
...
fix
2024-12-20 02:35:41 +05:30
hlky
074798b299
Fix local_files_only for checkpoints with shards ( #10294 )
2024-12-19 07:04:57 -10:00
Dhruv Nair
3ee966950b
Allow Mochi Transformer to be split across multiple GPUs ( #10300 )
...
update
2024-12-19 22:34:44 +05:30
Dhruv Nair
9764f229d4
[Single File] Add single file support for Mochi Transformer ( #10268 )
...
update
2024-12-19 22:20:40 +05:30
Shenghai Yuan
1826a1e7d3
[LoRA] Support HunyuanVideo ( #10254 )
...
* 1217
* 1217
* 1217
* update
* reverse
* add test
* update test
* make style
* update
* make style
---------
Co-authored-by: Aryan <aryan@huggingface.co >
2024-12-19 16:22:20 +05:30
hlky
0ed09a17bb
Check correct model type is passed to from_pretrained ( #10189 )
...
* Check correct model type is passed to `from_pretrained`
* Flax, skip scheduler
* test_wrong_model
* Fix for scheduler
* Update tests/pipelines/test_pipelines.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* EnumMeta
* Flax
* scheduler in expected types
* make
* type object 'CLIPTokenizer' has no attribute '_PipelineFastTests__name'
* support union
* fix typing in kandinsky
* make
* add LCMScheduler
* 'LCMScheduler' object has no attribute 'sigmas'
* tests for wrong scheduler
* make
* update
* warning
* tests
* Update src/diffusers/pipelines/pipeline_utils.py
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
* import FlaxSchedulerMixin
* skip scheduler
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2024-12-19 09:24:52 +00:00
赵三石
2f7a417d1f
Update lora_conversion_utils.py ( #9980 )
...
x-flux single-blocks lora load
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2024-12-18 23:07:50 -10:00
hlky
4450d26b63
Add Flux Control to AutoPipeline ( #10292 )
2024-12-18 22:28:56 -10:00
Aryan
f781b8c30c
Hunyuan VAE tiling fixes and transformer docs ( #10295 )
...
* update
* udpate
* fix test
2024-12-19 10:28:10 +05:30
Sayak Paul
9c0e20de61
[chore] Update README_sana.md to update the default model ( #10285 )
...
Update README_sana.md to update the default model
2024-12-19 10:24:57 +05:30
Aryan
f35a38725b
[tests] remove nullop import checks from lora tests ( #10273 )
...
remove nullop imports
2024-12-19 01:19:08 +05:30
Aryan
f66bd3261c
Rename Mochi integration test correctly ( #10220 )
...
rename integration test
2024-12-18 22:41:23 +05:30
Aryan
c4c99c3907
[tests] Fix broken cuda, nightly and lora tests on main for CogVideoX ( #10270 )
...
fix joint pos embedding device
2024-12-18 22:36:08 +05:30
Dhruv Nair
862a7d5038
[Single File] Add single file support for Flux Canny, Depth and Fill ( #10288 )
...
update
2024-12-18 19:19:47 +05:30
Dhruv Nair
8304adce2a
Make zeroing prompt embeds for Mochi Pipeline configurable ( #10284 )
...
update
2024-12-18 18:32:53 +05:30
Dhruv Nair
b389f339ec
Fix Doc links in GGUF and Quantization overview docs ( #10279 )
...
* update
* Update docs/source/en/quantization/gguf.md
Co-authored-by: Aryan <aryan@huggingface.co >
---------
Co-authored-by: Aryan <aryan@huggingface.co >
2024-12-18 18:32:36 +05:30
hlky
e222246b4e
Fix sigma_last with use_flow_sigmas ( #10267 )
2024-12-18 12:22:10 +00:00
Andrés Romero
83709d5a06
Flux Control(Depth/Canny) + Inpaint ( #10192 )
...
* flux_control_inpaint - failing test_flux_different_prompts
* removing test_flux_different_prompts?
* fix style
* fix from PR comments
* fix style
* reducing guidance_scale in demo
* Update src/diffusers/pipelines/flux/pipeline_flux_control_inpaint.py
Co-authored-by: hlky <hlky@hlky.ac >
* make
* prepare_latents is not copied from
* update docs
* typos
---------
Co-authored-by: affromero <ubuntu@ip-172-31-17-146.ec2.internal >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: hlky <hlky@hlky.ac >
2024-12-18 09:14:16 +00:00
Qin Zhou
8eb73c872a
Support pass kwargs to sd3 custom attention processor ( #9818 )
...
* Support pass kwargs to sd3 custom attention processor
---------
Co-authored-by: hlky <hlky@hlky.ac >
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2024-12-17 21:58:33 -10:00
Xinyuan Zhao
88b015dc9f
Make time_embed_dim of UNet2DModel changeable ( #10262 )
2024-12-17 21:55:18 -10:00
Sayak Paul
63cdf9c0ba
[chore] fix: reamde -> readme ( #10276 )
...
fix: reamde -> readme
2024-12-18 10:56:08 +05:30
hlky
0ac52d6f09
Use torch in get_2d_rotary_pos_embed ( #10155 )
...
* Use `torch` in `get_2d_rotary_pos_embed`
* Add deprecation
2024-12-17 18:26:52 -10:00
Sayak Paul
ba6fd6eb30
[chore] fix: licensing headers in mochi and ltx ( #10275 )
...
fix: licensing header.
2024-12-18 08:43:57 +05:30
Sayak Paul
9408aa2dfc
[LoRA] feat: lora support for SANA. ( #10234 )
...
* feat: lora support for SANA.
* make fix-copies
* rename test class.
* attention_kwargs -> cross_attention_kwargs.
* Revert "attention_kwargs -> cross_attention_kwargs."
This reverts commit 23433bf9bc .
* exhaust 119 max line limit
* sana lora fine-tuning script.
* readme
* add a note about the supported models.
* Apply suggestions from code review
Co-authored-by: Aryan <aryan@huggingface.co >
* style
* docs for attention_kwargs.
* remove lora_scale from pag pipeline.
* copy fix
---------
Co-authored-by: Aryan <aryan@huggingface.co >
2024-12-18 08:22:31 +05:30
hlky
ec1c7a793f
Add set_shift to FlowMatchEulerDiscreteScheduler ( #10269 )
2024-12-17 21:40:09 +00:00