Sayak Paul
8eb17315c8
[LoRA] get rid of the legacy lora remnants and make our codebase lighter ( #8623 )
...
* get rid of the legacy lora remnants and make our codebase lighter
* fix depcrecated lora argument
* fix
* empty commit to trigger ci
* remove print
* empty
2024-06-21 16:36:05 +05:30
HelloWorldBeginner
58237364b1
Add Ascend NPU support for SDXL fine-tuning and fix the model saving bug when using DeepSpeed. ( #7816 )
...
* Add Ascend NPU support for SDXL fine-tuning and fix the model saving bug when using DeepSpeed.
* fix check code quality
* Decouple the NPU flash attention and make it an independent module.
* add doc and unit tests for npu flash attention.
---------
Co-authored-by: mhh001 <mahonghao1@huawei.com >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-05-03 08:14:34 -10:00
Steven Liu
3dd4168d4c
[docs] Minor updates ( #7063 )
...
* updates
* feedback
2024-02-25 09:38:02 -08:00
Steven Liu
3a7e481611
[docs] Video generation ( #6701 )
...
* first draft
* fix path
* fix path
* i2vgen-xl
* review
* modelscopet2v
* feedback
2024-02-16 16:35:37 -08:00
Sayak Paul
30e5e81d58
change to 2024 in the license ( #6902 )
...
change to 2024
2024-02-08 08:19:31 -10:00
Sayak Paul
a2bc2e14b9
[feat] allow SDXL pipeline to run with fused QKV projections ( #6030 )
...
* debug
* from step
* print
* turn sigma a list
* make str
* init_noise_sigma
* comment
* remove prints
* feat: introduce fused projections
* change to a better name
* no grad
* device.
* device
* dtype
* okay
* print
* more print
* fix: unbind -> split
* fix: qkv >-> k
* enable disable
* apply attention processor within the method
* attn processors
* _enable_fused_qkv_projections
* remove print
* add fused projection to vae
* add todos.
* add: documentation and cleanups.
* add: test for qkv projection fusion.
* relax assertions.
* relax further
* fix: docs
* fix-copies
* correct error message.
* Empty-Commit
* better conditioning on disable_fused_qkv_projections
* check
* check processor
* bfloat16 computation.
* check latent dtype
* style
* remove copy temporarily
* cast latent to bfloat16
* fix: vae -> self.vae
* remove print.
* add _change_to_group_norm_32
* comment out stuff that didn't work
* Apply suggestions from code review
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com >
* reflect patrick's suggestions.
* fix imports
* fix: disable call.
* fix more
* fix device and dtype
* fix conditions.
* fix more
* Apply suggestions from code review
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com >
---------
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com >
2023-12-06 07:33:26 +05:30
M. Tolga Cangöz
a359ff7644
[Docs] Fix typos and update files at API's Main Classes, Models, and Schedulers pages ( #5720 )
...
* Fix typos, update, add Copyright info, and trim trailing whitespaces
* Update docs/source/en/api/loaders.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* Update docs/source/en/api/models/autoencoder_tiny.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* Update docs/source/en/api/models/autoencoder_tiny.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
---------
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
2023-11-13 14:32:59 -08:00
Ruoxi
16b9a57d29
Implement CustomDiffusionAttnProcessor2_0. ( #4604 )
...
* Implement `CustomDiffusionAttnProcessor2_0`
* Doc-strings and type annotations for `CustomDiffusionAttnProcessor2_0`. (#1 )
* Update attnprocessor.md
* Update attention_processor.py
* Interops for `CustomDiffusionAttnProcessor2_0`.
* Formatted `attention_processor.py`.
* Formatted doc-string in `attention_processor.py`
* Conditional CustomDiffusion2_0 for training example.
* Remove unnecessary reference impl in comments.
* Fix `save_attn_procs`.
2023-09-18 14:49:00 +02:00
camenduru
c6ae9b7df6
Where did this 'x' come from, Elon? ( #4277 )
...
* why mdx?
* why mdx?
* why mdx?
* no x for kandinksy either
---------
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com >
2023-07-26 18:18:14 +02:00