Sayak Paul
531e719163
[LoRA] use the PyTorch classes wherever needed and start depcrecation cycles ( #7204 )
...
* fix PyTorch classes and start deprecsation cycles.
* remove args crafting for accommodating scale.
* remove scale check in feedforward.
* assert against nn.Linear and not CompatibleLinear.
* remove conv_cls and lineaR_cls.
* remove scale
* 👋 scale.
* fix: unet2dcondition
* fix attention.py
* fix: attention.py again
* fix: unet_2d_blocks.
* fix-copies.
* more fixes.
* fix: resnet.py
* more fixes
* fix i2vgenxl unet.
* depcrecate scale gently.
* fix-copies
* Apply suggestions from code review
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* quality
* throw warning when scale is passed to the the BasicTransformerBlock class.
* remove scale from signature.
* cross_attention_kwargs, very nice catch by Yiyi
* fix: logger.warn
* make deprecation message clearer.
* address final comments.
* maintain same depcrecation message and also add it to activations.
* address yiyi
* fix copies
* Apply suggestions from code review
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* more depcrecation
* fix-copies
---------
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2024-03-13 07:56:19 +05:30
Sayak Paul
30e5e81d58
change to 2024 in the license ( #6902 )
...
change to 2024
2024-02-08 08:19:31 -10:00
Sayak Paul
09b7bfce91
[Core] move transformer scripts to transformers modules ( #6747 )
...
* move transformer scripts to transformers modules
* move transformer model test
* move prior transformer test to directory
* fix doc path
* correct doc path
* add: __init__.py
2024-01-29 22:28:28 +05:30
Arsalan
f427345ab1
Device agnostic testing ( #5612 )
...
* utils and test modifications to enable device agnostic testing
* device for manual seed in unet1d
* fix generator condition in vae test
* consistency changes to testing
* make style
* add device agnostic testing changes to source and one model test
* make dtype check fns private, log cuda fp16 case
* remove dtype checks from import utils, move to testing_utils
* adding tests for most model classes and one pipeline
* fix vae import
2023-12-05 19:04:13 +05:30
Dhruv Nair
b6e0b016ce
Lazy Import for Diffusers ( #4829 )
...
* initial commit
* move modules to import struct
* add dummy objects and _LazyModule
* add lazy import to schedulers
* clean up unused imports
* lazy import on models module
* lazy import for schedulers module
* add lazy import to pipelines module
* lazy import altdiffusion
* lazy import audio diffusion
* lazy import audioldm
* lazy import consistency model
* lazy import controlnet
* lazy import dance diffusion ddim ddpm
* lazy import deepfloyd
* lazy import kandinksy
* lazy imports
* lazy import semantic diffusion
* lazy imports
* lazy import stable diffusion
* move sd output to its own module
* clean up
* lazy import t2iadapter
* lazy import unclip
* lazy import versatile and vq diffsuion
* lazy import vq diffusion
* helper to fetch objects from modules
* lazy import sdxl
* lazy import txt2vid
* lazy import stochastic karras
* fix model imports
* fix bug
* lazy import
* clean up
* clean up
* fixes for tests
* fixes for tests
* clean up
* remove import of torch_utils from utils module
* clean up
* clean up
* fix mistake import statement
* dedicated modules for exporting and loading
* remove testing utils from utils module
* fixes from merge conflicts
* Update src/diffusers/pipelines/kandinsky2_2/__init__.py
* fix docs
* fix alt diffusion copied from
* fix check dummies
* fix more docs
* remove accelerate import from utils module
* add type checking
* make style
* fix check dummies
* remove torch import from xformers check
* clean up error message
* fixes after upstream merges
* dummy objects fix
* fix tests
* remove unused module import
---------
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com >
2023-09-11 09:56:22 +02:00
Batuhan Taskaya
ff8f58086b
Load Kohya-ss style LoRAs with auxilary states ( #4147 )
...
* Support to load Kohya-ss style LoRA file format (without restrictions)
Co-Authored-By: Takuma Mori <takuma104@gmail.com >
Co-Authored-By: Sayak Paul <spsayakpaul@gmail.com >
* tmp: add sdxl to mlp_modules
---------
Co-authored-by: Takuma Mori <takuma104@gmail.com >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2023-07-26 00:24:19 +02:00
Patrick von Platen
51843fd7d0
Refactor full determinism ( #3485 )
...
* up
* fix more
* Apply suggestions from code review
* fix more
* fix more
* Check it
* Remove 16:8
* fix more
* fix more
* fix more
* up
* up
* Test only stable diffusion
* Test only two files
* up
* Try out spinning up processes that can be killed
* up
* Apply suggestions from code review
* up
* up
2023-05-22 11:15:11 +01:00
Will Berman
909742dbd6
attention refactor: the trilogy ( #3387 )
...
* Replace `AttentionBlock` with `Attention`
* use _from_deprecated_attn_block check re: @patrickvonplaten
2023-05-12 08:54:09 -06:00
Patrick von Platen
3a9d7d9758
[Tests] parallelize ( #3078 )
...
* [Tests] parallelize
* finish folder structuring
* Parallelize tests more
* Correct saving of pipelines
* make sure logging level is correct
* try again
* Apply suggestions from code review
Co-authored-by: Pedro Cuenca <pedro@huggingface.co >
---------
Co-authored-by: Pedro Cuenca <pedro@huggingface.co >
2023-04-13 13:32:57 +01:00