Bagheera
99c0483b67
add skip_layers argument to SD3 transformer model class ( #9880 )
...
* add skip_layers argument to SD3 transformer model class
* add unit test for skip_layers in stable diffusion 3
* sd3: pipeline should support skip layer guidance
* up
---------
Co-authored-by: bghira <bghira@users.github.com >
Co-authored-by: yiyixuxu <yixu310@gmail.com >
2024-11-19 15:22:54 -05:00
Sayak Paul
7d0b9c4d4e
[LoRA] feat: save_lora_adapter() ( #9862 )
...
* feat: save_lora_adapter.
2024-11-18 21:03:38 -10:00
Yuxuan.Zhang
3b2830618d
CogVideoX 1.5 ( #9877 )
...
* CogVideoX1_1PatchEmbed test
* 1360 * 768
* refactor
* make style
* update docs
* add modeling tests for cogvideox 1.5
* update
* make fix-copies
* add ofs embed(for convert)
* add ofs embed(for convert)
* more resolution for cogvideox1.5-5b-i2v
* use even number of latent frames only
* update pipeline implementations
* make style
* set patch_size_t as None by default
* #skip frames 0
* refactor
* make style
* update docs
* fix ofs_embed
* update docs
* invert_scale_latents
* update
* fix
* Update docs/source/en/api/pipelines/cogvideox.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* Update docs/source/en/api/pipelines/cogvideox.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* Update docs/source/en/api/pipelines/cogvideox.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* Update docs/source/en/api/pipelines/cogvideox.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* Update src/diffusers/models/transformers/cogvideox_transformer_3d.py
* update conversion script
* remove copied from
* fix test
* Update docs/source/en/api/pipelines/cogvideox.md
* Update docs/source/en/api/pipelines/cogvideox.md
* Update docs/source/en/api/pipelines/cogvideox.md
* Update docs/source/en/api/pipelines/cogvideox.md
---------
Co-authored-by: Aryan <aryan@huggingface.co >
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
2024-11-19 00:56:34 +05:30
Aryan
3f329a426a
[core] Mochi T2V ( #9769 )
...
* update
* udpate
* update transformer
* make style
* fix
* add conversion script
* update
* fix
* update
* fix
* update
* fixes
* make style
* update
* update
* update
* init
* update
* update
* add
* up
* up
* up
* update
* mochi transformer
* remove original implementation
* make style
* update inits
* update conversion script
* docs
* Update src/diffusers/pipelines/mochi/pipeline_mochi.py
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
* Update src/diffusers/pipelines/mochi/pipeline_mochi.py
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
* fix docs
* pipeline fixes
* make style
* invert sigmas in scheduler; fix pipeline
* fix pipeline num_frames
* flip proj and gate in swiglu
* make style
* fix
* make style
* fix tests
* latent mean and std fix
* update
* cherry-pick 1069d210e1
* remove additional sigma already handled by flow match scheduler
* fix
* remove hardcoded value
* replace conv1x1 with linear
* Update src/diffusers/pipelines/mochi/pipeline_mochi.py
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
* framewise decoding and conv_cache
* make style
* Apply suggestions from code review
* mochi vae encoder changes
* rebase correctly
* Update scripts/convert_mochi_to_diffusers.py
* fix tests
* fixes
* make style
* update
* make style
* update
* add framewise and tiled encoding
* make style
* make original vae implementation behaviour the default; note: framewise encoding does not work
* remove framewise encoding implementation due to presence of attn layers
* fight test 1
* fight test 2
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
Co-authored-by: yiyixuxu <yixu310@gmail.com >
2024-11-05 20:33:41 +05:30
Sayak Paul
4adf6affbb
[Tests] clean up and refactor gradient checkpointing tests ( #9494 )
...
* check.
* fixes
* fixes
* updates
* fixes
* fixes
2024-10-31 18:24:19 +05:30
Aryan
0d1d267b12
[core] Allegro T2V ( #9736 )
...
* update
* refactor transformer part 1
* refactor part 2
* refactor part 3
* make style
* refactor part 4; modeling tests
* make style
* refactor part 5
* refactor part 6
* gradient checkpointing
* pipeline tests (broken atm)
* update
* add coauthor
Co-Authored-By: Huan Yang <hyang@fastmail.com >
* refactor part 7
* add docs
* make style
* add coauthor
Co-Authored-By: YiYi Xu <yixu310@gmail.com >
* make fix-copies
* undo unrelated change
* revert changes to embeddings, normalization, transformer
* refactor part 8
* make style
* refactor part 9
* make style
* fix
* apply suggestions from review
* Apply suggestions from code review
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* update example
* remove attention mask for self-attention
* update
* copied from
* update
* update
---------
Co-authored-by: Huan Yang <hyang@fastmail.com >
Co-authored-by: YiYi Xu <yixu310@gmail.com >
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
2024-10-29 13:14:36 +05:30
YiYi Xu
e2d037bbf1
minor doc/test update ( #9734 )
...
* update some docs and tests!
---------
Co-authored-by: Aryan <contact.aryanvs@gmail.com >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: Aryan <aryan@huggingface.co >
Co-authored-by: apolinário <joaopaulo.passos@gmail.com >
2024-10-21 13:06:13 -10:00
Yuxuan.Zhang
8d81564b27
CogView3Plus DiT ( #9570 )
...
* merge 9588
* max_shard_size="5GB" for colab running
* conversion script updates; modeling test; refactor transformer
* make fix-copies
* Update convert_cogview3_to_diffusers.py
* initial pipeline draft
* make style
* fight bugs 🐛 🪳
* add example
* add tests; refactor
* make style
* make fix-copies
* add co-author
YiYi Xu <yixu310@gmail.com >
* remove files
* add docs
* add co-author
Co-Authored-By: YiYi Xu <yixu310@gmail.com >
* fight docs
* address reviews
* make style
* make model work
* remove qkv fusion
* remove qkv fusion tets
* address review comments
* fix make fix-copies error
* remove None and TODO
* for FP16(draft)
* make style
* remove dynamic cfg
* remove pooled_projection_dim as a parameter
* fix tests
---------
Co-authored-by: Aryan <aryan@huggingface.co >
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2024-10-14 19:30:36 +05:30
Darren Hsu
61d37640ad
Support bfloat16 for Upsample2D ( #9480 )
...
* Support bfloat16 for Upsample2D
* Add test and use is_torch_version
* Resolve comments and add decorator
* Simplify require_torch_version_greater_equal decorator
* Run make style
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2024-10-01 16:08:12 -10:00
Sayak Paul
11542431a5
[Core] fix variant-identification. ( #9253 )
...
* fix variant-idenitification.
* fix variant
* fix sharded variant checkpoint loading.
* Apply suggestions from code review
* fixes.
* more fixes.
* remove print.
* fixes
* fixes
* comments
* fixes
* apply suggestions.
* hub_utils.py
* fix test
* updates
* fixes
* fixes
* Apply suggestions from code review
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* updates.
* removep patch file.
---------
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2024-09-28 09:57:31 +05:30
YiYi Xu
bac8a2412d
a few fix for SingleFile tests ( #9522 )
...
* update sd15 repo
* update more
2024-09-24 13:36:53 -10:00
Sayak Paul
aa73072f1f
[CI] fix nightly model tests ( #9483 )
...
* check if default attn procs fix it.
* print
* print
* replace
* style./
* replace revision with variant.
* replace with stable-diffusion-v1-5/stable-diffusion-inpainting.
* replace with stable-diffusion-v1-5/stable-diffusion-v1-5.
* fix
2024-09-21 07:44:47 +05:30
Dhruv Nair
1e8cf2763d
[CI] Nightly Test Updates ( #9380 )
...
* update
* update
* update
* update
* update
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2024-09-12 20:21:28 +05:30
Fanli Lin
2ee3215949
[tests] make 2 tests device-agnostic ( #9347 )
...
* enabel on xpu
* fix style
2024-09-03 16:34:03 -10:00
Aryan
24053832b5
[tests] remove/speedup some low signal tests ( #9285 )
...
* remove 2 shapes from SDFunctionTesterMixin::test_vae_tiling
* combine freeu enable/disable test to reduce many inference runs
* remove low signal unet test for signature
* remove low signal embeddings test
* remove low signal progress bar test from PipelineTesterMixin
* combine ip-adapter single and multi tests to save many inferences
* fix broken tests
* Update tests/pipelines/test_pipelines_common.py
* Update tests/pipelines/test_pipelines_common.py
* add progress bar tests
2024-09-03 13:59:18 +05:30
Dhruv Nair
f6f16a0c11
[CI] More Fast GPU Test Fixes ( #9346 )
...
* update
* update
* update
* update
2024-09-03 13:22:38 +05:30
Dhruv Nair
007ad0e2aa
[CI] More fixes for Fast GPU Tests on main ( #9300 )
...
update
2024-09-02 17:51:48 +05:30
Aryan
cbc2ec8f44
AnimateDiff prompt travel ( #9231 )
...
* update
* implement prompt interpolation
* make style
* resnet memory optimizations
* more memory optimizations; todo: refactor
* update
* update animatediff controlnet with latest changes
* refactor chunked inference changes
* remove print statements
* undo memory optimization changes
* update docstrings
* fix tests
* fix pia tests
* apply suggestions from review
* add tests
* update comment
2024-08-28 14:48:12 +05:30
YiYi Xu
c291617518
Flux followup ( #9074 )
...
* refactor rotary embeds
* adding jsmidt as co-author of this PR for https://github.com/huggingface/diffusers/pull/9133
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: Joseph Smidt <josephsmidt@gmail.com >
2024-08-21 08:44:58 -10:00
Dhruv Nair
940b8e0358
[CI] Multiple Slow Test fixes. ( #9198 )
...
* update
* update
* update
* update
2024-08-19 13:31:09 +05:30
M Saqlain
ba4348d9a7
[Tests] Improve transformers model test suite coverage - Lumina ( #8987 )
...
* Added test suite for lumina
* Fixed failing tests
* Improved code quality
* Added function docstrings
* Improved formatting
2024-08-19 08:29:03 +05:30
Sayak Paul
f848febacd
feat: allow sharding for auraflow. ( #8853 )
2024-08-18 08:47:26 +05:30
Sayak Paul
39b87b14b5
feat: allow flux transformer to be sharded during inference ( #9159 )
...
* feat: support sharding for flux.
* tests
2024-08-16 10:00:51 +05:30
Aryan
a85b34e7fd
[refactor] CogVideoX followups + tiled decoding support ( #9150 )
...
* refactor context parallel cache; update torch compile time benchmark
* add tiling support
* make style
* remove num_frames % 8 == 0 requirement
* update default num_frames to original value
* add explanations + refactor
* update torch compile example
* update docs
* update
* clean up if-statements
* address review comments
* add test for vae tiling
* update docs
* update docs
* update docstrings
* add modeling test for cogvideox transformer
* make style
2024-08-14 03:53:21 +05:30
Marc Sun
e4325606db
Fix loading sharded checkpoints when we have variants ( #9061 )
...
* Fix loading sharded checkpoint when we have variant
* add test
* remote print
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-08-06 13:38:44 -10:00
Vinh H. Pham
87e50a2f1d
[Tests] Improve transformers model test suite coverage - Hunyuan DiT ( #8916 )
...
* add hunyuan model test
* apply suggestions
* reduce dims further
* reduce dims further
* run make style
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-08-06 12:59:30 +05:30
Vinh H. Pham
e1d508ae92
[Tests] Improve transformers model test suite coverage - Latte ( #8919 )
...
* add LatteTransformer3DModel model test
* change patch_size to 1
* reduce req len
* reduce channel dims
* increase num_layers
* reduce dims further
* run make style
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: Aryan <aryan@huggingface.co >
2024-08-05 17:13:03 +05:30
Sayak Paul
0e460675e2
[Flux] allow tests to run ( #9050 )
...
* fix tests
* fix
* float64 skip
* remove sample_size.
* remove
* remove more
* default_sample_size.
* credit black forest for flux model.
* skip
* fix: tests
* remove OriginalModelMixin
* add transformer model test
* add: transformer model tests
2024-08-02 11:49:59 +05:30
YiYi Xu
95a7832879
fix load sharded checkpoint from a subfolder (local path) ( #8913 )
...
fix
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-08-01 20:15:42 +05:30
Yoach Lacombe
ea1b4ea7ca
Fix Stable Audio repository id ( #9016 )
...
Fix Stable Audio repo id
2024-07-30 23:17:44 +05:30
Yoach Lacombe
69e72b1dd1
Stable Audio integration ( #8716 )
...
* WIP modeling code and pipeline
* add custom attention processor + custom activation + add to init
* correct ProjectionModel forward
* add stable audio to __initèè
* add autoencoder and update pipeline and modeling code
* add half Rope
* add partial rotary v2
* add temporary modfis to scheduler
* add EDM DPM Solver
* remove TODOs
* clean GLU
* remove att.group_norm to attn processor
* revert back src/diffusers/schedulers/scheduling_dpmsolver_multistep.py
* refactor GLU -> SwiGLU
* remove redundant args
* add channel multiples in autoencoder docstrings
* changes in docsrtings and copyright headers
* clean pipeline
* further cleaning
* remove peft and lora and fromoriginalmodel
* Delete src/diffusers/pipelines/stable_audio/diffusers.code-workspace
* make style
* dummy models
* fix copied from
* add fast oobleck tests
* add brownian tree
* oobleck autoencoder slow tests
* remove TODO
* fast stable audio pipeline tests
* add slow tests
* make style
* add first version of docs
* wrap is_torchsde_available to the scheduler
* fix slow test
* test with input waveform
* add input waveform
* remove some todos
* create stableaudio gaussian projection + make style
* add pipeline to toctree
* fix copied from
* make quality
* refactor timestep_features->time_proj
* refactor joint_attention_kwargs->cross_attention_kwargs
* remove forward_chunk
* move StableAudioDitModel to transformers folder
* correct convert + remove partial rotary embed
* apply suggestions from yiyixuxu -> removing attn.kv_heads
* remove temb
* remove cross_attention_kwargs
* further removal of cross_attention_kwargs
* remove text encoder autocast to fp16
* continue removing autocast
* make style
* refactor how text and audio are embedded
* add paper
* update example code
* make style
* unify projection model forward + fix device placement
* make style
* remove fuse qkv
* apply suggestions from review
* Update src/diffusers/pipelines/stable_audio/pipeline_stable_audio.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* make style
* smaller models in fast tests
* pass sequential offloading fast tests
* add docs for vae and autoencoder
* make style and update example
* remove useless import
* add cosine scheduler
* dummy classes
* cosine scheduler docs
* better description of scheduler
---------
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2024-07-30 15:29:06 +05:30
Dhruv Nair
93983b6780
[CI] Skip flaky download tests in PR CI ( #8945 )
...
update
2024-07-24 09:25:06 +05:30
Vinh H. Pham
7a95f8d9d8
[Tests] Improve transformers model test suite coverage - Temporal Transformer ( #8932 )
...
* add test for temporal transformer
* remove unused variable
* fix code quality
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-07-23 15:36:30 +05:30
Sayak Paul
af400040f5
[Tests] proper skipping of request caching test ( #8908 )
...
proper skipping of request caching test
2024-07-22 12:52:57 -10:00
Sayak Paul
0f09b01ab3
[Core] fix: shard loading and saving when variant is provided. ( #8869 )
...
fix: shard loading and saving when variant is provided.
2024-07-17 08:26:28 +05:30
Sayak Paul
2261510bbc
[Core] Add AuraFlow ( #8796 )
...
* add lavender flow transformer
---------
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2024-07-11 08:50:19 -10:00
Sayak Paul
a785992c1d
[Tests] fix more sharding tests ( #8797 )
...
* fix
* fix
* ugly
* okay
* fix more
* fix oops
2024-07-09 13:09:36 +05:30
Tolga Cangöz
57084dacc5
Remove unnecessary lines ( #8569 )
...
* Remove unused line
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-07-08 10:42:02 -10:00
YiYi Xu
9e9ed353a2
fix loading sharded checkpoints from subfolder ( #8798 )
...
* fix load sharded checkpoints from subfolder{
* style
* os.path.join
* add a small test
---------
Co-authored-by: sayakpaul <spsayakpaul@gmail.com >
2024-07-06 11:32:04 -10:00
Sayak Paul
31adeb41cd
[Tests] fix sharding tests ( #8764 )
...
fix sharding tests
2024-07-04 08:50:59 +05:30
Mathis Koroglu
3e0d128da7
Motion Model / Adapter versatility ( #8301 )
...
* Motion Model / Adapter versatility
- allow to use a different number of layers per block
- allow to use a different number of transformer per layers per block
- allow a different number of motion attention head per block
- use dropout argument in get_down/up_block in 3d blocks
* Motion Model added arguments renamed & refactoring
* Add test for asymmetric UNetMotionModel
2024-06-27 11:11:29 +05:30
Dhruv Nair
effe4b9784
Update xformers SD3 test ( #8712 )
...
update
2024-06-26 10:24:27 -10:00
Dhruv Nair
0f0b531827
Add decorator for compile tests ( #8703 )
...
* update
* update
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-06-26 11:26:47 +05:30
Sayak Paul
4ad7a1f5fd
[Chore] create a utility for calculating the expected number of shards. ( #8692 )
...
create a utility for calculating the expected number of shards.
2024-06-25 17:05:39 +05:30
Tolga Cangöz
c375903db5
Errata - Fix typos & improve contributing page ( #8572 )
...
* Fix typos & improve contributing page
* `make style && make quality`
* fix typos
* Fix typo
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-06-24 14:13:03 +05:30
YiYi Xu
c71c19c5e6
a few fix for shard checkpoints ( #8656 )
...
fix
Co-authored-by: yiyixuxu <yixu310@gmail,com>
2024-06-21 12:50:58 +05:30
Marc Sun
96399c3ec6
Fix sharding when no device_map is passed ( #8531 )
...
* Fix sharding when no device_map is passed
* style
* add tests
* align
* add docstring
* format
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-06-18 05:47:23 -10:00
Dhruv Nair
04717fd861
Add Stable Diffusion 3 ( #8483 )
...
* up
* add sd3
* update
* update
* add tests
* fix copies
* fix docs
* update
* add dreambooth lora
* add LoRA
* update
* update
* update
* update
* import fix
* update
* Update src/diffusers/pipelines/stable_diffusion_3/pipeline_stable_diffusion_3.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* import fix 2
* update
* Update src/diffusers/models/autoencoders/autoencoder_kl.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/models/autoencoders/autoencoder_kl.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/models/autoencoders/autoencoder_kl.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/models/autoencoders/autoencoder_kl.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/models/autoencoders/autoencoder_kl.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/models/autoencoders/autoencoder_kl.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/models/autoencoders/autoencoder_kl.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/models/autoencoders/autoencoder_kl.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/models/autoencoders/autoencoder_kl.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/models/autoencoders/autoencoder_kl.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/models/autoencoders/autoencoder_kl.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* update
* update
* update
* fix ckpt id
* fix more ids
* update
* missing doc
* Update src/diffusers/schedulers/scheduling_flow_match_euler_discrete.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/schedulers/scheduling_flow_match_euler_discrete.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update docs/source/en/api/pipelines/stable_diffusion/stable_diffusion_3.md
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* Update docs/source/en/api/pipelines/stable_diffusion/stable_diffusion_3.md
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* update'
* fix
* update
* Update src/diffusers/models/autoencoders/autoencoder_kl.py
* Update src/diffusers/models/autoencoders/autoencoder_kl.py
* note on gated access.
* requirements
* licensing
---------
Co-authored-by: sayakpaul <spsayakpaul@gmail.com >
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2024-06-12 20:44:00 +01:00
Sayak Paul
7d887118b9
[Core] support saving and loading of sharded checkpoints ( #7830 )
...
* feat: support saving a model in sharded checkpoints.
* feat: make loading of sharded checkpoints work.
* add tests
* cleanse the loading logic a bit more.
* more resilience while loading from the Hub.
* parallelize shard downloads by using snapshot_download()/
* default to a shard size.
* more fix
* Empty-Commit
* debug
* fix
* uality
* more debugging
* fix more
* initial comments from Benjamin
* move certain methods to loading_utils
* add test to check if the correct number of shards are present.
* add a test to check if loading of sharded checkpoints from the Hub is okay
* clarify the unit when passed as an int.
* use hf_hub for sharding.
* remove unnecessary code
* remove unnecessary function
* lucain's comments.
* fixes
* address high-level comments.
* fix test
* subfolder shenanigans./
* Update src/diffusers/utils/hub_utils.py
Co-authored-by: Lucain <lucainp@gmail.com >
* Apply suggestions from code review
Co-authored-by: Lucain <lucainp@gmail.com >
* remove _huggingface_hub_version as not needed.
* address more feedback.
* add a test for local_files_only=True/
* need hf hub to be at least 0.23.2
* style
* final comment.
* clean up subfolder.
* deal with suffixes in code.
* _add_variant default.
* use weights_name_pattern
* remove add_suffix_keyword
* clean up downloading of sharded ckpts.
* don't return something special when using index.json
* fix more
* don't use bare except
* remove comments and catch the errors better
* fix a couple of things when using is_file()
* empty
---------
Co-authored-by: Lucain <lucainp@gmail.com >
2024-06-07 14:49:10 +05:30
Sayak Paul
a0542c1917
[LoRA] Remove legacy LoRA code and related adjustments ( #8316 )
...
* remove legacy code from load_attn_procs.
* finish first draft
* fix more.
* fix more
* add test
* add serialization support.
* fix-copies
* require peft backend for lora tests
* style
* fix test
* fix loading.
* empty
* address benjamin's feedback.
2024-06-05 08:15:30 +04:00