Sayak Paul
43459079ab
[core] feat: support group offloading at the pipeline level ( #12283 )
...
* feat: support group offloading at the pipeline level.
* add tests
* up
* [docs] Pipeline group offloading (#12286 )
init
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
---------
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
2025-09-10 09:09:57 +05:30
Dhruv Nair
7aa6af1138
[Refactor] Move testing utils out of src ( #12238 )
...
* update
* update
* update
* update
* update
* merge main
* Revert "merge main"
This reverts commit 65efbcead5 .
2025-08-28 19:53:02 +05:30
Sayak Paul
46a0c6aa82
feat: cuda device_map for pipelines. ( #12122 )
...
* feat: cuda device_map for pipelines.
* up
* up
* empty
* up
2025-08-14 10:31:24 +05:30
Aryan
18c8f10f20
[refactor] Flux/Chroma single file implementation + Attention Dispatcher ( #11916 )
...
* update
* update
* add coauthor
Co-Authored-By: Dhruv Nair <dhruv.nair@gmail.com >
* improve test
* handle ip adapter params correctly
* fix chroma qkv fusion test
* fix fastercache implementation
* fix more tests
* fight more tests
* add back set_attention_backend
* update
* update
* make style
* make fix-copies
* make ip adapter processor compatible with attention dispatcher
* refactor chroma as well
* remove rmsnorm assert
* minify and deprecate npu/xla processors
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2025-07-17 17:30:39 +05:30
Aryan
0454fbb30b
First Block Cache ( #11180 )
...
* update
* modify flux single blocks to make compatible with cache techniques (without too much model-specific intrusion code)
* remove debug logs
* update
* cache context for different batches of data
* fix hs residual bug for single return outputs; support ltx
* fix controlnet flux
* support flux, ltx i2v, ltx condition
* update
* update
* Update docs/source/en/api/cache.md
* Update src/diffusers/hooks/hooks.py
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
* address review comments pt. 1
* address review comments pt. 2
* cache context refacotr; address review pt. 3
* address review comments
* metadata registration with decorators instead of centralized
* support cogvideox
* support mochi
* fix
* remove unused function
* remove central registry based on review
* update
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2025-07-09 03:27:15 +05:30
Aryan
3f3f0c16a6
[tests] Fix failing float16 cuda tests ( #11835 )
...
* update
* update
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-07-01 11:13:58 +05:30
kaixuanliu
27bf7fcd0e
adjust tolerance criteria for test_float16_inference in unit test ( #11809 )
...
Signed-off-by: Liu, Kaixuan <kaixuan.liu@intel.com >
2025-06-26 13:19:59 +05:30
kaixuanliu
dd285099eb
adjust to get CI test cases passed on XPU ( #11759 )
...
* adjust to get CI test cases passed on XPU
Signed-off-by: Liu, Kaixuan <kaixuan.liu@intel.com >
* fix format issue
Signed-off-by: Liu, Kaixuan <kaixuan.liu@intel.com >
* Apply style fixes
---------
Signed-off-by: Liu, Kaixuan <kaixuan.liu@intel.com >
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
Co-authored-by: Aryan <aryan@huggingface.co >
2025-06-25 14:02:17 +05:30
Aryan
474a248f10
[tests] Fix HunyuanVideo Framepack device tests ( #11789 )
...
update
2025-06-24 13:49:37 +05:30
Edna
8adc6003ba
Chroma Pipeline ( #11698 )
...
* working state from hameerabbasi and iddl
* working state form hameerabbasi and iddl (transformer)
* working state (normalization)
* working state (embeddings)
* add chroma loader
* add chroma to mappings
* add chroma to transformer init
* take out variant stuff
* get decently far in changing variant stuff
* add chroma init
* make chroma output class
* add chroma transformer to dummy tp
* add chroma to init
* add chroma to init
* fix single file
* update
* update
* add chroma to auto pipeline
* add chroma to pipeline init
* change to chroma transformer
* take out variant from blocks
* swap embedder location
* remove prompt_2
* work on swapping text encoders
* remove mask function
* dont modify mask (for now)
* wrap attn mask
* no attn mask (can't get it to work)
* remove pooled prompt embeds
* change to my own unpooled embeddeer
* fix load
* take pooled projections out of transformer
* ensure correct dtype for chroma embeddings
* update
* use dn6 attn mask + fix true_cfg_scale
* use chroma pipeline output
* use DN6 embeddings
* remove guidance
* remove guidance embed (pipeline)
* remove guidance from embeddings
* don't return length
* dont change dtype
* remove unused stuff, fix up docs
* add chroma autodoc
* add .md (oops)
* initial chroma docs
* undo don't change dtype
* undo arxiv change
unsure why that happened
* fix hf papers regression in more places
* Update docs/source/en/api/pipelines/chroma.md
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
* do_cfg -> self.do_classifier_free_guidance
* Update docs/source/en/api/models/chroma_transformer.md
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
* Update chroma.md
* Move chroma layers into transformer
* Remove pruned AdaLayerNorms
* Add chroma fast tests
* (untested) batch cond and uncond
* Add # Copied from for shift
* Update # Copied from statements
* update norm imports
* Revert cond + uncond batching
* Add transformer tests
* move chroma test (oops)
* chroma init
* fix chroma pipeline fast tests
* Update src/diffusers/models/transformers/transformer_chroma.py
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
* Move Approximator and Embeddings
* Fix auto pipeline + make style, quality
* make style
* Apply style fixes
* switch to new input ids
* fix # Copied from error
* remove # Copied from on protected members
* try to fix import
* fix import
* make fix-copes
* revert style fix
* update chroma transformer params
* update chroma transformer approximator init params
* update to pad tokens
* fix batch inference
* Make more pipeline tests work
* Make most transformer tests work
* fix docs
* make style, make quality
* skip batch tests
* fix test skipping
* fix test skipping again
* fix for tests
* Fix all pipeline test
* update
* push local changes, fix docs
* add encoder test, remove pooled dim
* default proj dim
* fix tests
* fix equal size list input
* update
* push local changes, fix docs
* add encoder test, remove pooled dim
* default proj dim
* fix tests
* fix equal size list input
* Revert "fix equal size list input"
This reverts commit 3fe4ad67d5 .
* update
* update
* update
* update
* update
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-06-14 06:52:56 +05:30
Dhruv Nair
5b0dab1253
Introduce DeprecatedPipelineMixin to simplify pipeline deprecation process ( #11596 )
...
* update
* update
* update
* update
* update
* update
* update
2025-06-09 13:03:40 +05:30
co63oc
8183d0f16e
Fix typos in strings and comments ( #11476 )
...
* Fix typos in strings and comments
Signed-off-by: co63oc <co63oc@users.noreply.github.com >
* Update src/diffusers/hooks/hooks.py
Co-authored-by: Aryan <contact.aryanvs@gmail.com >
* Update src/diffusers/hooks/hooks.py
Co-authored-by: Aryan <contact.aryanvs@gmail.com >
* Update layerwise_casting.py
* Apply style fixes
* update
---------
Signed-off-by: co63oc <co63oc@users.noreply.github.com >
Co-authored-by: Aryan <contact.aryanvs@gmail.com >
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-05-30 18:49:00 +05:30
Sayak Paul
4af76d0d7d
[tests] Changes to the torch.compile() CI and tests ( #11508 )
...
* remove compile cuda docker.
* replace compile cuda docker path.
* better manage compilation cache.
* propagate similar to the pipeline tests.
* remove unneeded compile test.
* small.
* don't check for deleted files.
2025-05-26 08:31:04 -07:00
Aryan
7b904941bc
Cosmos ( #10660 )
...
* begin transformer conversion
* refactor
* refactor
* refactor
* refactor
* refactor
* refactor
* update
* add conversion script
* add pipeline
* make fix-copies
* remove einops
* update docs
* gradient checkpointing
* add transformer test
* update
* debug
* remove prints
* match sigmas
* add vae pt. 1
* finish CV* vae
* update
* update
* update
* update
* update
* update
* make fix-copies
* update
* make fix-copies
* fix
* update
* update
* make fix-copies
* update
* update tests
* handle device and dtype for safety checker; required in latest diffusers
* remove enable_gqa and use repeat_interleave instead
* enforce safety checker; use dummy checker in fast tests
* add review suggestion for ONNX export
Co-Authored-By: Asfiya Baig <asfiyab@nvidia.com >
* fix safety_checker issues when not passed explicitly
We could either do what's done in this commit, or update the Cosmos examples to explicitly pass the safety checker
* use cosmos guardrail package
* auto format docs
* update conversion script to support 14B models
* update name CosmosPipeline -> CosmosTextToWorldPipeline
* update docs
* fix docs
* fix group offload test failing for vae
---------
Co-authored-by: Asfiya Baig <asfiyab@nvidia.com >
2025-05-07 20:59:09 +05:30
Yao Matrix
35fada4169
enable unidiffuser test cases on xpu ( #11444 )
...
* enable unidiffuser cases on XPU
Signed-off-by: Yao Matrix <matrix.yao@intel.com >
* fix a typo
Signed-off-by: Yao Matrix <matrix.yao@intel.com >
* fix style
Signed-off-by: Yao Matrix <matrix.yao@intel.com >
---------
Signed-off-by: Yao Matrix <matrix.yao@intel.com >
2025-04-30 13:58:00 +05:30
Yao Matrix
9ce89e2efa
enable group_offload cases and quanto cases on XPU ( #11405 )
...
* enable group_offload cases and quanto cases on XPU
Signed-off-by: YAO Matrix <matrix.yao@intel.com >
* use backend APIs
Signed-off-by: Yao Matrix <matrix.yao@intel.com >
* fix style
Signed-off-by: Yao Matrix <matrix.yao@intel.com >
---------
Signed-off-by: YAO Matrix <matrix.yao@intel.com >
Signed-off-by: Yao Matrix <matrix.yao@intel.com >
2025-04-28 06:37:16 +05:30
Sayak Paul
aa5f5d41d6
[tests] add tests to check for graph breaks, recompilation, cuda syncs in pipelines during torch.compile() ( #11085 )
...
* test for better torch.compile stuff.
* fixes
* recompilation and graph break.
* clear compilation cache.
* change to modeling level test.
* allow running compilation tests during nightlies.
2025-04-28 08:36:33 +08:00
Yao Matrix
68663f8a17
fix test_vanilla_funetuning failure on XPU and A100 ( #11263 )
...
* fix test_vanilla_funetuning failure on XPU and A100
Signed-off-by: Matrix Yao <matrix.yao@intel.com >
* change back to 5e-2
Signed-off-by: Matrix Yao <matrix.yao@intel.com >
---------
Signed-off-by: Matrix Yao <matrix.yao@intel.com >
2025-04-10 05:55:07 +01:00
Dhruv Nair
edc154da09
Update Ruff to latest Version ( #10919 )
...
* update
* update
* update
* update
2025-04-09 16:51:34 +05:30
Yao Matrix
c36c745ceb
fix FluxReduxSlowTests::test_flux_redux_inference case failure on XPU ( #11245 )
...
* loose test_float16_inference's tolerance from 5e-2 to 6e-2, so XPU can
pass UT
Signed-off-by: Matrix Yao <matrix.yao@intel.com >
* fix test_pipeline_flux_redux fail on XPU
Signed-off-by: Matrix Yao <matrix.yao@intel.com >
---------
Signed-off-by: Matrix Yao <matrix.yao@intel.com >
2025-04-09 11:41:15 +01:00
hlky
da857bebb6
Revert save_model in ModelMixin save_pretrained and use safe_serialization=False in test ( #11196 )
2025-04-02 12:45:36 +01:00
hlky
d8c617ccb0
allow models to run with a user-provided dtype map instead of a single dtype ( #10301 )
...
* allow models to run with a user-provided dtype map instead of a single dtype
* make style
* Add warning, change `_` to `default`
* make style
* add test
* handle shared tensors
* remove warning
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-04-02 09:05:46 +01:00
Aryan
844221ae4e
[core] FasterCache ( #10163 )
...
* init
* update
* update
* update
* make style
* update
* fix
* make it work with guidance distilled models
* update
* make fix-copies
* add tests
* update
* apply_faster_cache -> apply_fastercache
* fix
* reorder
* update
* refactor
* update docs
* add fastercache to CacheMixin
* update tests
* Apply suggestions from code review
* make style
* try to fix partial import error
* Apply style fixes
* raise warning
* update
---------
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-03-21 09:35:04 +05:30
Fanli Lin
15ad97f782
[tests] make cuda only tests device-agnostic ( #11058 )
...
* enable bnb on xpu
* add 2 more cases
* add missing change
* add missing change
* add one more
* enable cuda only tests on xpu
* enable big gpu cases
2025-03-20 10:12:35 +00:00
Daniel Regado
1450c2ac4f
Multi IP-Adapter for Flux pipelines ( #10867 )
...
* Initial implementation of Flux multi IP-Adapter
* Update src/diffusers/pipelines/flux/pipeline_flux.py
Co-authored-by: hlky <hlky@hlky.ac >
* Update src/diffusers/pipelines/flux/pipeline_flux.py
Co-authored-by: hlky <hlky@hlky.ac >
* Changes for ipa image embeds
* Update src/diffusers/pipelines/flux/pipeline_flux.py
Co-authored-by: hlky <hlky@hlky.ac >
* Update src/diffusers/pipelines/flux/pipeline_flux.py
Co-authored-by: hlky <hlky@hlky.ac >
* make style && make quality
* Updated ip_adapter test
* Created typing_utils.py
---------
Co-authored-by: hlky <hlky@hlky.ac >
2025-02-25 09:51:15 +00:00
Sayak Paul
b2ca39c8ac
[tests] test encode_prompt() in isolation ( #10438 )
...
* poc encode_prompt() tests
* fix
* updates.
* fixes
* fixes
* updates
* updates
* updates
* revert
* updates
* updates
* updates
* updates
* remove SDXLOptionalComponentsTesterMixin.
* remove tests that directly leveraged encode_prompt() in some way or the other.
* fix imports.
* remove _save_load
* fixes
* fixes
* fixes
* fixes
2025-02-20 13:21:43 +05:30
Aryan
9a147b82f7
Module Group Offloading ( #10503 )
...
* update
* fix
* non_blocking; handle parameters and buffers
* update
* Group offloading with cuda stream prefetching (#10516 )
* cuda stream prefetch
* remove breakpoints
* update
* copy model hook implementation from pab
* update; ~very workaround based implementation but it seems to work as expected; needs cleanup and rewrite
* more workarounds to make it actually work
* cleanup
* rewrite
* update
* make sure to sync current stream before overwriting with pinned params
not doing so will lead to erroneous computations on the GPU and cause bad results
* better check
* update
* remove hook implementation to not deal with merge conflict
* re-add hook changes
* why use more memory when less memory do trick
* why still use slightly more memory when less memory do trick
* optimise
* add model tests
* add pipeline tests
* update docs
* add layernorm and groupnorm
* address review comments
* improve tests; add docs
* improve docs
* Apply suggestions from code review
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* apply suggestions from code review
* update tests
* apply suggestions from review
* enable_group_offloading -> enable_group_offload for naming consistency
* raise errors if multiple offloading strategies used; add relevant tests
* handle .to() when group offload applied
* refactor some repeated code
* remove unintentional change from merge conflict
* handle .cuda()
---------
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
2025-02-14 12:59:45 +05:30
Aryan
658e24e86c
[core] Pyramid Attention Broadcast ( #9562 )
...
* start pyramid attention broadcast
* add coauthor
Co-Authored-By: Xuanlei Zhao <43881818+oahzxl@users.noreply.github.com >
* update
* make style
* update
* make style
* add docs
* add tests
* update
* Update docs/source/en/api/pipelines/cogvideox.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* Update docs/source/en/api/pipelines/cogvideox.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* Pyramid Attention Broadcast rewrite + introduce hooks (#9826 )
* rewrite implementation with hooks
* make style
* update
* merge pyramid-attention-rewrite-2
* make style
* remove changes from latte transformer
* revert docs changes
* better debug message
* add todos for future
* update tests
* make style
* cleanup
* fix
* improve log message; fix latte test
* refactor
* update
* update
* update
* revert changes to tests
* update docs
* update tests
* Apply suggestions from code review
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* update
* fix flux test
* reorder
* refactor
* make fix-copies
* update docs
* fixes
* more fixes
* make style
* update tests
* update code example
* make fix-copies
* refactor based on reviews
* use maybe_free_model_hooks
* CacheMixin
* make style
* update
* add current_timestep property; update docs
* make fix-copies
* update
* improve tests
* try circular import fix
* apply suggestions from review
* address review comments
* Apply suggestions from code review
* refactor hook implementation
* add test suite for hooks
* PAB Refactor (#10667 )
* update
* update
* update
---------
Co-authored-by: DN6 <dhruv.nair@gmail.com >
* update
* fix remove hook behaviour
---------
Co-authored-by: Xuanlei Zhao <43881818+oahzxl@users.noreply.github.com >
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
Co-authored-by: DN6 <dhruv.nair@gmail.com >
2025-01-28 05:09:04 +05:30
Aryan
beacaa5528
[core] Layerwise Upcasting ( #10347 )
...
* update
* update
* make style
* remove dynamo disable
* add coauthor
Co-Authored-By: Dhruv Nair <dhruv.nair@gmail.com >
* update
* update
* update
* update mixin
* add some basic tests
* update
* update
* non_blocking
* improvements
* update
* norm.* -> norm
* apply suggestions from review
* add example
* update hook implementation to the latest changes from pyramid attention broadcast
* deinitialize should raise an error
* update doc page
* Apply suggestions from code review
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* update docs
* update
* refactor
* fix _always_upcast_modules for asym ae and vq_model
* fix lumina embedding forward to not depend on weight dtype
* refactor tests
* add simple lora inference tests
* _always_upcast_modules -> _precision_sensitive_module_patterns
* remove todo comments about review; revert changes to self.dtype in unets because .dtype on ModelMixin should be able to handle fp8 weight case
* check layer dtypes in lora test
* fix UNet1DModelTests::test_layerwise_upcasting_inference
* _precision_sensitive_module_patterns -> _skip_layerwise_casting_patterns based on feedback
* skip test in NCSNppModelTests
* skip tests for AutoencoderTinyTests
* skip tests for AutoencoderOobleckTests
* skip tests for UNet1DModelTests - unsupported pytorch operations
* layerwise_upcasting -> layerwise_casting
* skip tests for UNetRLModelTests; needs next pytorch release for currently unimplemented operation support
* add layerwise fp8 pipeline test
* use xfail
* Apply suggestions from code review
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
* add assertion with fp32 comparison; add tolerance to fp8-fp32 vs fp32-fp32 comparison (required for a few models' test to pass)
* add note about memory consumption on tesla CI runner for failing test
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
2025-01-22 19:49:37 +05:30
Marc Sun
fbff43acc9
[FEAT] DDUF format ( #10037 )
...
* load and save dduf archive
* style
* switch to zip uncompressed
* updates
* Update src/diffusers/pipelines/pipeline_utils.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* Update src/diffusers/pipelines/pipeline_utils.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* first draft
* remove print
* switch to dduf_file for consistency
* switch to huggingface hub api
* fix log
* add a basic test
* Update src/diffusers/configuration_utils.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* Update src/diffusers/pipelines/pipeline_utils.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* Update src/diffusers/pipelines/pipeline_utils.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* fix
* fix variant
* change saving logic
* DDUF - Load transformers components manually (#10171 )
* update hfh version
* Load transformers components manually
* load encoder from_pretrained with state_dict
* working version with transformers and tokenizer !
* add generation_config case
* fix tests
* remove saving for now
* typing
* need next version from transformers
* Update src/diffusers/configuration_utils.py
Co-authored-by: Lucain <lucain@huggingface.co >
* check path corectly
* Apply suggestions from code review
Co-authored-by: Lucain <lucain@huggingface.co >
* udapte
* typing
* remove check for subfolder
* quality
* revert setup changes
* oups
* more readable condition
* add loading from the hub test
* add basic docs.
* Apply suggestions from code review
Co-authored-by: Lucain <lucain@huggingface.co >
* add example
* add
* make functions private
* Apply suggestions from code review
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* minor.
* fixes
* fix
* change the precdence of parameterized.
* error out when custom pipeline is passed with dduf_file.
* updates
* fix
* updates
* fixes
* updates
* fix xfail condition.
* fix xfail
* fixes
* sharded checkpoint compat
* add test for sharded checkpoint
* add suggestions
* Update src/diffusers/models/model_loading_utils.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* from suggestions
* add class attributes to flag dduf tests
* last one
* fix logic
* remove comment
* revert changes
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: Lucain <lucain@huggingface.co >
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2025-01-14 13:21:42 +05:30
Sayak Paul
daf9d0f119
[chore] remove prints from tests. ( #10505 )
...
remove prints from tests.
2025-01-09 14:19:43 +05:30
hlky
be2070991f
Support Flux IP Adapter ( #10261 )
...
* Flux IP-Adapter
* test cfg
* make style
* temp remove copied from
* fix test
* fix test
* v2
* fix
* make style
* temp remove copied from
* Apply suggestions from code review
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Move encoder_hid_proj to inside FluxTransformer2DModel
* merge
* separate encode_prompt, add copied from, image_encoder offload
* make
* fix test
* fix
* Update src/diffusers/pipelines/flux/pipeline_flux.py
* test_flux_prompt_embeds change not needed
* true_cfg -> true_cfg_scale
* fix merge conflict
* test_flux_ip_adapter_inference
* add fast test
* FluxIPAdapterMixin not test mixin
* Update pipeline_flux.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
---------
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2024-12-21 17:49:58 +00:00
Sayak Paul
c1926cef6b
[tests] refactor vae tests ( #9808 )
...
* add: autoencoderkl tests
* autoencodertiny.
* fix
* asymmetric autoencoder.
* more
* integration tests for stable audio decoder.
* consistency decoder vae tests
* remove grad check from consistency decoder.
* cog
* bye test_models_vae.py
* fix
* fix
* remove allegro
* fixes
* fixes
* fixes
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2024-12-04 15:58:36 +05:30
Fanli Lin
64b3e0f539
make pipelines tests device-agnostic (part1) ( #9399 )
...
* enable on xpu
* add 1 more
* add one more
* enable more
* add 1 more
* add more
* enable 1
* enable more cases
* enable
* enable
* update comment
* one more
* enable 1
* add more cases
* enable xpu
* add one more caswe
* add more cases
* add 1
* add more
* add more cases
* add case
* enable
* add more
* add more
* add more
* enbale more
* add more
* update code
* update test marker
* add skip back
* update comment
* remove single files
* remove
* style
* add
* revert
* reformat
* update decorator
* update
* update
* update
* Update tests/pipelines/deepfloyd_if/test_if.py
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
* Update src/diffusers/utils/testing_utils.py
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
* Update tests/pipelines/animatediff/test_animatediff_controlnet.py
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
* Update tests/pipelines/animatediff/test_animatediff.py
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
* Update tests/pipelines/animatediff/test_animatediff_controlnet.py
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
* update float16
* no unitest.skipt
* update
* apply style check
* reapply format
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2024-11-22 15:32:54 +05:30
Sayak Paul
ded3db164b
[Core] introduce controlnet module ( #8768 )
...
* move vae flax module.
* controlnet module.
* prepare for PR.
* revert a commit
* gracefully deprecate controlnet deps.
* fix
* fix doc path
* fix-copies
* fix path
* style
* style
* conflicts
* fix
* fix-copies
* sparsectrl.
* updates
* fix
* updates
* updates
* updates
* fix
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2024-11-06 22:08:55 -04:00
YiYi Xu
d2e5cb3c10
Revert "[LoRA] fix: lora loading when using with a device_mapped mode⦠( #9823 )
...
Revert "[LoRA] fix: lora loading when using with a device_mapped model. (#9449 )"
This reverts commit 41e4779d98 .
2024-10-31 08:19:32 -10:00
Sayak Paul
41e4779d98
[LoRA] fix: lora loading when using with a device_mapped model. ( #9449 )
...
* fix: lora loading when using with a device_mapped model.
* better attibutung
* empty
Co-authored-by: Benjamin Bossan <BenjaminBossan@users.noreply.github.com >
* Apply suggestions from code review
Co-authored-by: Marc Sun <57196510+SunMarc@users.noreply.github.com >
* minors
* better error messages.
* fix-copies
* add: tests, docs.
* add hardware note.
* quality
* Update docs/source/en/training/distributed_inference.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* fixes
* skip properly.
* fixes
---------
Co-authored-by: Benjamin Bossan <BenjaminBossan@users.noreply.github.com >
Co-authored-by: Marc Sun <57196510+SunMarc@users.noreply.github.com >
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
2024-10-31 21:17:41 +05:30
Aryan
0d1d267b12
[core] Allegro T2V ( #9736 )
...
* update
* refactor transformer part 1
* refactor part 2
* refactor part 3
* make style
* refactor part 4; modeling tests
* make style
* refactor part 5
* refactor part 6
* gradient checkpointing
* pipeline tests (broken atm)
* update
* add coauthor
Co-Authored-By: Huan Yang <hyang@fastmail.com >
* refactor part 7
* add docs
* make style
* add coauthor
Co-Authored-By: YiYi Xu <yixu310@gmail.com >
* make fix-copies
* undo unrelated change
* revert changes to embeddings, normalization, transformer
* refactor part 8
* make style
* refactor part 9
* make style
* fix
* apply suggestions from review
* Apply suggestions from code review
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* update example
* remove attention mask for self-attention
* update
* copied from
* update
* update
---------
Co-authored-by: Huan Yang <hyang@fastmail.com >
Co-authored-by: YiYi Xu <yixu310@gmail.com >
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
2024-10-29 13:14:36 +05:30
Sayak Paul
11542431a5
[Core] fix variant-identification. ( #9253 )
...
* fix variant-idenitification.
* fix variant
* fix sharded variant checkpoint loading.
* Apply suggestions from code review
* fixes.
* more fixes.
* remove print.
* fixes
* fixes
* comments
* fixes
* apply suggestions.
* hub_utils.py
* fix test
* updates
* fixes
* fixes
* Apply suggestions from code review
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* updates.
* removep patch file.
---------
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2024-09-28 09:57:31 +05:30
Aryan
24053832b5
[tests] remove/speedup some low signal tests ( #9285 )
...
* remove 2 shapes from SDFunctionTesterMixin::test_vae_tiling
* combine freeu enable/disable test to reduce many inference runs
* remove low signal unet test for signature
* remove low signal embeddings test
* remove low signal progress bar test from PipelineTesterMixin
* combine ip-adapter single and multi tests to save many inferences
* fix broken tests
* Update tests/pipelines/test_pipelines_common.py
* Update tests/pipelines/test_pipelines_common.py
* add progress bar tests
2024-09-03 13:59:18 +05:30
Γlvaro Somoza
39e1f7eaa4
[Kolors] Add PAG ( #8934 )
...
* txt2img pag added
* autopipe added, fixed case
* style
* apply suggestions
* added fast tests, added todo tests
* revert dummy objects for kolors
* fix pag dummies
* fix test imports
* update pag tests
* add kolor pag to docs
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-08-07 09:29:52 +05:30
Sayak Paul
50d21f7c6a
[Core] fix QKV fusion for attention ( #8829 )
...
* start debugging the problem,
* start
* fix
* fix
* fix imports.
* handle hunyuan
* remove residuals.
* add a check for making sure there's appropriate procs.
* add more rigor to the tests.
* fix test
* remove redundant check
* fix-copies
* move check_qkv_fusion_matches_attn_procs_length and check_qkv_fusion_processors_exist.
2024-07-24 06:52:19 +05:30
shinetzh
3b04cdc816
fix loop bug in SlicedAttnProcessor ( #8836 )
...
* fix loop bug in SlicedAttnProcessor
---------
Co-authored-by: neoshang <neoshang@tencent.com >
2024-07-19 18:14:29 -10:00
Dhruv Nair
150142c537
[Tests] Fix precision related issues in slow pipeline tests ( #8720 )
...
update
2024-06-28 08:13:46 +05:30
YiYi Xu
35358a2dec
fix offload test ( #7868 )
...
fix
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2024-05-08 07:59:08 -10:00
YiYi Xu
21a7ff12a7
update the logic of is_sequential_cpu_offload ( #7788 )
...
* up
* add comment to the tests + fix dit
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-05-01 06:25:57 -10:00
Dhruv Nair
88018fcf20
Fix failing VAE tiling test ( #7747 )
...
update
2024-04-24 12:27:45 +05:30
Fabio Rigano
b5c8b555d7
Move IP Adapter Face ID to core ( #7186 )
...
* Switch to peft and multi proj layers
* Move Face ID loading and inference to core
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-04-18 14:13:27 -10:00
UmerHA
fda1531d8a
Fixing implementation of ControlNet-XS ( #6772 )
...
* CheckIn - created DownSubBlocks
* Added extra channels, implemented subblock fwd
* Fixed connection sizes
* checkin
* Removed iter, next in forward
* Models for SD21 & SDXL run through
* Added back pipelines, cleared up connections
* Cleaned up connection creation
* added debug logs
* updated logs
* logs: added input loading
* Update umer_debug_logger.py
* log: Loading hint
* Update umer_debug_logger.py
* added logs
* Changed debug logging
* debug: added more logs
* Fixed num_norm_groups
* Debug: Logging all of SDXL input
* Update umer_debug_logger.py
* debug: updated logs
* checkim
* Readded tests
* Removed debug logs
* Fixed Slow Tests
* Added value ckecks | Updated model_cpu_offload_seq
* accelerate-offloading works ; fast tests work
* Made unet & addon explicit in controlnet
* Updated slow tests
* Added dtype/device to ControlNetXS
* Filled in test model paths
* Added image_encoder/feature_extractor to XL pipe
* Fixed fast tests
* Added comments and docstrings
* Fixed copies
* Added docs ; Updates slow tests
* Moved changes to UNetMidBlock2DCrossAttn
* tiny cleanups
* Removed stray prints
* Removed ip adapters + freeU
- Removed ip adapters + freeU as they don't make sense for ControlNet-XS
- Fixed imports of UNet components
* Fixed test_save_load_float16
* Make style, quality, fix-copies
* Changed loading/saving API for ControlNetXS
- Changed loading/saving API for ControlNetXS
- other small fixes
* Removed ControlNet-XS from research examples
* Make style, quality, fix-copies
* Small fixes
- deleted ControlNetXSModel.init_original
- added time_embedding_mix to StableDiffusionControlNetXSPipeline .from_pretrained / StableDiffusionXLControlNetXSPipeline.from_pretrained
- fixed copy hints
* checkin May 11 '23
* CheckIn Mar 12 '24
* Fixed tests for SD
* Added tests for UNetControlNetXSModel
* Fixed SDXL tests
* cleanup
* Delete Pipfile
* CheckIn Mar 20
Started replacing sub blocks by `ControlNetXSCrossAttnDownBlock2D` and `ControlNetXSCrossAttnUplock2D`
* check-in Mar 23
* checkin 24 Mar
* Created init for UNetCnxs and CnxsAddon
* CheckIn
* Made from_modules, from_unet and no_control work
* make style,quality,fix-copies & small changes
* Fixed freezing
* Added gradient ckpt'ing; fixed tests
* Fix slow tests(+compile) ; clear naming confusion
* Don't create UNet in init ; removed class_emb
* Incorporated review feedback
- Deleted get_base_pipeline / get_controlnet_addon for pipes
- Pipes inherit from StableDiffusionXLPipeline
- Made module dicts for cnxs-addon's down/mid/up classes
- Added support for qkv fusion and freeU
* Make style, quality, fix-copies
* Implemented review feedback
* Removed compatibility check for vae/ctrl embedding
* make style, quality, fix-copies
* Delete Pipfile
* Integrated review feedback
- Importing ControlNetConditioningEmbedding now
- get_down/mid/up_block_addon now outside class
- renamed `do_control` to `apply_control`
* Reduced size of test tensors
For this, added `norm_num_groups` as parameter everywhere
* Renamed cnxs-`Addon` to cnxs-`Adapter`
- `ControlNetXSAddon` -> `ControlNetXSAdapter`
- `ControlNetXSAddonDownBlockComponents` -> `DownBlockControlNetXSAdapter`, and similarly for mid/up
- `get_mid_block_addon` -> `get_mid_block_adapter`, and similarly for mid/up
* Fixed save_pretrained/from_pretrained bug
* Removed redundant code
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2024-04-16 21:56:20 +05:30
Fabio Rigano
a0cf607667
Multi-image masking for single IP Adapter ( #7499 )
...
* Support multiimage masking
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2024-04-09 09:20:57 -10:00