kaixuanliu
dd285099eb
adjust to get CI test cases passed on XPU ( #11759 )
...
* adjust to get CI test cases passed on XPU
Signed-off-by: Liu, Kaixuan <kaixuan.liu@intel.com >
* fix format issue
Signed-off-by: Liu, Kaixuan <kaixuan.liu@intel.com >
* Apply style fixes
---------
Signed-off-by: Liu, Kaixuan <kaixuan.liu@intel.com >
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
Co-authored-by: Aryan <aryan@huggingface.co >
2025-06-25 14:02:17 +05:30
Sayak Paul
62cce3045d
[chore] change to 2025 licensing for remaining ( #11741 )
...
change to 2025 licensing for remaining
2025-06-18 20:56:00 +05:30
Yao Matrix
c3726153fd
enable several pipeline integration tests on XPU ( #11526 )
...
* enable kandinsky2_2 integration test cases on XPU
Signed-off-by: Yao Matrix <matrix.yao@intel.com >
* fix style
Signed-off-by: Yao Matrix <matrix.yao@intel.com >
* enable latent_diffusion, dance_diffusion, musicldm, shap_e integration
uts on xpu
Signed-off-by: Yao Matrix <matrix.yao@intel.com >
* fix style
Signed-off-by: Yao Matrix <matrix.yao@intel.com >
---------
Signed-off-by: Yao Matrix <matrix.yao@intel.com >
Co-authored-by: Aryan <aryan@huggingface.co >
2025-05-12 16:51:37 +05:30
Yao Matrix
aa541b9fab
make KandinskyV22PipelineInpaintCombinedFastTests::test_float16_inference pass on XPU ( #11308 )
...
loose expected_max_diff from 5e-1 to 8e-1 to make
KandinskyV22PipelineInpaintCombinedFastTests::test_float16_inference
pass on XPU
Signed-off-by: Matrix Yao <matrix.yao@intel.com >
2025-04-14 07:49:20 +01:00
Dhruv Nair
edc154da09
Update Ruff to latest Version ( #10919 )
...
* update
* update
* update
* update
2025-04-09 16:51:34 +05:30
Fanli Lin
7855ac597e
[tests] make tests device-agnostic (part 4) ( #10508 )
...
* initial comit
* fix empty cache
* fix one more
* fix style
* update device functions
* update
* update
* Update src/diffusers/utils/testing_utils.py
Co-authored-by: hlky <hlky@hlky.ac >
* Update src/diffusers/utils/testing_utils.py
Co-authored-by: hlky <hlky@hlky.ac >
* Update src/diffusers/utils/testing_utils.py
Co-authored-by: hlky <hlky@hlky.ac >
* Update tests/pipelines/controlnet/test_controlnet.py
Co-authored-by: hlky <hlky@hlky.ac >
* Update src/diffusers/utils/testing_utils.py
Co-authored-by: hlky <hlky@hlky.ac >
* Update src/diffusers/utils/testing_utils.py
Co-authored-by: hlky <hlky@hlky.ac >
* Update tests/pipelines/controlnet/test_controlnet.py
Co-authored-by: hlky <hlky@hlky.ac >
* with gc.collect
* update
* make style
* check_torch_dependencies
* add mps empty cache
* add changes
* bug fix
* enable on xpu
* update more cases
* revert
* revert back
* Update test_stable_diffusion_xl.py
* Update tests/pipelines/stable_diffusion/test_stable_diffusion.py
Co-authored-by: hlky <hlky@hlky.ac >
* Update tests/pipelines/stable_diffusion/test_stable_diffusion.py
Co-authored-by: hlky <hlky@hlky.ac >
* Update tests/pipelines/stable_diffusion/test_stable_diffusion_img2img.py
Co-authored-by: hlky <hlky@hlky.ac >
* Update tests/pipelines/stable_diffusion/test_stable_diffusion_img2img.py
Co-authored-by: hlky <hlky@hlky.ac >
* Update tests/pipelines/stable_diffusion/test_stable_diffusion_img2img.py
Co-authored-by: hlky <hlky@hlky.ac >
* Apply suggestions from code review
Co-authored-by: hlky <hlky@hlky.ac >
* add test marker
---------
Co-authored-by: hlky <hlky@hlky.ac >
2025-03-04 08:26:06 +00:00
Marc Sun
fbff43acc9
[FEAT] DDUF format ( #10037 )
...
* load and save dduf archive
* style
* switch to zip uncompressed
* updates
* Update src/diffusers/pipelines/pipeline_utils.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* Update src/diffusers/pipelines/pipeline_utils.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* first draft
* remove print
* switch to dduf_file for consistency
* switch to huggingface hub api
* fix log
* add a basic test
* Update src/diffusers/configuration_utils.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* Update src/diffusers/pipelines/pipeline_utils.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* Update src/diffusers/pipelines/pipeline_utils.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* fix
* fix variant
* change saving logic
* DDUF - Load transformers components manually (#10171 )
* update hfh version
* Load transformers components manually
* load encoder from_pretrained with state_dict
* working version with transformers and tokenizer !
* add generation_config case
* fix tests
* remove saving for now
* typing
* need next version from transformers
* Update src/diffusers/configuration_utils.py
Co-authored-by: Lucain <lucain@huggingface.co >
* check path corectly
* Apply suggestions from code review
Co-authored-by: Lucain <lucain@huggingface.co >
* udapte
* typing
* remove check for subfolder
* quality
* revert setup changes
* oups
* more readable condition
* add loading from the hub test
* add basic docs.
* Apply suggestions from code review
Co-authored-by: Lucain <lucain@huggingface.co >
* add example
* add
* make functions private
* Apply suggestions from code review
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* minor.
* fixes
* fix
* change the precdence of parameterized.
* error out when custom pipeline is passed with dduf_file.
* updates
* fix
* updates
* fixes
* updates
* fix xfail condition.
* fix xfail
* fixes
* sharded checkpoint compat
* add test for sharded checkpoint
* add suggestions
* Update src/diffusers/models/model_loading_utils.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* from suggestions
* add class attributes to flag dduf tests
* last one
* fix logic
* remove comment
* revert changes
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: Lucain <lucain@huggingface.co >
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2025-01-14 13:21:42 +05:30
YiYi Xu
d2e5cb3c10
Revert "[LoRA] fix: lora loading when using with a device_mapped mode… ( #9823 )
...
Revert "[LoRA] fix: lora loading when using with a device_mapped model. (#9449 )"
This reverts commit 41e4779d98 .
2024-10-31 08:19:32 -10:00
Sayak Paul
41e4779d98
[LoRA] fix: lora loading when using with a device_mapped model. ( #9449 )
...
* fix: lora loading when using with a device_mapped model.
* better attibutung
* empty
Co-authored-by: Benjamin Bossan <BenjaminBossan@users.noreply.github.com >
* Apply suggestions from code review
Co-authored-by: Marc Sun <57196510+SunMarc@users.noreply.github.com >
* minors
* better error messages.
* fix-copies
* add: tests, docs.
* add hardware note.
* quality
* Update docs/source/en/training/distributed_inference.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* fixes
* skip properly.
* fixes
---------
Co-authored-by: Benjamin Bossan <BenjaminBossan@users.noreply.github.com >
Co-authored-by: Marc Sun <57196510+SunMarc@users.noreply.github.com >
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
2024-10-31 21:17:41 +05:30
Sayak Paul
d8bcb33f4b
[Tests] fix slices of 26 tests (first half) ( #8959 )
...
* check for assertions.
* update with correct slices.
* okay
* style
* get it ready
* update
* update
* update
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2024-07-25 14:56:49 +05:30
Dhruv Nair
a9dd86029e
Fix Kandinksy V22 tests ( #7699 )
...
update
2024-04-22 15:41:59 +05:30
Dhruv Nair
9a34953823
Additional Memory clean up for slow tests ( #7436 )
...
* update
* update
* update
2024-03-25 12:19:21 +05:30
Sayak Paul
30e5e81d58
change to 2024 in the license ( #6902 )
...
change to 2024
2024-02-08 08:19:31 -10:00
Sayak Paul
0fc62d1702
[Kandinsky tests] add is_flaky to test_model_cpu_offload_forward_pass ( #6762 )
...
* add is_flaky to test_model_cpu_offload_forward_pass
* style
* update
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2024-01-31 14:51:12 +05:30
YiYi Xu
2b23ec82e8
add callbacks to denoising step ( #5427 )
...
* draft1
* update
* style
* move to the end of loop
* update
* update callbak_on_step_end_inputs
* Revert "update"
This reverts commit 5f9b153183 .
* Revert "update callbak_on_step_end_inputs"
This reverts commit 44889f4dab .
* update
* update test required_optional_params
* remove self.lora_scale
* img2img
* inpaint
* Apply suggestions from code review
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com >
* fix
* apply feedbacks on img2img + inpaint: keep only important pipeline attributes
* depth
* pix2pix
* make _callback_tensor_inputs an class variable so that we can use it for testing
* add a basic tst for callback
* add a read-only tensor input timesteps + fix tests
* add second test for callback cfg
* sdxl
* sdxl img2img
* sdxl inpaint
* kandinsky prior
* kandinsky decoder
* kandinsky img2img + combined
* kandinsky inpaint
* fix copies
* fix
* consistent default inputs
* fix copies
* wuerstchen_prior prior
* test_wuerstchen_decoder + fix test for prior
* wuerstchen_combined pipeline + skip tests
* skip test for kandinsky combined
* lcm
* remove timesteps etc
* add doc string
* copies
* Update src/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* Update src/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* make style and improve tests
* up
* up
* fix more
* fix cfg test
* tests for callbacks
* fix for real
* update
* lcm img2img
* add doc
* add doc page to index
---------
Co-authored-by: yiyixuxu <yixu310@gmail,com>
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2023-11-05 20:00:41 +01:00
Dhruv Nair
dd5a36291f
New Pipeline Slow Test runners ( #5131 )
...
* pipline fetcher
* update script
* clean up
* clean up
* clean up
* new pipeline runner
* rename tests to match modules
* test actions in pr
* change runner to gpu
* clean up
* clean up
* clean up
* fix report
* fix reporting
* clean up
* show test stats in failure reports
* give names to jobs
* add lora tests
* split torch cuda tests and add compile tests
* clean up
* fix tests
* change push to run only on main
---------
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com >
2023-10-04 11:42:17 +02:00