Dhruv Nair
|
7aa6af1138
|
[Refactor] Move testing utils out of src (#12238)
* update
* update
* update
* update
* update
* merge main
* Revert "merge main"
This reverts commit 65efbcead5.
|
2025-08-28 19:53:02 +05:30 |
|
Sayak Paul
|
061163142d
|
[tests] tighten compilation tests for quantization (#12002)
* tighten compilation tests for quantization
* up
* up
|
2025-08-07 10:13:14 +05:30 |
|
Sayak Paul
|
7a935a0bbe
|
[tests] Unify compilation + offloading tests in quantization (#11910)
* unify the quant compile + offloading tests.
* fix
* update
|
2025-07-11 17:02:29 +05:30 |
|
Sayak Paul
|
4c20624cc6
|
[tests] annotate compilation test classes with bnb (#11715)
annotate compilation test classes with bnb
|
2025-07-09 09:24:52 +05:30 |
|
jiqing-feng
|
f3e1310469
|
reset deterministic in tearDownClass (#11785)
* reset deterministic in tearDownClass
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
* fix deterministic setting
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
---------
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
|
2025-07-01 10:06:54 +05:30 |
|
Aryan
|
cdaf84a708
|
TorchAO compile + offloading tests (#11697)
* update
* update
* update
* update
* update
* user property instead
|
2025-06-27 18:31:57 +05:30 |
|
jiqing-feng
|
ee40088fe5
|
enable deterministic in bnb 4 bit tests (#11738)
* enable deterministic in bnb 4 bit tests
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
* fix 8bit test
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
---------
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
|
2025-06-23 08:17:36 +05:30 |
|
Aryan
|
a4df8dbc40
|
Update more licenses to 2025 (#11746)
update
|
2025-06-19 07:46:01 +05:30 |
|
Sayak Paul
|
b6f7933044
|
[tests] tests for compilation + quantization (bnb) (#11672)
* start adding compilation tests for quantization.
* fixes
* make common utility.
* modularize.
* add group offloading+compile
* xfail
* update
* Update tests/quantization/test_torch_compile_utils.py
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com>
* fixes
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com>
|
2025-06-11 21:14:24 +05:30 |
|
Sayak Paul
|
8e88495da2
|
[LoRA] support Flux Control LoRA with bnb 8bit. (#11655)
support Flux Control LoRA with bnb 8bit.
|
2025-06-11 08:32:47 +05:30 |
|
jiqing-feng
|
0f91f2f6fc
|
use deterministic to get stable result (#11663)
* use deterministic to get stable result
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
* add deterministic for int8 test
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
---------
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
|
2025-06-06 09:14:00 +05:30 |
|
Sayak Paul
|
d4dc4d7654
|
[chore] misc changes in the bnb tests for consistency. (#11355)
misc changes in the bnb tests for consistency.
|
2025-06-02 08:41:10 -07:00 |
|
co63oc
|
8183d0f16e
|
Fix typos in strings and comments (#11476)
* Fix typos in strings and comments
Signed-off-by: co63oc <co63oc@users.noreply.github.com>
* Update src/diffusers/hooks/hooks.py
Co-authored-by: Aryan <contact.aryanvs@gmail.com>
* Update src/diffusers/hooks/hooks.py
Co-authored-by: Aryan <contact.aryanvs@gmail.com>
* Update layerwise_casting.py
* Apply style fixes
* update
---------
Signed-off-by: co63oc <co63oc@users.noreply.github.com>
Co-authored-by: Aryan <contact.aryanvs@gmail.com>
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
|
2025-05-30 18:49:00 +05:30 |
|
Yao Matrix
|
2d380895e5
|
enable 7 cases on XPU (#11503)
* enable 7 cases on XPU
Signed-off-by: Yao Matrix <matrix.yao@intel.com>
* calibrate A100 expectations
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
---------
Signed-off-by: Yao Matrix <matrix.yao@intel.com>
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
|
2025-05-09 15:52:08 +05:30 |
|
Yao Matrix
|
eef3d65954
|
enable 2 test cases on XPU (#11332)
* enable 2 test cases on XPU
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
* Apply style fixes
---------
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com>
|
2025-04-17 13:27:41 -10:00 |
|
Sayak Paul
|
1a04812439
|
[bistandbytes] improve replacement warnings for bnb (#11132)
* improve replacement warnings for bnb
* updates to docs.
|
2025-04-08 21:18:34 +05:30 |
|
hlky
|
5d49b3e83b
|
Flux quantized with lora (#10990)
* Flux quantized with lora
* fix
* changes
* Apply suggestions from code review
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
* Apply style fixes
* enable model cpu offload()
* Update src/diffusers/loaders/lora_pipeline.py
Co-authored-by: hlky <hlky@hlky.ac>
* update
* Apply suggestions from code review
* update
* add peft as an additional dependency for gguf
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com>
|
2025-04-08 21:17:03 +05:30 |
|
Yao Matrix
|
506f39af3a
|
enable 1 case on XPU (#11219)
enable case on XPU: 1. tests/quantization/bnb/test_mixed_int8.py::BnB8bitTrainingTests::test_training
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
|
2025-04-07 08:24:21 +01:00 |
|
jiqing-feng
|
4d5a96e40a
|
fix autocast (#11190)
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
|
2025-04-02 07:26:27 +01:00 |
|
Fanli Lin
|
5a6edac087
|
[tests] no hard-coded cuda (#11186)
no cuda only
|
2025-04-01 12:14:31 +01:00 |
|
Dhruv Nair
|
7dc52ea769
|
[Quantization] dtype fix for GGUF + fix BnB tests (#11159)
* update
* update
* update
* update
|
2025-03-26 22:22:16 +05:30 |
|
Fanli Lin
|
56f740051d
|
[tests] enable bnb tests on xpu (#11001)
* enable bnb on xpu
* add 2 more cases
* add missing change
* add missing change
* add one more
|
2025-03-19 16:33:11 +00:00 |
|
Sayak Paul
|
cc19726f3d
|
[Tests] add requires peft decorator. (#11037)
* add requires peft decorator.
* install peft conditionally.
* conditional deps.
Co-authored-by: DN6 <dhruv.nair@gmail.com>
---------
Co-authored-by: DN6 <dhruv.nair@gmail.com>
|
2025-03-15 12:56:41 +05:30 |
|
Sayak Paul
|
e7e6d85282
|
[Tests] improve quantization tests by additionally measuring the inference memory savings (#11021)
* memory usage tests
* fixes
* gguf
|
2025-03-10 21:42:24 +05:30 |
|
Marc Sun
|
f5929e0306
|
[FEAT] Model loading refactor (#10604)
* first draft model loading refactor
* revert name change
* fix bnb
* revert name
* fix dduf
* fix huanyan
* style
* Update src/diffusers/models/model_loading_utils.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
* suggestions from reviews
* Update src/diffusers/models/modeling_utils.py
Co-authored-by: YiYi Xu <yixu310@gmail.com>
* remove safetensors check
* fix default value
* more fix from suggestions
* revert logic for single file
* style
* typing + fix couple of issues
* improve speed
* Update src/diffusers/models/modeling_utils.py
Co-authored-by: Aryan <aryan@huggingface.co>
* fp8 dtype
* add tests
* rename resolved_archive_file to resolved_model_file
* format
* map_location default cpu
* add utility function
* switch to smaller model + test inference
* Apply suggestions from code review
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
* rm comment
* add log
* Apply suggestions from code review
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
* add decorator
* cosine sim instead
* fix use_keep_in_fp32_modules
* comm
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
Co-authored-by: YiYi Xu <yixu310@gmail.com>
Co-authored-by: Aryan <aryan@huggingface.co>
|
2025-02-19 17:34:53 +05:30 |
|
hlky
|
0b065c099a
|
Move buffers to device (#10523)
* Move buffers to device
* add test
* named_buffers
|
2025-01-16 07:42:56 -10:00 |
|
Sayak Paul
|
bba59fb88b
|
[Tests] add: test to check 8bit bnb quantized models work with lora loading. (#10576)
* add: test to check 8bit bnb quantized models work with lora loading.
* Update tests/quantization/bnb/test_mixed_int8.py
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com>
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com>
|
2025-01-15 13:05:26 +05:30 |
|
Sayak Paul
|
2432f80ca3
|
[LoRA] feat: support loading loras into 4bit quantized Flux models. (#10578)
* feat: support loading loras into 4bit quantized models.
* updates
* update
* remove weight check.
|
2025-01-15 12:40:40 +05:30 |
|
Sayak Paul
|
9f06a0d1a4
|
[CI] Match remaining assertions from big runner (#10521)
* print
* remove print.
* print
* update slice.
* empty
|
2025-01-10 16:37:36 +05:30 |
|
Sayak Paul
|
e8da75dff5
|
[bitsandbytes] allow directly CUDA placements of pipelines loaded with bnb components (#9840)
* allow device placement when using bnb quantization.
* warning.
* tests
* fixes
* docs.
* require accelerate version.
* remove print.
* revert to()
* tests
* fixes
* fix: missing AutoencoderKL lora adapter (#9807)
* fix: missing AutoencoderKL lora adapter
* fix
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
* fixes
* fix condition test
* updates
* updates
* remove is_offloaded.
* fixes
* better
* empty
---------
Co-authored-by: Emmanuel Benazera <emmanuel.benazera@jolibrain.com>
|
2024-12-04 22:27:43 +05:30 |
|
Sayak Paul
|
827b6c25f9
|
[CI] Add quantization (#9832)
* add quantization to nightly CI.
* prep.
* fix lib name.
* remove deps that are not needed.
* fix slice.
|
2024-12-02 14:53:43 +05:30 |
|
Sayak Paul
|
60ffa84253
|
[bitsandbbytes] follow-ups (#9730)
* bnb follow ups.
* add a warning when dtypes mismatch.
* fx-copies
* clear cache.
* check_if_quantized_param
* add a check on shape.
* updates
* docs
* improve readability.
* resources.
* fix
|
2024-10-22 16:00:05 +05:30 |
|
Sayak Paul
|
b821f006d0
|
[Quantization] Add quantization support for bitsandbytes (#9213)
* quantization config.
* fix-copies
* fix
* modules_to_not_convert
* add bitsandbytes utilities.
* make progress.
* fixes
* quality
* up
* up
rotary embedding refactor 2: update comments, fix dtype for use_real=False (#9312)
fix notes and dtype
up
up
* minor
* up
* up
* fix
* provide credits where due.
* make configurations work.
* fixes
* fix
* update_missing_keys
* fix
* fix
* make it work.
* fix
* provide credits to transformers.
* empty commit
* handle to() better.
* tests
* change to bnb from bitsandbytes
* fix tests
fix slow quality tests
SD3 remark
fix
complete int4 tests
add a readme to the test files.
add model cpu offload tests
warning test
* better safeguard.
* change merging status
* courtesy to transformers.
* move upper.
* better
* make the unused kwargs warning friendlier.
* harmonize changes with https://github.com/huggingface/transformers/pull/33122
* style
* trainin tests
* feedback part i.
* Add Flux inpainting and Flux Img2Img (#9135)
---------
Co-authored-by: yiyixuxu <yixu310@gmail.com>
Update `UNet2DConditionModel`'s error messages (#9230)
* refactor
[CI] Update Single file Nightly Tests (#9357)
* update
* update
feedback.
improve README for flux dreambooth lora (#9290)
* improve readme
* improve readme
* improve readme
* improve readme
fix one uncaught deprecation warning for accessing vae_latent_channels in VaeImagePreprocessor (#9372)
deprecation warning vae_latent_channels
add mixed int8 tests and more tests to nf4.
[core] Freenoise memory improvements (#9262)
* update
* implement prompt interpolation
* make style
* resnet memory optimizations
* more memory optimizations; todo: refactor
* update
* update animatediff controlnet with latest changes
* refactor chunked inference changes
* remove print statements
* update
* chunk -> split
* remove changes from incorrect conflict resolution
* remove changes from incorrect conflict resolution
* add explanation of SplitInferenceModule
* update docs
* Revert "update docs"
This reverts commit c55a50a271.
* update docstring for freenoise split inference
* apply suggestions from review
* add tests
* apply suggestions from review
quantization docs.
docs.
* Revert "Add Flux inpainting and Flux Img2Img (#9135)"
This reverts commit 5799954dd4.
* tests
* don
* Apply suggestions from code review
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
* contribution guide.
* changes
* empty
* fix tests
* harmonize with https://github.com/huggingface/transformers/pull/33546.
* numpy_cosine_distance
* config_dict modification.
* remove if config comment.
* note for load_state_dict changes.
* float8 check.
* quantizer.
* raise an error for non-True low_cpu_mem_usage values when using quant.
* low_cpu_mem_usage shenanigans when using fp32 modules.
* don't re-assign _pre_quantization_type.
* make comments clear.
* remove comments.
* handle mixed types better when moving to cpu.
* add tests to check if we're throwing warning rightly.
* better check.
* fix 8bit test_quality.
* handle dtype more robustly.
* better message when keep_in_fp32_modules.
* handle dtype casting.
* fix dtype checks in pipeline.
* fix warning message.
* Update src/diffusers/models/modeling_utils.py
Co-authored-by: YiYi Xu <yixu310@gmail.com>
* mitigate the confusing cpu warning
---------
Co-authored-by: Vishnu V Jaddipal <95531133+Gothos@users.noreply.github.com>
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
Co-authored-by: YiYi Xu <yixu310@gmail.com>
|
2024-10-21 10:11:57 +05:30 |
|