Linoy Tsaban
941ac9c3d9
[training-scripts] Make more examples UV-compatible (follow up on #12000 ) ( #12407 )
...
* make qwen and kontext uv compatible
* add torchvision
* add torchvision
* add datasets, bitsandbytes, prodigyopt
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-10-03 17:46:47 +03:00
Leo Jiang
c222570a9b
DeepSpeed adaption for flux-kontext ( #12240 )
...
Co-authored-by: Jη³ι‘΅ <jiangshuo9@h-partners.com >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-09-09 12:58:08 +05:30
Linoy Tsaban
006d092751
[Flux LoRA] fix for prior preservation and mixed precision sampling, follow up on #11873 ( #12264 )
...
* propagate fixes from https://github.com/huggingface/diffusers/pull/11873/ to flux script
* propagate fixes from https://github.com/huggingface/diffusers/pull/11873/ to flux script
* propagate fixes from https://github.com/huggingface/diffusers/pull/11873/ to flux script
* Apply style fixes
---------
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-09-02 11:30:33 +03:00
Bulat Akhmatov
0ff1aa910c
[fix] fix for prior preservation and mixed precision sampling ( #11873 )
...
Co-authored-by: Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-09-01 14:12:14 +03:00
Leo Jiang
0fd7ee79ea
NPU attention refactor for FLUX ( #12209 )
...
* NPU attention refactor for FLUX transformer
* Apply style fixes
---------
Co-authored-by: Jη³ι‘΅ <jiangshuo9@h-partners.com >
Co-authored-by: Aryan <aryan@huggingface.co >
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-08-26 12:53:55 +05:30
Meta
0e46c55931
Update README.md ( #12193 )
2025-08-25 11:35:56 -07:00
Sayak Paul
7a2b78bf0f
post release v0.35.0 ( #12184 )
...
* post release v0.35.0
* quality
2025-08-19 22:10:08 +05:30
Taechai
3c50f0cdad
Update README.md ( #12182 )
...
* Update README.md
Specify the full dir
* Update examples/dreambooth/README.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
---------
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
2025-08-18 13:02:49 -07:00
Leo Jiang
03be15e890
[Docs] typo error in qwen image ( #12144 )
...
typo error in qwen image
Co-authored-by: Jη³ι‘΅ <jiangshuo9@h-partners.com >
Co-authored-by: Aryan <aryan@huggingface.co >
2025-08-18 11:55:42 +05:30
Sayak Paul
c9c8217306
[chore] complete the licensing statement. ( #12001 )
...
complete the licensing statement.
2025-08-11 22:15:15 +05:30
Sayak Paul
5937e11d85
[docs] small corrections to the example in the Qwen docs ( #12068 )
...
* up
* up
2025-08-05 09:47:21 +05:30
Sayak Paul
9c1d4e3be1
[wip] feat: support lora in qwen image and training script ( #12056 )
...
* feat: support lora in qwen image and training script
* up
* up
* up
* up
* up
* up
* add lora tests
* fix
* add tests
* fix
* reviewer feedback
* up[
* Apply suggestions from code review
Co-authored-by: Aryan <aryan@huggingface.co >
---------
Co-authored-by: Aryan <aryan@huggingface.co >
2025-08-05 07:06:02 +05:30
Sayak Paul
20e0740b88
[training-scripts] Make pytorch examples UV-compatible ( #12000 )
...
* add uv dependencies on top of scripts.
* add uv deps.
2025-07-31 22:09:52 +05:30
Γlvaro Somoza
edcbe8038b
Fix huggingface-hub failing tests ( #11994 )
...
* login
* more logins
* uploads
* missed login
* another missed login
* downloads
* examples and more logins
* fix
* setup
* Apply style fixes
* fix
* Apply style fixes
2025-07-29 02:34:58 -04:00
Chengxi Guo
cde02b061b
Fix kontext finetune issue when batch size >1 ( #11921 )
...
set drop_last to True
Signed-off-by: mymusise <mymusise1@gmail.com >
2025-07-18 19:38:58 -04:00
Sayak Paul
9c13f86579
[training] add an offload utility that can be used as a context manager. ( #11775 )
...
* add an offload utility that can be used as a context manager.
* update
---------
Co-authored-by: Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com >
2025-07-16 09:09:13 +01:00
Sayak Paul
01240fecb0
[training ] add Kontext i2i training ( #11858 )
...
* feat: enable i2i fine-tuning in Kontext script.
* readme
* more checks.
* Apply suggestions from code review
Co-authored-by: Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com >
* fixes
* fix
* add proj_mlp to the mix
* Update README_flux.md
add note on installing from commit `05e7a854d0a5661f5b433f6dd5954c224b104f0b`
* fix
* fix
---------
Co-authored-by: Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com >
2025-07-08 21:04:16 +05:30
Sayak Paul
00f95b9755
Kontext training ( #11813 )
...
* support flux kontext
* make fix-copies
* add example
* add tests
* update docs
* update
* add note on integrity checker
* initial commit
* initial commit
* add readme section and fixes in the training script.
* add test
* rectify ckpt_id
* fix ckpt
* fixes
* change id
* update
* Update examples/dreambooth/train_dreambooth_lora_flux_kontext.py
Co-authored-by: Aryan <aryan@huggingface.co >
* Update examples/dreambooth/README_flux.md
---------
Co-authored-by: Aryan <aryan@huggingface.co >
Co-authored-by: linoytsaban <linoy@huggingface.co >
Co-authored-by: Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com >
2025-06-26 19:31:42 +03:00
Sayak Paul
10c36e0b78
[chore] post release v0.34.0 ( #11800 )
...
* post release v0.34.0
* code quality
---------
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2025-06-26 06:56:46 +05:30
imbr92
6760300202
Add --lora_alpha and metadata handling to train_dreambooth_lora_sana.py ( #11744 )
...
Co-authored-by: Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com >
2025-06-23 15:46:44 +03:00
Sayak Paul
62cce3045d
[chore] change to 2025 licensing for remaining ( #11741 )
...
change to 2025 licensing for remaining
2025-06-18 20:56:00 +05:30
Leo Jiang
d72184eba3
[training] add ds support to lora hidream ( #11737 )
...
* [training] add ds support to lora hidream
* Apply style fixes
---------
Co-authored-by: Jη³ι‘΅ <jiangshuo9@h-partners.com >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-06-18 09:26:02 +05:30
Linoy Tsaban
1bc6f3dc0f
[LoRA training] update metadata use for lora alpha + README ( #11723 )
...
* lora alpha
* Apply style fixes
* Update examples/advanced_diffusion_training/README_flux.md
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* fix readme format
---------
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-06-17 12:19:27 +03:00
Sayak Paul
f0dba33d82
[training] show how metadata stuff should be incorporated in training scripts. ( #11707 )
...
* show how metadata stuff should be incorporated in training scripts.
* typing
* fix
---------
Co-authored-by: Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com >
2025-06-16 16:42:34 +05:30
Linoy Tsaban
cc59505e26
[training docs] smol update to README files ( #11616 )
...
add comment to install prodigy
2025-05-27 06:26:54 -07:00
Quentin GallouΓ©dec
c8bb1ff53e
Use HF Papers ( #11567 )
...
* Use HF Papers
* Apply style fixes
---------
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-05-19 06:22:33 -10:00
Kenneth Gerald Hamilton
07dd6f8c0e
[train_dreambooth.py] Fix the LR Schedulers when num_train_epochs is passed in a distributed training env ( #11239 )
...
Co-authored-by: Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-05-13 07:34:01 +05:30
Linoy Tsaban
66e50d4e24
[LoRA] make lora alpha and dropout configurable ( #11467 )
...
* add lora_alpha and lora_dropout
* Apply style fixes
* add lora_alpha and lora_dropout
* Apply style fixes
* revert lora_alpha until #11324 is merged
* Apply style fixes
* empty commit
---------
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-05-08 11:54:50 +03:00
Sayak Paul
071807c853
[training] feat: enable quantization for hidream lora training. ( #11494 )
...
* feat: enable quantization for hidream lora training.
* better handle compute dtype.
* finalize.
* fix dtype.
---------
Co-authored-by: Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com >
2025-05-05 20:44:35 +05:30
Evan Han
ee1516e5c7
[train_dreambooth_lora_lumina2] Add LANCZOS as the default interpolation mode for image resizing ( #11491 )
...
[ADD] interpolation
2025-05-05 10:41:33 -04:00
MinJu-Ha
ec9323996b
[train_dreambooth_lora_sdxl] Add --image_interpolation_mode option for image resizing (default to lanczos) ( #11490 )
...
feat(train_dreambooth_lora_sdxl): support --image_interpolation_mode with default to lanczos
2025-05-05 10:19:30 -04:00
co63oc
86294d3c7f
Fix typos in docs and comments ( #11416 )
...
* Fix typos in docs and comments
* Apply style fixes
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-04-30 20:30:53 -10:00
Linoy Tsaban
0ac1d5b482
[Hi-Dream LoRA] fix bug in validation ( #11439 )
...
remove unnecessary pipeline moving to cpu in validation
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-04-28 06:22:32 -10:00
Mert Erbak
bd96a084d3
[train_dreambooth_lora.py] Set LANCZOS as default interpolation mode for resizing ( #11421 )
...
* Set LANCZOS as default interpolation mode for resizing
* [train_dreambooth_lora.py] Set LANCZOS as default interpolation mode for resizing
2025-04-26 01:58:41 -04:00
co63oc
f00a995753
Fix typos in strings and comments ( #11407 )
2025-04-24 08:53:47 -10:00
Linoy Tsaban
edd7880418
[HiDream LoRA] optimizations + small updates ( #11381 )
...
* 1. add pre-computation of prompt embeddings when custom prompts are used as well
2. save model card even if model is not pushed to hub
3. remove scheduler initialization from code example - not necessary anymore (it's now if the base model's config)
4. add skip_final_inference - to allow to run with validation, but skip the final loading of the pipeline with the lora weights to reduce memory reqs
* pre encode validation prompt as well
* Update examples/dreambooth/train_dreambooth_lora_hidream.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* Update examples/dreambooth/train_dreambooth_lora_hidream.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* Update examples/dreambooth/train_dreambooth_lora_hidream.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* pre encode validation prompt as well
* Apply style fixes
* empty commit
* change default trained modules
* empty commit
* address comments + change encoding of validation prompt (before it was only pre-encoded if custom prompts are provided, but should be pre-encoded either way)
* Apply style fixes
* empty commit
* fix validation_embeddings definition
* fix final inference condition
* fix pipeline deletion in last inference
* Apply style fixes
* empty commit
* layers
* remove readme remarks on only pre-computing when instance prompt is provided and change example to 3d icons
* smol fix
* empty commit
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-04-24 07:48:19 +03:00
Ishan Dutta
4b60f4b602
[train_dreambooth_flux] Add LANCZOS as the default interpolation mode for image resizing ( #11395 )
2025-04-23 10:47:05 -04:00
Ameer Azam
026507c06c
Update README_hidream.md ( #11386 )
...
Small change
requirements_sana.txt to
requirements_hidream.txt
2025-04-22 20:08:26 -04:00
Linoy Tsaban
e30d3bf544
[LoRA] add LoRA support to HiDream and fine-tuning script ( #11281 )
...
* initial commit
* initial commit
* initial commit
* initial commit
* initial commit
* initial commit
* Update examples/dreambooth/train_dreambooth_lora_hidream.py
Co-authored-by: Bagheera <59658056+bghira@users.noreply.github.com >
* move prompt embeds, pooled embeds outside
* Update examples/dreambooth/train_dreambooth_lora_hidream.py
Co-authored-by: hlky <hlky@hlky.ac >
* Update examples/dreambooth/train_dreambooth_lora_hidream.py
Co-authored-by: hlky <hlky@hlky.ac >
* fix import
* fix import and tokenizer 4, text encoder 4 loading
* te
* prompt embeds
* fix naming
* shapes
* initial commit to add HiDreamImageLoraLoaderMixin
* fix init
* add tests
* loader
* fix model input
* add code example to readme
* fix default max length of text encoders
* prints
* nullify training cond in unpatchify for temp fix to incompatible shaping of transformer output during training
* smol fix
* unpatchify
* unpatchify
* fix validation
* flip pred and loss
* fix shift!!!
* revert unpatchify changes (for now)
* smol fix
* Apply style fixes
* workaround moe training
* workaround moe training
* remove prints
* to reduce some memory, keep vae in `weight_dtype` same as we have for flux (as it's the same vae)
bbd0c161b5/examples/dreambooth/train_dreambooth_lora_flux.py (L1207)
* refactor to align with HiDream refactor
* refactor to align with HiDream refactor
* refactor to align with HiDream refactor
* add support for cpu offloading of text encoders
* Apply style fixes
* adjust lr and rank for train example
* fix copies
* Apply style fixes
* update README
* update README
* update README
* fix license
* keep prompt2,3,4 as None in validation
* remove reverse ode comment
* Update examples/dreambooth/train_dreambooth_lora_hidream.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* Update examples/dreambooth/train_dreambooth_lora_hidream.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* vae offload change
* fix text encoder offloading
* Apply style fixes
* cleaner to_kwargs
* fix module name in copied from
* add requirements
* fix offloading
* fix offloading
* fix offloading
* update transformers version in reqs
* try AutoTokenizer
* try AutoTokenizer
* Apply style fixes
* empty commit
* Delete tests/lora/test_lora_layers_hidream.py
* change tokenizer_4 to load with AutoTokenizer as well
* make text_encoder_four and tokenizer_four configurable
* save model card
* save model card
* revert T5
* fix test
* remove non diffusers lumina2 conversion
---------
Co-authored-by: Bagheera <59658056+bghira@users.noreply.github.com >
Co-authored-by: hlky <hlky@hlky.ac >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-04-22 11:44:02 +03:00
Kenneth Gerald Hamilton
0dec414d5b
[train_dreambooth_lora_sdxl.py] Fix the LR Schedulers when num_train_epochs is passed in a distributed training env ( #11240 )
...
Co-authored-by: Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com >
2025-04-21 12:51:03 +05:30
Linoy Tsaban
44eeba07b2
[Flux LoRAs] fix lr scheduler bug in distributed scenarios ( #11242 )
...
* add fix
* add fix
* Apply style fixes
---------
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-04-21 10:08:45 +03:00
Sayak Paul
4b868f14c1
post release 0.33.0 ( #11255 )
...
* post release
* update
* fix deprecations
* remaining
* update
---------
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2025-04-15 06:50:08 -10:00
Dhruv Nair
edc154da09
Update Ruff to latest Version ( #10919 )
...
* update
* update
* update
* update
2025-04-09 16:51:34 +05:30
Linoy Tsaban
71f34fc5a4
[Flux LoRA] fix issues in flux lora scripts ( #11111 )
...
* remove custom scheduler
* update requirements.txt
* log_validation with mixed precision
* add intermediate embeddings saving when checkpointing is enabled
* remove comment
* fix validation
* add unwrap_model for accelerator, torch.no_grad context for validation, fix accelerator.accumulate call in advanced script
* revert unwrap_model change temp
* add .module to address distributed training bug + replace accelerator.unwrap_model with unwrap model
* changes to align advanced script with canonical script
* make changes for distributed training + unify unwrap_model calls in advanced script
* add module.dtype fix to dreambooth script
* unify unwrap_model calls in dreambooth script
* fix condition in validation run
* mixed precision
* Update examples/advanced_diffusion_training/train_dreambooth_lora_flux_advanced.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* smol style change
* change autocast
* Apply style fixes
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-04-08 17:40:30 +03:00
Γlvaro Somoza
723dbdd363
[Training] Better image interpolation in training scripts ( #11206 )
...
* initial
* Update examples/dreambooth/train_dreambooth_lora_sdxl.py
Co-authored-by: hlky <hlky@hlky.ac >
* update
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: hlky <hlky@hlky.ac >
2025-04-08 12:26:07 +05:30
Jun Yeop Na
37b8edfb86
[train_dreambooth_lora.py] Fix the LR Schedulers when num_train_epochs is passed in a distributed training env ( #10973 )
...
* updated train_dreambooth_lora to fix the LR schedulers for `num_train_epochs` in distributed training env
* fixed formatting
* remove trailing newlines
* fixed style error
2025-03-06 10:06:24 +05:30
Alexey Zolotenkov
b8215b1c06
Fix incorrect seed initialization when args.seed is 0 ( #10964 )
...
* Fix seed initialization to handle args.seed = 0 correctly
* Apply style fixes
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2025-03-04 10:09:52 -10:00
SahilCarterr
170833c22a
[Fix] fp16 unscaling in train_dreambooth_lora_sdxl ( #10889 )
...
Fix fp16 bug
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-02-24 06:49:23 -10:00
Sayak Paul
f10d3c6d04
[LoRA] add LoRA support to Lumina2 and fine-tuning script ( #10818 )
...
* feat: lora support for Lumina2.
* fix-copies.
* updates
* updates
* docs.
* fix
* add: training script.
* tests
* updates
* updates
* major updates.
* updates
* fixes
* docs.
* updates
* updates
2025-02-20 09:41:51 +05:30
Leo Jiang
cd0a4a82cf
[bugfix] NPU Adaption for Sana ( #10724 )
...
* NPU Adaption for Sanna
* NPU Adaption for Sanna
* NPU Adaption for Sanna
* NPU Adaption for Sanna
* NPU Adaption for Sanna
* NPU Adaption for Sanna
* NPU Adaption for Sanna
* NPU Adaption for Sanna
* NPU Adaption for Sanna
* NPU Adaption for Sanna
* NPU Adaption for Sanna
* NPU Adaption for Sanna
* NPU Adaption for Sanna
* NPU Adaption for Sanna
* NPU Adaption for Sanna
* NPU Adaption for Sanna
* NPU Adaption for Sanna
* NPU Adaption for Sanna
* [bugfix]NPU Adaption for Sanna
---------
Co-authored-by: Jη³ι‘΅ <jiangshuo9@h-partners.com >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-02-06 19:29:58 +05:30