Aryan
c4d4ac21e7
Refactor gradient checkpointing ( #10611 )
...
* update
* remove unused fn
* apply suggestions based on review
* update + cleanup 🧹
* more cleanup 🧹
* make fix-copies
* update test
2025-01-28 06:51:46 +05:30
Yuqian Hong
4fa24591a3
create a script to train autoencoderkl ( #10605 )
...
* create a script to train vae
* update main.py
* update train_autoencoderkl.py
* update train_autoencoderkl.py
* add a check of --pretrained_model_name_or_path and --model_config_name_or_path
* remove the comment, remove diffusers in requiremnets.txt, add validation_image ote
* update autoencoderkl.py
* quality
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-01-27 16:41:34 +05:30
Sayak Paul
4ace7d0483
[chore] change licensing to 2025 from 2024. ( #10615 )
...
change licensing to 2025 from 2024.
2025-01-20 16:57:27 -10:00
baymax591
75a636da48
bugfix for npu not support float64 ( #10123 )
...
* bugfix for npu not support float64
* is_mps is_npu
---------
Co-authored-by: 白超 <baichao19@huawei.com >
Co-authored-by: hlky <hlky@hlky.ac >
2025-01-20 09:35:24 -10:00
Juan Acevedo
aeac0a00f8
implementing flux on TPUs with ptxla ( #10515 )
...
* implementing flux on TPUs with ptxla
* add xla flux attention class
* run make style/quality
* Update src/diffusers/models/attention_processor.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* Update src/diffusers/models/attention_processor.py
Co-authored-by: YiYi Xu <yixu310@gmail.com >
* run style and quality
---------
Co-authored-by: Juan Acevedo <jfacevedo@google.com >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: YiYi Xu <yixu310@gmail.com >
2025-01-16 08:46:02 -10:00
hlky
980736b792
Fix train_dreambooth_lora_sd3_miniature ( #10554 )
2025-01-13 13:47:27 +00:00
hlky
ee7e141d80
Use pipelines without vae ( #10441 )
...
* Use pipelines without vae
* getattr
* vqvae
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-01-07 13:26:51 -10:00
dependabot[bot]
e0b96ba7b0
Bump jinja2 from 3.1.4 to 3.1.5 in /examples/research_projects/realfill ( #10377 )
...
Bumps [jinja2](https://github.com/pallets/jinja ) from 3.1.4 to 3.1.5.
- [Release notes](https://github.com/pallets/jinja/releases )
- [Changelog](https://github.com/pallets/jinja/blob/main/CHANGES.rst )
- [Commits](https://github.com/pallets/jinja/compare/3.1.4...3.1.5 )
---
updated-dependencies:
- dependency-name: jinja2
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2025-01-07 19:59:41 +05:30
Rahul Raman
f1e0c7ce4a
Refactor instructpix2pix lora to support peft ( #10205 )
...
* make base code changes referred from train_instructpix2pix script in examples
* change code to use PEFT as discussed in issue 10062
* update README training command
* update README training command
* refactor variable name and freezing unet
* Update examples/research_projects/instructpix2pix_lora/train_instruct_pix2pix_lora.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* update README installation instructions.
* cleanup code using make style and quality
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2025-01-07 12:00:45 +05:30
Juan Acevedo
3cb7b8628c
Update ptxla training ( #9864 )
...
* update ptxla example
---------
Co-authored-by: Juan Acevedo <jfacevedo@google.com >
Co-authored-by: Pei Zhang <zpcore@gmail.com >
Co-authored-by: Pei Zhang <piz@google.com >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: Pei Zhang <pei@Peis-MacBook-Pro.local >
Co-authored-by: hlky <hlky@hlky.ac >
2024-12-06 10:50:13 -10:00
Parag Ekbote
cc7d88f247
Move IP Adapter Scripts to research project ( #9960 )
...
* Move files to research-projects.
* docs: add IP Adapter training instructions
* Delete venv
* Update examples/ip_adapter/tutorial_train_sdxl.py
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
* Cherry-picked commits and re-moved files
to research_projects.
* make style.
* Update toctree and delete ip_adapter.
* Nit Fix
* Fix nit.
* Fix nit.
* Create training script for single GPU and set
model format to .safetensors
* Add sample inference script and restore _toctree
* Restore toctree.yaml
* fix spacing.
* Update toctree.yaml
---------
Co-authored-by: AMohamedAakhil <a.aakhilmohamed@gmail.com >
Co-authored-by: BootesVoid <78485654+AMohamedAakhil@users.noreply.github.com >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-11-19 10:37:22 -08:00
Parag Ekbote
e255920719
Move Wuerstchen Dreambooth to research_projects ( #9935 )
...
update file paths to research_projects folder.
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-11-16 18:56:16 +05:30
Michael Tkachuk
5b972fbd6a
Enabling gradient checkpointing in eval() mode ( #9878 )
...
* refactored
2024-11-08 09:03:26 -10:00
Sayak Paul
ded3db164b
[Core] introduce controlnet module ( #8768 )
...
* move vae flax module.
* controlnet module.
* prepare for PR.
* revert a commit
* gracefully deprecate controlnet deps.
* fix
* fix doc path
* fix-copies
* fix path
* style
* style
* conflicts
* fix
* fix-copies
* sparsectrl.
* updates
* fix
* updates
* updates
* updates
* fix
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2024-11-06 22:08:55 -04:00
Sayak Paul
8ce37ab055
[training] use the lr when using 8bit adam. ( #9796 )
...
* use the lr when using 8bit adam.
* remove lr as we pack it in params_to_optimize.
---------
Co-authored-by: Linoy Tsaban <57615435+linoytsaban@users.noreply.github.com >
2024-10-31 15:51:42 +05:30
Sayak Paul
09b8aebd67
[training] fixes to the quantization training script and add AdEMAMix optimizer as an option ( #9806 )
...
* fixes
* more fixes.
2024-10-31 15:46:00 +05:30
Raul Ciotescu
c5376c5695
adds the pipeline for pixart alpha controlnet ( #8857 )
...
* add the controlnet pipeline for pixart alpha
---------
Co-authored-by: YiYi Xu <yixu310@gmail.com >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
Co-authored-by: junsongc <cjs1020440147@icloud.com >
2024-10-28 08:48:04 -10:00
Sayak Paul
fddbab7993
[research_projects] Update README.md to include a note about NF5 T5-xxl ( #9775 )
...
Update README.md
2024-10-26 22:13:03 +09:00
Sayak Paul
df073ba137
[research_projects] add flux training script with quantization ( #9754 )
...
* add flux training script with quantization
* remove exclamation
2024-10-26 00:07:57 +09:00
GSSun
164ec9f423
fix IsADirectoryError when running the training code for sd3_dreambooth_lora_16gb.ipynb ( #9634 )
...
Add files via upload
fix IsADirectoryError when running the training code
2024-10-11 13:33:39 +05:30
suzukimain
b52119ae92
[docs] Replace runwayml/stable-diffusion-v1-5 with Lykon/dreamshaper-8 ( #9428 )
...
* [docs] Replace runwayml/stable-diffusion-v1-5 with Lykon/dreamshaper-8
Updated documentation as runwayml/stable-diffusion-v1-5 has been removed from Huggingface.
* Update docs/source/en/using-diffusers/inpaint.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* Replace with stable-diffusion-v1-5/stable-diffusion-v1-5
* Update inpaint.md
---------
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
2024-09-16 10:18:45 -07:00
Juan Acevedo
45aa8bb187
Ptxla sd training ( #9381 )
...
* enable pxla training of stable diffusion 2.x models.
* run linter/style and run pipeline test for stable diffusion and fix issues.
* update xla libraries
* fix read me newline.
* move files to research folder.
* update per comments.
* rename readme.
---------
Co-authored-by: Juan Acevedo <jfacevedo@google.com >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-09-12 08:35:06 +05:30
dependabot[bot]
0c1e63bd11
Bump jinja2 from 3.1.3 to 3.1.4 in /examples/research_projects/realfill ( #7873 )
...
Bumps [jinja2](https://github.com/pallets/jinja ) from 3.1.3 to 3.1.4.
- [Release notes](https://github.com/pallets/jinja/releases )
- [Changelog](https://github.com/pallets/jinja/blob/main/CHANGES.rst )
- [Commits](https://github.com/pallets/jinja/compare/3.1.3...3.1.4 )
---
updated-dependencies:
- dependency-name: jinja2
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-08-14 16:36:59 +05:30
dependabot[bot]
e7e45bd127
Bump torch from 2.0.1 to 2.2.0 in /examples/research_projects/realfill ( #8971 )
...
Bumps [torch](https://github.com/pytorch/pytorch ) from 2.0.1 to 2.2.0.
- [Release notes](https://github.com/pytorch/pytorch/releases )
- [Changelog](https://github.com/pytorch/pytorch/blob/main/RELEASE.md )
- [Commits](https://github.com/pytorch/pytorch/compare/v2.0.1...v2.2.0 )
---
updated-dependencies:
- dependency-name: torch
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-08-14 16:36:46 +05:30
Tolga Cangöz
3dc97bd148
Update CLIPFeatureExtractor to CLIPImageProcessor and DPTFeatureExtractor to DPTImageProcessor ( #9002 )
...
* fix: update `CLIPFeatureExtractor` to `CLIPImageProcessor` in codebase
* `make style && make quality`
* Update `DPTFeatureExtractor` to `DPTImageProcessor` in codebase
* `make style`
---------
Co-authored-by: Aryan <aryan@huggingface.co >
2024-08-05 09:20:29 -10:00
omahs
6d32b29239
Fix typos ( #9077 )
...
* fix typo
2024-08-05 09:00:08 -10:00
Tolga Cangöz
7071b7461b
Errata: Fix typos & \s+$ ( #9008 )
...
* Fix typos
* chore: Fix typos
* chore: Update README.md for promptdiffusion example
* Trim trailing white spaces
* Fix a typo
* update number
* chore: update number
* Trim trailing white space
* Update README.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* Update README.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
---------
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
2024-08-02 21:24:25 -07:00
Sayak Paul
d87fe95f90
[Chore] add LoraLoaderMixin to the inits ( #8981 )
...
* introduce to promote reusability.
* up
* add more tests
* up
* remove comments.
* fix fuse_nan test
* clarify the scope of fuse_lora and unfuse_lora
* remove space
* rewrite fuse_lora a bit.
* feedback
* copy over load_lora_into_text_encoder.
* address dhruv's feedback.
* fix-copies
* fix issubclass.
* num_fused_loras
* fix
* fix
* remove mapping
* up
* fix
* style
* fix-copies
* change to SD3TransformerLoRALoadersMixin
* Apply suggestions from code review
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
* up
* handle wuerstchen
* up
* move lora to lora_pipeline.py
* up
* fix-copies
* fix documentation.
* comment set_adapters().
* fix-copies
* fix set_adapters() at the model level.
* fix?
* fix
* loraloadermixin.
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2024-07-26 08:59:33 +05:30
YiYi Xu
62863bb1ea
Revert "[LoRA] introduce LoraBaseMixin to promote reusability." ( #8976 )
...
Revert "[LoRA] introduce LoraBaseMixin to promote reusability. (#8774 )"
This reverts commit 527430d0a4 .
2024-07-25 09:10:35 -10:00
Sayak Paul
527430d0a4
[LoRA] introduce LoraBaseMixin to promote reusability. ( #8774 )
...
* introduce to promote reusability.
* up
* add more tests
* up
* remove comments.
* fix fuse_nan test
* clarify the scope of fuse_lora and unfuse_lora
* remove space
* rewrite fuse_lora a bit.
* feedback
* copy over load_lora_into_text_encoder.
* address dhruv's feedback.
* fix-copies
* fix issubclass.
* num_fused_loras
* fix
* fix
* remove mapping
* up
* fix
* style
* fix-copies
* change to SD3TransformerLoRALoadersMixin
* Apply suggestions from code review
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
* up
* handle wuerstchen
* up
* move lora to lora_pipeline.py
* up
* fix-copies
* fix documentation.
* comment set_adapters().
* fix-copies
* fix set_adapters() at the model level.
* fix?
* fix
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2024-07-25 21:40:58 +05:30
Tolga Cangöz
c1dc2ae619
Fix multi-gpu case for train_cm_ct_unconditional.py ( #8653 )
...
* Fix multi-gpu case
* Prefer previously created `unwrap_model()` function
For `torch.compile()` generalizability
* `chore: update unwrap_model() function to use accelerator.unwrap_model()`
2024-07-17 19:03:12 +05:30
Thomas Eding
2e2684f014
Add vae_roundtrip.py example ( #7104 )
...
* Add vae_roundtrip.py example
* Add cuda support to vae_roundtrip
* Move vae_roundtrip.py into research_projects/vae
* Fix channel scaling in vae roundrip and also support taesd.
* Apply ruff --fix for CI gatekeep check
---------
Co-authored-by: Álvaro Somoza <asomoza@users.noreply.github.com >
2024-07-04 01:53:09 -04:00
Sayak Paul
84bbd2f4ce
Update README.md to include Colab link ( #8775 )
2024-07-03 07:46:38 +05:30
Sayak Paul
600ef8a4dc
Allow SD3 DreamBooth LoRA fine-tuning on a free-tier Colab ( #8762 )
...
* add experimental scripts to train SD3 transformer lora on colab
* add readme
* add colab
* Apply suggestions from code review
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
* fix link in the notebook.
---------
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com >
2024-07-03 07:07:47 +05:30
Tolga Cangöz
f040c27d4c
Errata - Fix typos and improve style ( #8571 )
...
* Fix typos
* Fix typos & up style
* chore: Update numbers
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-06-24 10:07:22 -07:00
Tolga Cangöz
468ae09ed8
Errata - Trim trailing white space in the whole repo ( #8575 )
...
* Trim all the trailing white space in the whole repo
* Remove unnecessary empty places
* make style && make quality
* Trim trailing white space
* trim
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-06-24 18:39:15 +05:30
Tolga Cangöz
c375903db5
Errata - Fix typos & improve contributing page ( #8572 )
...
* Fix typos & improve contributing page
* `make style && make quality`
* fix typos
* Fix typo
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-06-24 14:13:03 +05:30
Sayak Paul
23a2cd3337
[LoRA] training fix the position of param casting when loading them ( #8460 )
...
fix the position of param casting when loading them
2024-06-18 14:57:34 +01:00
Sayak Paul
d457beed92
Update README.md to update the MaPO project ( #8470 )
...
Update README.md
2024-06-11 10:10:45 +01:00
Tolga Cangöz
98730c5dd7
Errata ( #8322 )
...
* Fix typos
* Trim trailing whitespaces
* Remove a trailing whitespace
* chore: Update MarigoldDepthPipeline checkpoint to prs-eth/marigold-lcm-v1-0
* Revert "chore: Update MarigoldDepthPipeline checkpoint to prs-eth/marigold-lcm-v1-0"
This reverts commit fd742b30b4 .
* pokemon -> naruto
* `DPMSolverMultistep` -> `DPMSolverMultistepScheduler`
* Improve Markdown stylization
* Improve style
* Improve style
* Refactor pipeline variable names for consistency
* up style
2024-06-05 13:59:09 -07:00
Hzzone
d3881f35b7
Gligen training ( #7906 )
...
* add training code of gligen
* fix code quality tests.
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-06-05 16:26:42 +04:00
Sai-Suraj-27
df2bc5ef28
fix: Fixed few docstrings according to the Google Style Guide ( #7717 )
...
Fixed few docstrings according to the Google Style Guide.
2024-05-20 10:26:05 -07:00
dependabot[bot]
5b830aa356
Bump transformers from 4.36.0 to 4.38.0 in /examples/research_projects/realfill ( #7635 )
...
Bump transformers in /examples/research_projects/realfill
Bumps [transformers](https://github.com/huggingface/transformers ) from 4.36.0 to 4.38.0.
- [Release notes](https://github.com/huggingface/transformers/releases )
- [Commits](https://github.com/huggingface/transformers/compare/v4.36.0...v4.38.0 )
---
updated-dependencies:
- dependency-name: transformers
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-05-14 08:17:06 +05:30
Mark Van Aken
be4afa0bb4
#7535 Update FloatTensor type hints to Tensor ( #7883 )
...
* find & replace all FloatTensors to Tensor
* apply formatting
* Update torch.FloatTensor to torch.Tensor in the remaining files
* formatting
* Fix the rest of the places where FloatTensor is used as well as in documentation
* formatting
* Update new file from FloatTensor to Tensor
2024-05-10 09:53:31 -10:00
Bagheera
8edaf3b79c
7879 - adjust documentation to use naruto dataset, since pokemon is now gated ( #7880 )
...
* 7879 - adjust documentation to use naruto dataset, since pokemon is now gated
* replace references to pokemon in docs
* more references to pokemon replaced
* Japanese translation update
---------
Co-authored-by: bghira <bghira@users.github.com >
2024-05-07 09:36:39 -07:00
Steven Liu
7404f1e9dc
[docs] Clean up toctree ( #7715 )
...
* toctree
* optim
* feedback
* improve overview
2024-04-23 09:30:33 -07:00
Dhruv Nair
90250d9e48
Cast height, width to int inside prepare latents ( #7691 )
...
update
2024-04-18 19:30:39 -10:00
UmerHA
fda1531d8a
Fixing implementation of ControlNet-XS ( #6772 )
...
* CheckIn - created DownSubBlocks
* Added extra channels, implemented subblock fwd
* Fixed connection sizes
* checkin
* Removed iter, next in forward
* Models for SD21 & SDXL run through
* Added back pipelines, cleared up connections
* Cleaned up connection creation
* added debug logs
* updated logs
* logs: added input loading
* Update umer_debug_logger.py
* log: Loading hint
* Update umer_debug_logger.py
* added logs
* Changed debug logging
* debug: added more logs
* Fixed num_norm_groups
* Debug: Logging all of SDXL input
* Update umer_debug_logger.py
* debug: updated logs
* checkim
* Readded tests
* Removed debug logs
* Fixed Slow Tests
* Added value ckecks | Updated model_cpu_offload_seq
* accelerate-offloading works ; fast tests work
* Made unet & addon explicit in controlnet
* Updated slow tests
* Added dtype/device to ControlNetXS
* Filled in test model paths
* Added image_encoder/feature_extractor to XL pipe
* Fixed fast tests
* Added comments and docstrings
* Fixed copies
* Added docs ; Updates slow tests
* Moved changes to UNetMidBlock2DCrossAttn
* tiny cleanups
* Removed stray prints
* Removed ip adapters + freeU
- Removed ip adapters + freeU as they don't make sense for ControlNet-XS
- Fixed imports of UNet components
* Fixed test_save_load_float16
* Make style, quality, fix-copies
* Changed loading/saving API for ControlNetXS
- Changed loading/saving API for ControlNetXS
- other small fixes
* Removed ControlNet-XS from research examples
* Make style, quality, fix-copies
* Small fixes
- deleted ControlNetXSModel.init_original
- added time_embedding_mix to StableDiffusionControlNetXSPipeline .from_pretrained / StableDiffusionXLControlNetXSPipeline.from_pretrained
- fixed copy hints
* checkin May 11 '23
* CheckIn Mar 12 '24
* Fixed tests for SD
* Added tests for UNetControlNetXSModel
* Fixed SDXL tests
* cleanup
* Delete Pipfile
* CheckIn Mar 20
Started replacing sub blocks by `ControlNetXSCrossAttnDownBlock2D` and `ControlNetXSCrossAttnUplock2D`
* check-in Mar 23
* checkin 24 Mar
* Created init for UNetCnxs and CnxsAddon
* CheckIn
* Made from_modules, from_unet and no_control work
* make style,quality,fix-copies & small changes
* Fixed freezing
* Added gradient ckpt'ing; fixed tests
* Fix slow tests(+compile) ; clear naming confusion
* Don't create UNet in init ; removed class_emb
* Incorporated review feedback
- Deleted get_base_pipeline / get_controlnet_addon for pipes
- Pipes inherit from StableDiffusionXLPipeline
- Made module dicts for cnxs-addon's down/mid/up classes
- Added support for qkv fusion and freeU
* Make style, quality, fix-copies
* Implemented review feedback
* Removed compatibility check for vae/ctrl embedding
* make style, quality, fix-copies
* Delete Pipfile
* Integrated review feedback
- Importing ControlNetConditioningEmbedding now
- get_down/mid/up_block_addon now outside class
- renamed `do_control` to `apply_control`
* Reduced size of test tensors
For this, added `norm_num_groups` as parameter everywhere
* Renamed cnxs-`Addon` to cnxs-`Adapter`
- `ControlNetXSAddon` -> `ControlNetXSAdapter`
- `ControlNetXSAddonDownBlockComponents` -> `DownBlockControlNetXSAdapter`, and similarly for mid/up
- `get_mid_block_addon` -> `get_mid_block_adapter`, and similarly for mid/up
* Fixed save_pretrained/from_pretrained bug
* Removed redundant code
---------
Co-authored-by: Dhruv Nair <dhruv.nair@gmail.com >
2024-04-16 21:56:20 +05:30
kabachuha
2f23437618
Add (Scheduled) Pseudo-Huber Loss training scripts to research projects ( #7527 )
...
* add scheduled pseudo-huber loss training scripts
See #7488
* add reduction modes to huber loss
* [DB Lora] *2 multiplier to huber loss cause of 1/2 a^2 conv.
pairing of c6495def1f
* [DB Lora] add option for smooth l1 (huber / delta)
Pairing of dd22958caa
* [DB Lora] unify huber scheduling
Pairing of 19a834c3ab
* [DB Lora] add snr huber scheduler
Pairing of 47fb1a6854
* fixup examples link
* use snr schedule by default in DB
* update all huber scripts with snr
* code quality
* huber: make style && make quality
---------
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-04-13 07:26:08 +05:30
Bagheera
8e963d1c2a
7529 do not disable autocast for cuda devices ( #7530 )
...
* 7529 do not disable autocast for cuda devices
* Remove typecasting error check for non-mps platforms, as a correct autocast implementation makes it a non-issue
* add autocast fix to other training examples
* disable native_amp for dreambooth (sdxl)
* disable native_amp for pix2pix (sdxl)
* remove tests from remaining files
* disable native_amp on huggingface accelerator for every training example that uses it
* convert more usages of autocast to nullcontext, make style fixes
* make style fixes
* style.
* Empty-Commit
---------
Co-authored-by: bghira <bghira@users.github.com >
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com >
2024-04-02 20:15:06 +05:30