1
0
mirror of https://github.com/huggingface/diffusers.git synced 2026-01-27 17:22:53 +03:00
This commit is contained in:
sayakpaul
2025-10-03 16:56:31 +05:30
parent 1737b710a2
commit c4bcf72084
14 changed files with 106 additions and 224 deletions

View File

@@ -99,42 +99,34 @@ class TestAuraFlowLoRA(PeftLoraLoaderMixinTests):
return noise, input_ids, pipeline_inputs
pytest.mark.skip("Not supported in AuraFlow.")
@pytest.mark.skip("Not supported in AuraFlow.")
def test_simple_inference_with_text_denoiser_block_scale(self):
pass
pytest.mark.skip("Not supported in AuraFlow.")
@pytest.mark.skip("Not supported in AuraFlow.")
def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(self):
pass
pytest.mark.skip("Not supported in AuraFlow.")
@pytest.mark.skip("Not supported in AuraFlow.")
def test_modify_padding_mode(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in AuraFlow.")
@pytest.mark.skip("Text encoder LoRA is not supported in AuraFlow.")
def test_simple_inference_with_partial_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in AuraFlow.")
@pytest.mark.skip("Text encoder LoRA is not supported in AuraFlow.")
def test_simple_inference_with_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in AuraFlow.")
@pytest.mark.skip("Text encoder LoRA is not supported in AuraFlow.")
def test_simple_inference_with_text_lora_and_scale(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in AuraFlow.")
@pytest.mark.skip("Text encoder LoRA is not supported in AuraFlow.")
def test_simple_inference_with_text_lora_fused(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in AuraFlow.")
@pytest.mark.skip("Text encoder LoRA is not supported in AuraFlow.")
def test_simple_inference_with_text_lora_save_load(self):
pass

View File

@@ -129,11 +129,7 @@ class TestCogVideoXLoRA(PeftLoraLoaderMixinTests):
@pytest.mark.parametrize(
"offload_type, use_stream",
[
("block_level", True),
("leaf_level", False),
("leaf_level", True),
],
[("block_level", True), ("leaf_level", False), ("leaf_level", True)],
)
@require_torch_accelerator
def test_group_offloading_inference_denoiser(self, offload_type, use_stream, tmpdirname):
@@ -141,47 +137,38 @@ class TestCogVideoXLoRA(PeftLoraLoaderMixinTests):
# The reason for this can be found here: https://github.com/huggingface/diffusers/pull/11804#issuecomment-3013325338
super()._test_group_offloading_inference_denoiser(offload_type, use_stream, tmpdirname)
pytest.mark.skip("Not supported in CogVideoX.")
@pytest.mark.skip("Not supported in CogVideoX.")
def test_simple_inference_with_text_denoiser_block_scale(self):
pass
pytest.mark.skip("Not supported in CogVideoX.")
@pytest.mark.skip("Not supported in CogVideoX.")
def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(self):
pass
pytest.mark.skip("Not supported in CogVideoX.")
@pytest.mark.skip("Not supported in CogVideoX.")
def test_modify_padding_mode(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in CogVideoX.")
@pytest.mark.skip("Text encoder LoRA is not supported in CogVideoX.")
def test_simple_inference_with_partial_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in CogVideoX.")
@pytest.mark.skip("Text encoder LoRA is not supported in CogVideoX.")
def test_simple_inference_with_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in CogVideoX.")
@pytest.mark.skip("Text encoder LoRA is not supported in CogVideoX.")
def test_simple_inference_with_text_lora_and_scale(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in CogVideoX.")
@pytest.mark.skip("Text encoder LoRA is not supported in CogVideoX.")
def test_simple_inference_with_text_lora_fused(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in CogVideoX.")
@pytest.mark.skip("Text encoder LoRA is not supported in CogVideoX.")
def test_simple_inference_with_text_lora_save_load(self):
pass
pytest.mark.skip("Not supported in CogVideoX.")
@pytest.mark.skip("Not supported in CogVideoX.")
def test_simple_inference_with_text_denoiser_multi_adapter_block_lora(self):
pass

View File

@@ -142,11 +142,7 @@ class TestCogView4LoRA(PeftLoraLoaderMixinTests):
@pytest.mark.parametrize(
"offload_type, use_stream",
[
("block_level", True),
("leaf_level", False),
("leaf_level", True),
],
[("block_level", True), ("leaf_level", False), ("leaf_level", True)],
)
@require_torch_accelerator
def test_group_offloading_inference_denoiser(self, offload_type, use_stream, tmpdirname):
@@ -154,42 +150,34 @@ class TestCogView4LoRA(PeftLoraLoaderMixinTests):
# The reason for this can be found here: https://github.com/huggingface/diffusers/pull/11804#issuecomment-3013325338
super()._test_group_offloading_inference_denoiser(offload_type, use_stream, tmpdirname)
pytest.mark.skip("Not supported in CogView4.")
@pytest.mark.skip("Not supported in CogView4.")
def test_simple_inference_with_text_denoiser_block_scale(self):
pass
pytest.mark.skip("Not supported in CogView4.")
@pytest.mark.skip("Not supported in CogView4.")
def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(self):
pass
pytest.mark.skip("Not supported in CogView4.")
@pytest.mark.skip("Not supported in CogView4.")
def test_modify_padding_mode(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in CogView4.")
@pytest.mark.skip("Text encoder LoRA is not supported in CogView4.")
def test_simple_inference_with_partial_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in CogView4.")
@pytest.mark.skip("Text encoder LoRA is not supported in CogView4.")
def test_simple_inference_with_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in CogView4.")
@pytest.mark.skip("Text encoder LoRA is not supported in CogView4.")
def test_simple_inference_with_text_lora_and_scale(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in CogView4.")
@pytest.mark.skip("Text encoder LoRA is not supported in CogView4.")
def test_simple_inference_with_text_lora_fused(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in CogView4.")
@pytest.mark.skip("Text encoder LoRA is not supported in CogView4.")
def test_simple_inference_with_text_lora_save_load(self):
pass

View File

@@ -230,23 +230,19 @@ class TestFluxLoRA(PeftLoraLoaderMixinTests):
"LoRA should lead to different results."
)
pytest.mark.skip("Not supported in Flux.")
@pytest.mark.skip("Not supported in Flux.")
def test_simple_inference_with_text_denoiser_block_scale(self):
pass
pytest.mark.skip("Not supported in Flux.")
@pytest.mark.skip("Not supported in Flux.")
def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(self):
pass
pytest.mark.skip("Not supported in Flux.")
@pytest.mark.skip("Not supported in Flux.")
def test_modify_padding_mode(self):
pass
pytest.mark.skip("Not supported in Flux.")
@pytest.mark.skip("Not supported in Flux.")
def test_simple_inference_with_text_denoiser_multi_adapter_block_lora(self):
pass
@@ -725,23 +721,19 @@ class TestFluxControlLoRA(PeftLoraLoaderMixinTests):
assert pipe.transformer.x_embedder.weight.data.shape[1] == in_features * 2
assert pipe.transformer.config.in_channels == in_features * 2
pytest.mark.skip("Not supported in Flux.")
@pytest.mark.skip("Not supported in Flux.")
def test_simple_inference_with_text_denoiser_block_scale(self):
pass
pytest.mark.skip("Not supported in Flux.")
@pytest.mark.skip("Not supported in Flux.")
def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(self):
pass
pytest.mark.skip("Not supported in Flux.")
@pytest.mark.skip("Not supported in Flux.")
def test_modify_padding_mode(self):
pass
pytest.mark.skip("Not supported in Flux.")
@pytest.mark.skip("Not supported in Flux.")
def test_simple_inference_with_text_denoiser_multi_adapter_block_lora(self):
pass

View File

@@ -155,48 +155,39 @@ class TestHunyuanVideoLoRA(PeftLoraLoaderMixinTests):
super().test_simple_inference_with_text_denoiser_lora_unfused(expected_atol=9e-3)
# TODO(aryan): Fix the following test
pytest.mark.skip("This test fails with an error I haven't been able to debug yet.")
@pytest.mark.skip("This test fails with an error I haven't been able to debug yet.")
def test_simple_inference_save_pretrained(self):
pass
pytest.mark.skip("Not supported in HunyuanVideo.")
@pytest.mark.skip("Not supported in HunyuanVideo.")
def test_simple_inference_with_text_denoiser_block_scale(self):
pass
pytest.mark.skip("Not supported in HunyuanVideo.")
@pytest.mark.skip("Not supported in HunyuanVideo.")
def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(self):
pass
pytest.mark.skip("Not supported in HunyuanVideo.")
@pytest.mark.skip("Not supported in HunyuanVideo.")
def test_modify_padding_mode(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in HunyuanVideo.")
@pytest.mark.skip("Text encoder LoRA is not supported in HunyuanVideo.")
def test_simple_inference_with_partial_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in HunyuanVideo.")
@pytest.mark.skip("Text encoder LoRA is not supported in HunyuanVideo.")
def test_simple_inference_with_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in HunyuanVideo.")
@pytest.mark.skip("Text encoder LoRA is not supported in HunyuanVideo.")
def test_simple_inference_with_text_lora_and_scale(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in HunyuanVideo.")
@pytest.mark.skip("Text encoder LoRA is not supported in HunyuanVideo.")
def test_simple_inference_with_text_lora_fused(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in HunyuanVideo.")
@pytest.mark.skip("Text encoder LoRA is not supported in HunyuanVideo.")
def test_simple_inference_with_text_lora_save_load(self):
pass

View File

@@ -114,42 +114,34 @@ class TestLTXVideoLoRA(PeftLoraLoaderMixinTests):
def test_simple_inference_with_text_denoiser_lora_unfused(self):
super().test_simple_inference_with_text_denoiser_lora_unfused(expected_atol=9e-3)
pytest.mark.skip("Not supported in LTXVideo.")
@pytest.mark.skip("Not supported in LTXVideo.")
def test_simple_inference_with_text_denoiser_block_scale(self):
pass
pytest.mark.skip("Not supported in LTXVideo.")
@pytest.mark.skip("Not supported in LTXVideo.")
def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(self):
pass
pytest.mark.skip("Not supported in LTXVideo.")
@pytest.mark.skip("Not supported in LTXVideo.")
def test_modify_padding_mode(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in LTXVideo.")
@pytest.mark.skip("Text encoder LoRA is not supported in LTXVideo.")
def test_simple_inference_with_partial_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in LTXVideo.")
@pytest.mark.skip("Text encoder LoRA is not supported in LTXVideo.")
def test_simple_inference_with_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in LTXVideo.")
@pytest.mark.skip("Text encoder LoRA is not supported in LTXVideo.")
def test_simple_inference_with_text_lora_and_scale(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in LTXVideo.")
@pytest.mark.skip("Text encoder LoRA is not supported in LTXVideo.")
def test_simple_inference_with_text_lora_fused(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in LTXVideo.")
@pytest.mark.skip("Text encoder LoRA is not supported in LTXVideo.")
def test_simple_inference_with_text_lora_save_load(self):
pass

View File

@@ -100,43 +100,35 @@ class TestLumina2LoRA(PeftLoraLoaderMixinTests):
return noise, input_ids, pipeline_inputs
pytest.mark.skip("Not supported in Lumina2.")
@pytest.mark.skip("Not supported in Lumina2.")
def test_simple_inference_with_text_denoiser_block_scale(self):
pass
pytest.mark.skip("Not supported in Lumina2.")
@pytest.mark.skip("Not supported in Lumina2.")
def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(self):
pass
pytest.mark.skip("Not supported in Lumina2.")
@pytest.mark.skip("Not supported in Lumina2.")
def test_modify_padding_mode(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Lumina2.")
@pytest.mark.skip("Text encoder LoRA is not supported in Lumina2.")
def test_simple_inference_with_partial_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Lumina2.")
@pytest.mark.skip("Text encoder LoRA is not supported in Lumina2.")
def test_simple_inference_with_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Lumina2.")
@pytest.mark.skip("Text encoder LoRA is not supported in Lumina2.")
def test_simple_inference_with_text_lora_and_scale(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Lumina2.")
@pytest.mark.skip("Text encoder LoRA is not supported in Lumina2.")
def test_simple_inference_with_text_lora_fused(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Lumina2.")
@pytest.mark.skip("Text encoder LoRA is not supported in Lumina2.")
def test_simple_inference_with_text_lora_save_load(self):
pass

View File

@@ -105,47 +105,38 @@ class TestMochiLoRA(PeftLoraLoaderMixinTests):
def test_simple_inference_with_text_denoiser_lora_unfused(self):
super().test_simple_inference_with_text_denoiser_lora_unfused(expected_atol=9e-3)
pytest.mark.skip("Not supported in Mochi.")
@pytest.mark.skip("Not supported in Mochi.")
def test_simple_inference_with_text_denoiser_block_scale(self):
pass
pytest.mark.skip("Not supported in Mochi.")
@pytest.mark.skip("Not supported in Mochi.")
def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(self):
pass
pytest.mark.skip("Not supported in Mochi.")
@pytest.mark.skip("Not supported in Mochi.")
def test_modify_padding_mode(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Mochi.")
@pytest.mark.skip("Text encoder LoRA is not supported in Mochi.")
def test_simple_inference_with_partial_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Mochi.")
@pytest.mark.skip("Text encoder LoRA is not supported in Mochi.")
def test_simple_inference_with_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Mochi.")
@pytest.mark.skip("Text encoder LoRA is not supported in Mochi.")
def test_simple_inference_with_text_lora_and_scale(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Mochi.")
@pytest.mark.skip("Text encoder LoRA is not supported in Mochi.")
def test_simple_inference_with_text_lora_fused(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Mochi.")
@pytest.mark.skip("Text encoder LoRA is not supported in Mochi.")
def test_simple_inference_with_text_lora_save_load(self):
pass
pytest.mark.skip("Not supported in CogVideoX.")
@pytest.mark.skip("Not supported in CogVideoX.")
def test_simple_inference_with_text_denoiser_multi_adapter_block_lora(self):
pass

View File

@@ -96,42 +96,34 @@ class TestQwenImageLoRA(PeftLoraLoaderMixinTests):
return noise, input_ids, pipeline_inputs
pytest.mark.skip("Not supported in Qwen Image.")
@pytest.mark.skip("Not supported in Qwen Image.")
def test_simple_inference_with_text_denoiser_block_scale(self):
pass
pytest.mark.skip("Not supported in Qwen Image.")
@pytest.mark.skip("Not supported in Qwen Image.")
def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(self):
pass
pytest.mark.skip("Not supported in Qwen Image.")
@pytest.mark.skip("Not supported in Qwen Image.")
def test_modify_padding_mode(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Qwen Image.")
@pytest.mark.skip("Text encoder LoRA is not supported in Qwen Image.")
def test_simple_inference_with_partial_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Qwen Image.")
@pytest.mark.skip("Text encoder LoRA is not supported in Qwen Image.")
def test_simple_inference_with_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Qwen Image.")
@pytest.mark.skip("Text encoder LoRA is not supported in Qwen Image.")
def test_simple_inference_with_text_lora_and_scale(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Qwen Image.")
@pytest.mark.skip("Text encoder LoRA is not supported in Qwen Image.")
def test_simple_inference_with_text_lora_fused(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Qwen Image.")
@pytest.mark.skip("Text encoder LoRA is not supported in Qwen Image.")
def test_simple_inference_with_text_lora_save_load(self):
pass

View File

@@ -105,42 +105,34 @@ class TestSanaLoRA(PeftLoraLoaderMixinTests):
return noise, input_ids, pipeline_inputs
pytest.mark.skip("Not supported in SANA.")
@pytest.mark.skip("Not supported in SANA.")
def test_modify_padding_mode(self):
pass
pytest.mark.skip("Not supported in SANA.")
@pytest.mark.skip("Not supported in SANA.")
def test_simple_inference_with_text_denoiser_block_scale(self):
pass
pytest.mark.skip("Not supported in SANA.")
@pytest.mark.skip("Not supported in SANA.")
def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in SANA.")
@pytest.mark.skip("Text encoder LoRA is not supported in SANA.")
def test_simple_inference_with_partial_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in SANA.")
@pytest.mark.skip("Text encoder LoRA is not supported in SANA.")
def test_simple_inference_with_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in SANA.")
@pytest.mark.skip("Text encoder LoRA is not supported in SANA.")
def test_simple_inference_with_text_lora_and_scale(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in SANA.")
@pytest.mark.skip("Text encoder LoRA is not supported in SANA.")
def test_simple_inference_with_text_lora_fused(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in SANA.")
@pytest.mark.skip("Text encoder LoRA is not supported in SANA.")
def test_simple_inference_with_text_lora_save_load(self):
pass

View File

@@ -113,23 +113,19 @@ class TestSD3LoRA(PeftLoraLoaderMixinTests):
lora_filename = "lora_peft_format.safetensors"
pipe.load_lora_weights(lora_model_id, weight_name=lora_filename)
pytest.mark.skip("Not supported in SD3.")
@pytest.mark.skip("Not supported in SD3.")
def test_simple_inference_with_text_denoiser_block_scale(self):
pass
pytest.mark.skip("Not supported in SD3.")
@pytest.mark.skip("Not supported in SD3.")
def test_simple_inference_with_text_denoiser_multi_adapter_block_lora(self):
pass
pytest.mark.skip("Not supported in SD3.")
@pytest.mark.skip("Not supported in SD3.")
def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(self):
pass
pytest.mark.skip("Not supported in SD3.")
@pytest.mark.skip("Not supported in SD3.")
def test_modify_padding_mode(self):
pass

View File

@@ -110,42 +110,34 @@ class TestWanLoRA(PeftLoraLoaderMixinTests):
def test_simple_inference_with_text_denoiser_lora_unfused(self):
super().test_simple_inference_with_text_denoiser_lora_unfused(expected_atol=9e-3)
pytest.mark.skip("Not supported in Wan.")
@pytest.mark.skip("Not supported in Wan.")
def test_simple_inference_with_text_denoiser_block_scale(self):
pass
pytest.mark.skip("Not supported in Wan.")
@pytest.mark.skip("Not supported in Wan.")
def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(self):
pass
pytest.mark.skip("Not supported in Wan.")
@pytest.mark.skip("Not supported in Wan.")
def test_modify_padding_mode(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Wan.")
@pytest.mark.skip("Text encoder LoRA is not supported in Wan.")
def test_simple_inference_with_partial_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Wan.")
@pytest.mark.skip("Text encoder LoRA is not supported in Wan.")
def test_simple_inference_with_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Wan.")
@pytest.mark.skip("Text encoder LoRA is not supported in Wan.")
def test_simple_inference_with_text_lora_and_scale(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Wan.")
@pytest.mark.skip("Text encoder LoRA is not supported in Wan.")
def test_simple_inference_with_text_lora_fused(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Wan.")
@pytest.mark.skip("Text encoder LoRA is not supported in Wan.")
def test_simple_inference_with_text_lora_save_load(self):
pass

View File

@@ -126,49 +126,38 @@ class TestWanVACELoRA(PeftLoraLoaderMixinTests):
def test_simple_inference_with_text_denoiser_lora_unfused(self):
super().test_simple_inference_with_text_denoiser_lora_unfused(expected_atol=9e-3)
pytest.mark.skip("Not supported in Wan VACE.")
@pytest.mark.skip("Not supported in Wan VACE.")
def test_simple_inference_with_text_denoiser_block_scale(self):
pass
pytest.mark.skip("Not supported in Wan VACE.")
@pytest.mark.skip("Not supported in Wan VACE.")
def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(self):
pass
pytest.mark.skip("Not supported in Wan VACE.")
@pytest.mark.skip("Not supported in Wan VACE.")
def test_modify_padding_mode(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Wan VACE.")
@pytest.mark.skip("Text encoder LoRA is not supported in Wan VACE.")
def test_simple_inference_with_partial_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Wan VACE.")
@pytest.mark.skip("Text encoder LoRA is not supported in Wan VACE.")
def test_simple_inference_with_text_lora(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Wan VACE.")
@pytest.mark.skip("Text encoder LoRA is not supported in Wan VACE.")
def test_simple_inference_with_text_lora_and_scale(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Wan VACE.")
@pytest.mark.skip("Text encoder LoRA is not supported in Wan VACE.")
def test_simple_inference_with_text_lora_fused(self):
pass
pytest.mark.skip("Text encoder LoRA is not supported in Wan VACE.")
@pytest.mark.skip("Text encoder LoRA is not supported in Wan VACE.")
def test_simple_inference_with_text_lora_save_load(self):
pass
def test_layerwise_casting_inference_denoiser(self):
super().test_layerwise_casting_inference_denoiser()
@require_peft_version_greater("0.13.2")
def test_lora_exclude_modules_wanvace(self, base_pipe_output, tmpdirname):
exclude_module_name = "vace_blocks.0.proj_out"

View File

@@ -2029,11 +2029,7 @@ class PeftLoraLoaderMixinTests:
@pytest.mark.parametrize(
"offload_type, use_stream",
[
("block_level", True),
("leaf_level", False),
("leaf_level", True),
],
[("block_level", True), ("leaf_level", False), ("leaf_level", True)],
)
@require_torch_accelerator
def test_group_offloading_inference_denoiser(self, offload_type, use_stream, tmpdirname):