From 9d3ff0794d7bada13c01ca0e517377d776cf48ff Mon Sep 17 00:00:00 2001 From: Patrick von Platen Date: Tue, 30 May 2023 18:59:07 +0100 Subject: [PATCH] fix tests (#3614) --- examples/community/{ mixture.py => mixture.py} | 0 tests/pipelines/stable_diffusion/test_stable_diffusion.py | 5 ++--- 2 files changed, 2 insertions(+), 3 deletions(-) rename examples/community/{ mixture.py => mixture.py} (100%) diff --git a/examples/community/ mixture.py b/examples/community/mixture.py similarity index 100% rename from examples/community/ mixture.py rename to examples/community/mixture.py diff --git a/tests/pipelines/stable_diffusion/test_stable_diffusion.py b/tests/pipelines/stable_diffusion/test_stable_diffusion.py index b5d968e2a3..dafd003215 100644 --- a/tests/pipelines/stable_diffusion/test_stable_diffusion.py +++ b/tests/pipelines/stable_diffusion/test_stable_diffusion.py @@ -41,7 +41,6 @@ from diffusers.models.attention_processor import AttnProcessor, LoRAXFormersAttn from diffusers.utils import load_numpy, nightly, slow, torch_device from diffusers.utils.testing_utils import ( CaptureLogger, - disable_full_determinism, enable_full_determinism, require_torch_2, require_torch_gpu, @@ -370,7 +369,7 @@ class StableDiffusionPipelineFastTests(PipelineLatentTesterMixin, PipelineTester @unittest.skipIf(not torch.cuda.is_available(), reason="xformers requires cuda") def test_stable_diffusion_attn_processors(self): - disable_full_determinism() + # disable_full_determinism() device = "cuda" # ensure determinism for the device-dependent torch.Generator components = self.get_dummy_components() sd_pipe = StableDiffusionPipeline(**components) @@ -416,7 +415,7 @@ class StableDiffusionPipelineFastTests(PipelineLatentTesterMixin, PipelineTester image = sd_pipe(**inputs).images assert image.shape == (1, 64, 64, 3) - enable_full_determinism() + # enable_full_determinism() def test_stable_diffusion_no_safety_checker(self): pipe = StableDiffusionPipeline.from_pretrained(