1
0
mirror of https://github.com/huggingface/diffusers.git synced 2026-01-29 07:22:12 +03:00
This commit is contained in:
Pedro Cuenca
2023-05-22 16:47:20 +02:00
parent 260307cab2
commit a090b7e211
6 changed files with 34 additions and 16 deletions

View File

@@ -15,6 +15,7 @@
import inspect
import tempfile
import traceback
import unittest
import unittest.mock as mock
from typing import Dict, List, Tuple
@@ -22,7 +23,6 @@ from typing import Dict, List, Tuple
import numpy as np
import requests_mock
import torch
import traceback
from requests.exceptions import HTTPError
from diffusers.models import UNet2DConditionModel
@@ -259,7 +259,7 @@ class ModelTesterMixin:
max_diff = (image - new_image).abs().sum().item()
self.assertLessEqual(max_diff, 5e-5, "Models give different forward passes")
@require_torch_2
def test_from_save_pretrained_dynamo(self):
init_dict, _ = self.prepare_init_args_and_inputs_for_common()

View File

@@ -20,7 +20,6 @@ import unittest
import numpy as np
import torch
from packaging import version
from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer
from diffusers import (
@@ -33,7 +32,12 @@ from diffusers import (
from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet import MultiControlNetModel
from diffusers.utils import load_image, load_numpy, randn_tensor, slow, torch_device
from diffusers.utils.import_utils import is_xformers_available
from diffusers.utils.testing_utils import enable_full_determinism, require_torch_2, require_torch_gpu, run_test_in_subprocess
from diffusers.utils.testing_utils import (
enable_full_determinism,
require_torch_2,
require_torch_gpu,
run_test_in_subprocess,
)
from ..pipeline_params import (
TEXT_TO_IMAGE_BATCH_PARAMS,

View File

@@ -15,17 +15,13 @@
import gc
import os
import signal
import subprocess
import sys
import tempfile
import time
import traceback
import unittest
import numpy as np
import torch
import traceback
from huggingface_hub import hf_hub_download
from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer
@@ -43,7 +39,13 @@ from diffusers import (
)
from diffusers.models.attention_processor import AttnProcessor
from diffusers.utils import load_numpy, nightly, slow, torch_device
from diffusers.utils.testing_utils import CaptureLogger, enable_full_determinism, require_torch_2, require_torch_gpu, run_test_in_subprocess
from diffusers.utils.testing_utils import (
CaptureLogger,
enable_full_determinism,
require_torch_2,
require_torch_gpu,
run_test_in_subprocess,
)
from ...models.test_models_unet_2d_condition import create_lora_layers
from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS
@@ -957,6 +959,7 @@ class StableDiffusionPipelineSlowTests(unittest.TestCase):
inputs["seed"] = seed
run_test_in_subprocess(test_case=self, target_func=_test_stable_diffusion_compile, inputs=inputs)
@slow
@require_torch_gpu
class StableDiffusionPipelineCkptTests(unittest.TestCase):

View File

@@ -15,11 +15,11 @@
import gc
import random
import traceback
import unittest
import numpy as np
import torch
import traceback
from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer
from diffusers import (
@@ -34,7 +34,13 @@ from diffusers import (
)
from diffusers.image_processor import VaeImageProcessor
from diffusers.utils import floats_tensor, load_image, load_numpy, nightly, slow, torch_device
from diffusers.utils.testing_utils import enable_full_determinism, require_torch_2, require_torch_gpu, skip_mps, run_test_in_subprocess
from diffusers.utils.testing_utils import (
enable_full_determinism,
require_torch_2,
require_torch_gpu,
run_test_in_subprocess,
skip_mps,
)
from ..pipeline_params import (
IMAGE_TO_IMAGE_IMAGE_PARAMS,

View File

@@ -15,11 +15,11 @@
import gc
import random
import traceback
import unittest
import numpy as np
import torch
import traceback
from PIL import Image
from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer
@@ -33,7 +33,12 @@ from diffusers import (
)
from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_inpaint import prepare_mask_and_masked_image
from diffusers.utils import floats_tensor, load_image, load_numpy, nightly, slow, torch_device
from diffusers.utils.testing_utils import enable_full_determinism, require_torch_2, require_torch_gpu, run_test_in_subprocess
from diffusers.utils.testing_utils import (
enable_full_determinism,
require_torch_2,
require_torch_gpu,
run_test_in_subprocess,
)
from ...models.test_models_unet_2d_condition import create_lora_layers
from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS
@@ -42,6 +47,7 @@ from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMix
enable_full_determinism()
# Will be run via run_test_in_subprocess
def _test_inpaint_compile(in_queue, out_queue, timeout):
error = None
@@ -76,7 +82,6 @@ def _test_inpaint_compile(in_queue, out_queue, timeout):
out_queue.join()
class StableDiffusionInpaintPipelineFastTests(PipelineLatentTesterMixin, PipelineTesterMixin, unittest.TestCase):
pipeline_class = StableDiffusionInpaintPipeline
params = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS

View File

@@ -20,6 +20,7 @@ import random
import shutil
import sys
import tempfile
import traceback
import unittest
import unittest.mock as mock
@@ -28,7 +29,6 @@ import PIL
import requests_mock
import safetensors.torch
import torch
import traceback
from parameterized import parameterized
from PIL import Image
from requests.exceptions import HTTPError