diff --git a/scripts/convert_shap_e_to_diffusers.py b/scripts/convert_shap_e_to_diffusers.py index b903b4ee8a..043cb282da 100644 --- a/scripts/convert_shap_e_to_diffusers.py +++ b/scripts/convert_shap_e_to_diffusers.py @@ -5,7 +5,7 @@ import torch from accelerate import load_checkpoint_and_dispatch from diffusers.models.transformers.prior_transformer import PriorTransformer -from diffusers.pipelines.shap_e import ShapERenderer +from diffusers.pipelines.deprecated.shap_e import ShapERenderer """ diff --git a/src/diffusers/pipelines/__init__.py b/src/diffusers/pipelines/__init__.py index 30a9d98bf0..5f0cd391a7 100644 --- a/src/diffusers/pipelines/__init__.py +++ b/src/diffusers/pipelines/__init__.py @@ -107,6 +107,7 @@ else: "AltDiffusionPipeline", "AltDiffusionImg2ImgPipeline", "CycleDiffusionPipeline", + "StableDiffusionLDM3DPipeline", "StableDiffusionInpaintPipelineLegacy", "StableDiffusionPix2PixZeroPipeline", "StableDiffusionParadigmsPipeline", @@ -327,7 +328,6 @@ else: "StableDiffusionUpscalePipeline", "StableUnCLIPImg2ImgPipeline", "StableUnCLIPPipeline", - "StableDiffusionLDM3DPipeline", ] ) _import_structure["aura_flow"] = ["AuraFlowPipeline"] @@ -555,15 +555,27 @@ if TYPE_CHECKING or DIFFUSERS_SLOW_IMPORT: AltDiffusionImg2ImgPipeline, AltDiffusionPipeline, CycleDiffusionPipeline, + PaintByExamplePipeline, + StableDiffusionAttendAndExcitePipeline, + StableDiffusionControlNetXSPipeline, + StableDiffusionDiffEditPipeline, + StableDiffusionGLIGENPipeline, + StableDiffusionGLIGENTextImagePipeline, StableDiffusionInpaintPipelineLegacy, StableDiffusionModelEditingPipeline, + StableDiffusionPanoramaPipeline, StableDiffusionParadigmsPipeline, StableDiffusionPix2PixZeroPipeline, + StableDiffusionXLControlNetXSPipeline, VersatileDiffusionDualGuidedPipeline, VersatileDiffusionImageVariationPipeline, VersatileDiffusionPipeline, VersatileDiffusionTextToImagePipeline, + VideoToVideoSDPipeline, VQDiffusionPipeline, + WuerstchenCombinedPipeline, + WuerstchenDecoderPipeline, + WuerstchenPriorPipeline, ) from .easyanimate import ( EasyAnimateControlPipeline, @@ -659,7 +671,6 @@ if TYPE_CHECKING or DIFFUSERS_SLOW_IMPORT: StableDiffusionXLPAGInpaintPipeline, StableDiffusionXLPAGPipeline, ) - from .paint_by_example import PaintByExamplePipeline from .pia import PIAPipeline from .pixart_alpha import PixArtAlphaPipeline, PixArtSigmaPipeline from .sana import SanaControlNetPipeline, SanaPipeline, SanaSprintPipeline @@ -689,13 +700,6 @@ if TYPE_CHECKING or DIFFUSERS_SLOW_IMPORT: StableDiffusion3InpaintPipeline, StableDiffusion3Pipeline, ) - from .stable_diffusion_attend_and_excite import StableDiffusionAttendAndExcitePipeline - from .stable_diffusion_diffedit import StableDiffusionDiffEditPipeline - from .stable_diffusion_gligen import StableDiffusionGLIGENPipeline, StableDiffusionGLIGENTextImagePipeline - from .stable_diffusion_ldm3d import StableDiffusionLDM3DPipeline - from .stable_diffusion_panorama import StableDiffusionPanoramaPipeline - from .stable_diffusion_safe import StableDiffusionPipelineSafe - from .stable_diffusion_sag import StableDiffusionSAGPipeline from .stable_diffusion_xl import ( StableDiffusionXLImg2ImgPipeline, StableDiffusionXLInpaintPipeline, @@ -707,25 +711,7 @@ if TYPE_CHECKING or DIFFUSERS_SLOW_IMPORT: StableDiffusionAdapterPipeline, StableDiffusionXLAdapterPipeline, ) - from .text_to_video_synthesis import ( - TextToVideoSDPipeline, - TextToVideoZeroPipeline, - TextToVideoZeroSDXLPipeline, - VideoToVideoSDPipeline, - ) - from .unclip import UnCLIPImageVariationPipeline, UnCLIPPipeline - from .unidiffuser import ( - ImageTextPipelineOutput, - UniDiffuserModel, - UniDiffuserPipeline, - UniDiffuserTextDecoder, - ) from .wan import WanImageToVideoPipeline, WanPipeline, WanVideoToVideoPipeline - from .wuerstchen import ( - WuerstchenCombinedPipeline, - WuerstchenDecoderPipeline, - WuerstchenPriorPipeline, - ) try: if not is_onnx_available(): @@ -756,7 +742,7 @@ if TYPE_CHECKING or DIFFUSERS_SLOW_IMPORT: except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_transformers_and_k_diffusion_objects import * else: - from .stable_diffusion_k_diffusion import ( + from .deprecated import ( StableDiffusionKDiffusionPipeline, StableDiffusionXLKDiffusionPipeline, ) diff --git a/src/diffusers/pipelines/deprecated/__init__.py b/src/diffusers/pipelines/deprecated/__init__.py index 9eb5fffee7..a6a8a20ec2 100644 --- a/src/diffusers/pipelines/deprecated/__init__.py +++ b/src/diffusers/pipelines/deprecated/__init__.py @@ -113,12 +113,16 @@ else: "TextToVideoSDPipeline", "TextToVideoZeroPipeline", "TextToVideoZeroSDXLPipeline", + "VideoToVideoSDPipeline", ] _import_structure["unclip"] = [ "UnCLIPPipeline", "UnCLIPImageVariationPipeline", ] _import_structure["unidiffuser"] = [ + "UniDiffuserTextDecoder", + "UniDiffuserModel", + "UTransformer2DModel", "UniDiffuserPipeline", ] _import_structure["versatile_diffusion"] = [ diff --git a/tests/pipelines/controlnet/test_controlnet_blip_diffusion.py b/tests/pipelines/controlnet/test_controlnet_blip_diffusion.py index a5768cb51f..07226f3671 100644 --- a/tests/pipelines/controlnet/test_controlnet_blip_diffusion.py +++ b/tests/pipelines/controlnet/test_controlnet_blip_diffusion.py @@ -28,10 +28,10 @@ from diffusers import ( PNDMScheduler, UNet2DConditionModel, ) +from diffusers.pipelines.deprecated.blip_diffusion.blip_image_processing import BlipImageProcessor +from diffusers.pipelines.deprecated.blip_diffusion.modeling_blip2 import Blip2QFormerModel +from diffusers.pipelines.deprecated.blip_diffusion.modeling_ctx_clip import ContextCLIPTextModel from diffusers.utils.testing_utils import enable_full_determinism, torch_device -from src.diffusers.pipelines.blip_diffusion.blip_image_processing import BlipImageProcessor -from src.diffusers.pipelines.blip_diffusion.modeling_blip2 import Blip2QFormerModel -from src.diffusers.pipelines.blip_diffusion.modeling_ctx_clip import ContextCLIPTextModel from ..test_pipelines_common import PipelineTesterMixin diff --git a/tests/pipelines/paint_by_example/test_paint_by_example.py b/tests/pipelines/paint_by_example/test_paint_by_example.py index 6b668de276..8de89c2195 100644 --- a/tests/pipelines/paint_by_example/test_paint_by_example.py +++ b/tests/pipelines/paint_by_example/test_paint_by_example.py @@ -23,7 +23,7 @@ from PIL import Image from transformers import CLIPImageProcessor, CLIPVisionConfig from diffusers import AutoencoderKL, PaintByExamplePipeline, PNDMScheduler, UNet2DConditionModel -from diffusers.pipelines.paint_by_example import PaintByExampleImageEncoder +from diffusers.pipelines.deprecated.paint_by_example import PaintByExampleImageEncoder from diffusers.utils.testing_utils import ( enable_full_determinism, floats_tensor, diff --git a/tests/pipelines/semantic_stable_diffusion/test_semantic_diffusion.py b/tests/pipelines/semantic_stable_diffusion/test_semantic_diffusion.py index 6cd431f02d..77414bb9a4 100644 --- a/tests/pipelines/semantic_stable_diffusion/test_semantic_diffusion.py +++ b/tests/pipelines/semantic_stable_diffusion/test_semantic_diffusion.py @@ -23,7 +23,9 @@ import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNet2DConditionModel -from diffusers.pipelines.semantic_stable_diffusion import SemanticStableDiffusionPipeline as StableDiffusionPipeline +from diffusers.pipelines.deprecated.semantic_stable_diffusion import ( + SemanticStableDiffusionPipeline as StableDiffusionPipeline, +) from diffusers.utils.testing_utils import ( enable_full_determinism, floats_tensor, diff --git a/tests/pipelines/shap_e/test_shap_e.py b/tests/pipelines/shap_e/test_shap_e.py index 6cf643fe47..64195534c8 100644 --- a/tests/pipelines/shap_e/test_shap_e.py +++ b/tests/pipelines/shap_e/test_shap_e.py @@ -20,7 +20,7 @@ import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline -from diffusers.pipelines.shap_e import ShapERenderer +from diffusers.pipelines.deprecated.shap_e import ShapERenderer from diffusers.utils.testing_utils import load_numpy, nightly, require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference diff --git a/tests/pipelines/shap_e/test_shap_e_img2img.py b/tests/pipelines/shap_e/test_shap_e_img2img.py index 72eee3e35e..f3a256b26b 100644 --- a/tests/pipelines/shap_e/test_shap_e_img2img.py +++ b/tests/pipelines/shap_e/test_shap_e_img2img.py @@ -21,7 +21,7 @@ import torch from transformers import CLIPImageProcessor, CLIPVisionConfig, CLIPVisionModel from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEImg2ImgPipeline -from diffusers.pipelines.shap_e import ShapERenderer +from diffusers.pipelines.deprecated.shap_e import ShapERenderer from diffusers.utils.testing_utils import ( floats_tensor, load_image, diff --git a/tests/pipelines/stable_diffusion_safe/test_safe_diffusion.py b/tests/pipelines/stable_diffusion_safe/test_safe_diffusion.py index 269677c083..20630e7a29 100644 --- a/tests/pipelines/stable_diffusion_safe/test_safe_diffusion.py +++ b/tests/pipelines/stable_diffusion_safe/test_safe_diffusion.py @@ -23,7 +23,7 @@ import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNet2DConditionModel -from diffusers.pipelines.stable_diffusion_safe import StableDiffusionPipelineSafe as StableDiffusionPipeline +from diffusers.pipelines.deprecated.stable_diffusion_safe import StableDiffusionPipelineSafe as StableDiffusionPipeline from diffusers.utils.testing_utils import floats_tensor, nightly, require_accelerator, require_torch_gpu, torch_device