import os import sys import torch try: import intel_extension_for_pytorch as ipex # pylint: disable=import-error, unused-import has_ipex = True except Exception: has_ipex = False from .hijacks import ipex_hijacks torch_version = torch.__version__[:4] if torch_version[-1] not in {"0", "1", "2", "3", "4", "5", "6", "7", "8", "9"}: torch_version = torch_version[:-1] torch_version = torch_version.split(".") torch_version[0], torch_version[1] = int(torch_version[0]), int(torch_version[1]) # pylint: disable=protected-access, missing-function-docstring, line-too-long def ipex_init(): # pylint: disable=too-many-statements try: if hasattr(torch, "cuda") and hasattr(torch.cuda, "is_xpu_hijacked") and torch.cuda.is_xpu_hijacked: return True, "Skipping IPEX hijack" else: try: # force xpu device on torch compile and triton # import inductor utils to get around lazy import from torch._inductor import utils as torch_inductor_utils # pylint: disable=import-error, unused-import # noqa: F401,RUF100 torch._inductor.utils.GPU_TYPES = ["xpu"] torch._inductor.utils.get_gpu_type = lambda *args, **kwargs: "xpu" from triton import backends as triton_backends # pylint: disable=import-error triton_backends.backends["nvidia"].driver.is_active = lambda *args, **kwargs: False except Exception: pass # Replace cuda with xpu: torch.cuda.current_device = torch.xpu.current_device torch.cuda.current_stream = torch.xpu.current_stream torch.cuda.device = torch.xpu.device torch.cuda.device_count = torch.xpu.device_count torch.cuda.device_of = torch.xpu.device_of torch.cuda.get_device_name = torch.xpu.get_device_name torch.cuda.get_device_properties = torch.xpu.get_device_properties torch.cuda.init = torch.xpu.init torch.cuda.is_available = torch.xpu.is_available torch.cuda.is_initialized = torch.xpu.is_initialized torch.cuda.stream = torch.xpu.stream torch.cuda.Event = torch.xpu.Event torch.cuda.Stream = torch.xpu.Stream torch.cuda.Optional = torch.xpu.Optional torch.cuda.streams = torch.xpu.streams torch.cuda.Any = torch.xpu.Any torch.cuda.default_generators = torch.xpu.default_generators torch.cuda.set_stream = torch.xpu.set_stream torch.cuda.torch = torch.xpu.torch torch.cuda.Union = torch.xpu.Union torch.cuda.StreamContext = torch.xpu.StreamContext torch.cuda.random = torch.xpu.random torch.cuda._get_device_index = torch.xpu._get_device_index torch.cuda._lazy_init = torch.xpu._lazy_init torch.cuda._lazy_call = torch.xpu._lazy_call torch.cuda._device = torch.xpu._device torch.cuda._device_t = torch.xpu._device_t torch.cuda.is_current_stream_capturing = lambda: False torch.cuda.__annotations__ = torch.xpu.__annotations__ torch.cuda.__builtins__ = torch.xpu.__builtins__ torch.cuda.__name__ = torch.xpu.__name__ torch.cuda.__spec__ = torch.xpu.__spec__ torch.cuda.__file__ = torch.xpu.__file__ torch.cuda.__path__ = torch.xpu.__path__ torch.cuda.__doc__ = torch.xpu.__doc__ torch.cuda.__package__ = getattr(torch.xpu, "__package__", None) torch.cuda.__cached__ = getattr(torch.xpu, "__cached__", None) torch.cuda.__loader__ = getattr(torch.xpu, "__loader__", None) torch.Tensor.cuda = torch.Tensor.xpu torch.Tensor.is_cuda = torch.Tensor.is_xpu torch.nn.Module.cuda = torch.nn.Module.xpu if torch_version[0] < 2 or (torch_version[0] == 2 and torch_version[1] < 3): torch.cuda.threading = torch.xpu.lazy_init.threading torch.cuda.traceback = torch.xpu.lazy_init.traceback torch.cuda._initialization_lock = torch.xpu.lazy_init._initialization_lock torch.cuda._initialized = torch.xpu.lazy_init._initialized torch.cuda._is_in_bad_fork = torch.xpu.lazy_init._is_in_bad_fork torch.cuda._lazy_seed_tracker = torch.xpu.lazy_init._lazy_seed_tracker torch.cuda._queued_calls = torch.xpu.lazy_init._queued_calls torch.cuda._tls = torch.xpu.lazy_init._tls torch.cuda._lazy_new = torch.xpu._lazy_new torch.cuda.FloatTensor = torch.xpu.FloatTensor torch.cuda.FloatStorage = torch.xpu.FloatStorage torch.cuda.BFloat16Tensor = torch.xpu.BFloat16Tensor torch.cuda.BFloat16Storage = torch.xpu.BFloat16Storage torch.cuda.HalfTensor = torch.xpu.HalfTensor torch.cuda.HalfStorage = torch.xpu.HalfStorage torch.cuda.ByteTensor = torch.xpu.ByteTensor torch.cuda.ByteStorage = torch.xpu.ByteStorage torch.cuda.DoubleTensor = torch.xpu.DoubleTensor torch.cuda.DoubleStorage = torch.xpu.DoubleStorage torch.cuda.ShortTensor = torch.xpu.ShortTensor torch.cuda.ShortStorage = torch.xpu.ShortStorage torch.cuda.LongTensor = torch.xpu.LongTensor torch.cuda.LongStorage = torch.xpu.LongStorage torch.cuda.IntTensor = torch.xpu.IntTensor torch.cuda.IntStorage = torch.xpu.IntStorage torch.cuda.CharTensor = torch.xpu.CharTensor torch.cuda.CharStorage = torch.xpu.CharStorage torch.cuda.BoolTensor = torch.xpu.BoolTensor torch.cuda.BoolStorage = torch.xpu.BoolStorage torch.cuda.ComplexFloatStorage = torch.xpu.ComplexFloatStorage torch.cuda.ComplexDoubleStorage = torch.xpu.ComplexDoubleStorage if has_ipex: torch._C._cuda_getCurrentRawStream = ipex._C._getCurrentRawStream else: torch.cuda.threading = torch.xpu.threading torch.cuda.traceback = torch.xpu.traceback torch.cuda._initialization_lock = torch.xpu._initialization_lock torch.cuda._initialized = torch.xpu._initialized torch.cuda._is_in_bad_fork = torch.xpu._is_in_bad_fork torch.cuda._lazy_seed_tracker = torch.xpu._lazy_seed_tracker torch.cuda._queued_calls = torch.xpu._queued_calls torch.cuda._tls = torch.xpu._tls torch._C._cuda_getCurrentRawStream = torch._C._xpu_getCurrentRawStream if torch_version[0] < 2 or (torch_version[0] == 2 and torch_version[1] < 5): torch.cuda.os = torch.xpu.os torch.cuda.Device = torch.xpu.Device torch.cuda.warnings = torch.xpu.warnings torch.cuda.classproperty = torch.xpu.classproperty torch.UntypedStorage.cuda = torch.UntypedStorage.xpu if torch_version[0] < 2 or (torch_version[0] == 2 and torch_version[1] < 7): torch.cuda.Tuple = torch.xpu.Tuple torch.cuda.List = torch.xpu.List if torch_version[0] < 2 or (torch_version[0] == 2 and torch_version[1] < 8): if has_ipex: torch.cuda.memory_summary = torch.xpu.memory_summary torch.cuda.memory_snapshot = torch.xpu.memory_snapshot # Memory: if "linux" in sys.platform and "WSL2" in os.popen("uname -a").read(): torch.xpu.empty_cache = lambda: None torch.cuda.empty_cache = torch.xpu.empty_cache torch.cuda.memory = torch.xpu.memory torch.cuda.memory_stats = torch.xpu.memory_stats torch.cuda.memory_allocated = torch.xpu.memory_allocated torch.cuda.max_memory_allocated = torch.xpu.max_memory_allocated torch.cuda.memory_reserved = torch.xpu.memory_reserved torch.cuda.memory_cached = torch.xpu.memory_reserved torch.cuda.max_memory_reserved = torch.xpu.max_memory_reserved torch.cuda.max_memory_cached = torch.xpu.max_memory_reserved torch.cuda.reset_peak_memory_stats = torch.xpu.reset_peak_memory_stats torch.cuda.reset_max_memory_cached = torch.xpu.reset_peak_memory_stats torch.cuda.reset_max_memory_allocated = torch.xpu.reset_peak_memory_stats torch.cuda.memory_stats_as_nested_dict = torch.xpu.memory_stats_as_nested_dict torch.cuda.reset_accumulated_memory_stats = torch.xpu.reset_accumulated_memory_stats # RNG: torch.cuda.get_rng_state = torch.xpu.get_rng_state torch.cuda.get_rng_state_all = torch.xpu.get_rng_state_all torch.cuda.set_rng_state = torch.xpu.set_rng_state torch.cuda.set_rng_state_all = torch.xpu.set_rng_state_all torch.cuda.manual_seed = torch.xpu.manual_seed torch.cuda.manual_seed_all = torch.xpu.manual_seed_all torch.cuda.seed = torch.xpu.seed torch.cuda.seed_all = torch.xpu.seed_all torch.cuda.initial_seed = torch.xpu.initial_seed # Fix functions with ipex: # torch.xpu.mem_get_info always returns the total memory as free memory torch.has_cuda = True torch.version.cuda = "12.1" torch.backends.cuda.is_built = lambda *args, **kwargs: True torch._utils._get_available_device_type = lambda: "xpu" torch.xpu.mem_get_info = lambda device=None: [(torch.xpu.get_device_properties(device).total_memory - torch.xpu.memory_reserved(device)), torch.xpu.get_device_properties(device).total_memory] torch.cuda.mem_get_info = torch.xpu.mem_get_info torch.cuda.has_half = True torch.cuda.is_bf16_supported = getattr(torch.xpu, "is_bf16_supported", lambda *args, **kwargs: True) torch.cuda.is_fp16_supported = lambda *args, **kwargs: True torch.cuda.get_arch_list = getattr(torch.xpu, "get_arch_list", lambda: ["pvc", "dg2", "ats-m150"]) torch.cuda.get_device_capability = lambda *args, **kwargs: (12,1) torch.cuda.ipc_collect = lambda *args, **kwargs: None torch.cuda.utilization = lambda *args, **kwargs: 0 device_supports_fp64 = ipex_hijacks() try: from .diffusers import ipex_diffusers ipex_diffusers(device_supports_fp64=device_supports_fp64) except Exception: # pylint: disable=broad-exception-caught pass torch.cuda.is_xpu_hijacked = True except Exception as e: return False, e return True, None