1
0
mirror of https://github.com/huggingface/diffusers.git synced 2026-01-29 07:22:12 +03:00
This commit is contained in:
sayakpaul
2025-06-24 16:25:20 +05:30
parent ea6c364485
commit f82de3339e
4 changed files with 5 additions and 2 deletions

View File

@@ -21,6 +21,7 @@ from typing import Dict, Optional, Union
from .bitsandbytes import BnB4BitDiffusersQuantizer, BnB8BitDiffusersQuantizer
from .gguf import GGUFQuantizer
from .higgs import HiggsQuantizer
from .quantization_config import (
BitsAndBytesConfig,
GGUFQuantizationConfig,
@@ -40,7 +41,7 @@ AUTO_QUANTIZER_MAPPING = {
"gguf": GGUFQuantizer,
"quanto": QuantoQuantizer,
"torchao": TorchAoHfQuantizer,
"higgs": 1,
"higgs": HiggsQuantizer,
}
AUTO_QUANTIZATION_CONFIG_MAPPING = {

View File

@@ -0,0 +1 @@
from .higgs_quantizer import HiggsQuantizer

View File

@@ -35,7 +35,7 @@ if is_torch_available():
logger = logging.get_logger(__name__)
class HiggsHfQuantizer(DiffusersQuantizer):
class HiggsQuantizer(DiffusersQuantizer):
"""
Quantizer of the HIGGS method. Enables the loading of prequantized models and in-flight quantization of
full-precision models.

View File

@@ -46,6 +46,7 @@ class QuantizationMethod(str, Enum):
GGUF = "gguf"
TORCHAO = "torchao"
QUANTO = "quanto"
HIGGS = "higgs"
if is_torchao_available():