1
0
mirror of https://github.com/huggingface/diffusers.git synced 2026-01-29 07:22:12 +03:00

push local changes, fix docs

This commit is contained in:
BuildTools
2025-06-13 07:46:29 -06:00
committed by DN6
parent 41751a3ec0
commit fd3e94450a
3 changed files with 4 additions and 10 deletions

View File

@@ -25,6 +25,7 @@ Original model checkpoints for Chroma can be found [here](https://huggingface.co
Chroma can use all the same optimizations as Flux.
</Tip>
## Inference (Single File)

View File

@@ -125,7 +125,7 @@ class ChromaTransformerTests(ModelTesterMixin, unittest.TestCase):
"num_layers": 1,
"num_single_layers": 1,
"attention_head_dim": 16,
"num_attention_heads": 192,
"num_attention_heads": 2,
"joint_attention_dim": 32,
"axes_dims_rope": [4, 4, 8],
"approximator_num_channels": 8,

View File

@@ -4,15 +4,8 @@ import numpy as np
import torch
from transformers import AutoTokenizer, T5EncoderModel
from diffusers import (
AutoencoderKL,
ChromaPipeline,
ChromaTransformer2DModel,
FlowMatchEulerDiscreteScheduler,
)
from diffusers.utils.testing_utils import (
torch_device,
)
from diffusers import AutoencoderKL, ChromaPipeline, ChromaTransformer2DModel, FlowMatchEulerDiscreteScheduler
from diffusers.utils.testing_utils import torch_device
from ..test_pipelines_common import (
FluxIPAdapterTesterMixin,