1
0
mirror of https://github.com/huggingface/diffusers.git synced 2026-01-29 07:22:12 +03:00

remove debug print

This commit is contained in:
Aryan
2025-04-15 08:26:43 +02:00
parent 8d31c699a5
commit 1c1d1d52e0

View File

@@ -80,17 +80,14 @@ class AttentionProcessorSkipHook(ModelHook):
def new_forward(self, module: torch.nn.Module, *args, **kwargs):
if self.skip_attention_scores:
print("Skipping attention scores")
with AttentionScoreSkipFunctionMode():
return self.fn_ref.original_forward(*args, **kwargs)
else:
print("Skipping attention processor output")
return self.skip_processor_output_fn(module, *args, **kwargs)
class FeedForwardSkipHook(ModelHook):
def new_forward(self, module: torch.nn.Module, *args, **kwargs):
print("Skipping feed-forward block")
output = kwargs.get("hidden_states", None)
if output is None:
output = kwargs.get("x", None)
@@ -105,7 +102,6 @@ class TransformerBlockSkipHook(ModelHook):
return module
def new_forward(self, module: torch.nn.Module, *args, **kwargs):
print("Skipping transformer block")
return self._metadata.skip_block_output_fn(module, *args, **kwargs)