From e6037e8275da98b15335a1de67b06cb79363eaf4 Mon Sep 17 00:00:00 2001 From: Sayak Paul Date: Wed, 29 Jan 2025 21:12:57 +0530 Subject: [PATCH] [tests] update llamatokenizer in hunyuanvideo tests (#10681) update llamatokenizer in hunyuanvideo tests --- tests/pipelines/hunyuan_video/test_hunyuan_video.py | 6 ++---- 1 file changed, 2 insertions(+), 4 deletions(-) diff --git a/tests/pipelines/hunyuan_video/test_hunyuan_video.py b/tests/pipelines/hunyuan_video/test_hunyuan_video.py index 1ecfee666f..ba7ec43ec9 100644 --- a/tests/pipelines/hunyuan_video/test_hunyuan_video.py +++ b/tests/pipelines/hunyuan_video/test_hunyuan_video.py @@ -132,7 +132,7 @@ class HunyuanVideoPipelineFastTests(PipelineTesterMixin, PyramidAttentionBroadca torch.manual_seed(0) text_encoder = LlamaModel(llama_text_encoder_config) - tokenizer = LlamaTokenizer.from_pretrained("hf-internal-testing/tiny-random-LlamaForCausalLM") + tokenizer = LlamaTokenizer.from_pretrained("finetrainers/dummy-hunyaunvideo", subfolder="tokenizer") torch.manual_seed(0) text_encoder_2 = CLIPTextModel(clip_text_encoder_config) @@ -155,10 +155,8 @@ class HunyuanVideoPipelineFastTests(PipelineTesterMixin, PyramidAttentionBroadca else: generator = torch.Generator(device=device).manual_seed(seed) - # Cannot test with dummy prompt because tokenizers are not configured correctly. - # TODO(aryan): create dummy tokenizers and using from hub inputs = { - "prompt": "", + "prompt": "dance monkey", "prompt_template": { "template": "{}", "crop_start": 0,