Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -216,11 +216,11 @@ def load_models():
|
|
216 |
|
217 |
# 加载模型
|
218 |
try:
|
219 |
-
text_encoder = LlamaModel.from_pretrained("hunyuanvideo-community/HunyuanVideo
|
220 |
-
text_encoder_2 = CLIPTextModel.from_pretrained("hunyuanvideo-community/HunyuanVideo
|
221 |
-
tokenizer = LlamaTokenizerFast.from_pretrained("hunyuanvideo-community/HunyuanVideo
|
222 |
-
tokenizer_2 = CLIPTokenizer.from_pretrained("hunyuanvideo-community/HunyuanVideo
|
223 |
-
vae = AutoencoderKLHunyuanVideo.from_pretrained("hunyuanvideo-community/HunyuanVideo
|
224 |
|
225 |
feature_extractor = SiglipImageProcessor.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='feature_extractor')
|
226 |
image_encoder = SiglipVisionModel.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='image_encoder', torch_dtype=dtype).to(model_device)
|
@@ -237,11 +237,11 @@ def load_models():
|
|
237 |
transformer_dtype = torch.float32
|
238 |
cpu_fallback_mode = True
|
239 |
|
240 |
-
text_encoder = LlamaModel.from_pretrained("hunyuanvideo-community/HunyuanVideo
|
241 |
-
text_encoder_2 = CLIPTextModel.from_pretrained("hunyuanvideo-community/HunyuanVideo
|
242 |
-
tokenizer = LlamaTokenizerFast.from_pretrained("hunyuanvideo-community/HunyuanVideo
|
243 |
-
tokenizer_2 = CLIPTokenizer.from_pretrained("hunyuanvideo-community/HunyuanVideo
|
244 |
-
vae = AutoencoderKLHunyuanVideo.from_pretrained("hunyuanvideo-community/HunyuanVideo
|
245 |
|
246 |
feature_extractor = SiglipImageProcessor.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='feature_extractor')
|
247 |
image_encoder = SiglipVisionModel.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='image_encoder', torch_dtype=dtype).to('cpu')
|
@@ -1101,8 +1101,8 @@ def end_process():
|
|
1101 |
|
1102 |
|
1103 |
quick_prompts = [
|
1104 |
-
'The
|
1105 |
-
|
1106 |
]
|
1107 |
quick_prompts = [[x] for x in quick_prompts]
|
1108 |
|
@@ -1545,7 +1545,7 @@ with block:
|
|
1545 |
label="GPU Memory (GB) / GPU推理保留内存(GB)",
|
1546 |
minimum=6,
|
1547 |
maximum=128,
|
1548 |
-
value=
|
1549 |
step=0.1,
|
1550 |
info="Set this to a larger value if you encounter OOM errors. Larger values cause slower speed. / 如果出现OOM错误,请将此值设置得更大。值越大,速度越慢。"
|
1551 |
)
|
|
|
216 |
|
217 |
# 加载模型
|
218 |
try:
|
219 |
+
text_encoder = LlamaModel.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='text_encoder', torch_dtype=dtype).to(model_device)
|
220 |
+
text_encoder_2 = CLIPTextModel.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='text_encoder_2', torch_dtype=dtype).to(model_device)
|
221 |
+
tokenizer = LlamaTokenizerFast.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='tokenizer')
|
222 |
+
tokenizer_2 = CLIPTokenizer.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='tokenizer_2')
|
223 |
+
vae = AutoencoderKLHunyuanVideo.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='vae', torch_dtype=dtype).to(model_device)
|
224 |
|
225 |
feature_extractor = SiglipImageProcessor.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='feature_extractor')
|
226 |
image_encoder = SiglipVisionModel.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='image_encoder', torch_dtype=dtype).to(model_device)
|
|
|
237 |
transformer_dtype = torch.float32
|
238 |
cpu_fallback_mode = True
|
239 |
|
240 |
+
text_encoder = LlamaModel.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='text_encoder', torch_dtype=dtype).to('cpu')
|
241 |
+
text_encoder_2 = CLIPTextModel.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='text_encoder_2', torch_dtype=dtype).to('cpu')
|
242 |
+
tokenizer = LlamaTokenizerFast.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='tokenizer')
|
243 |
+
tokenizer_2 = CLIPTokenizer.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='tokenizer_2')
|
244 |
+
vae = AutoencoderKLHunyuanVideo.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='vae', torch_dtype=dtype).to('cpu')
|
245 |
|
246 |
feature_extractor = SiglipImageProcessor.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='feature_extractor')
|
247 |
image_encoder = SiglipVisionModel.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='image_encoder', torch_dtype=dtype).to('cpu')
|
|
|
1101 |
|
1102 |
|
1103 |
quick_prompts = [
|
1104 |
+
'The girl dances gracefully, with clear movements, full of charm.',
|
1105 |
+
'A character doing some simple body movements.',
|
1106 |
]
|
1107 |
quick_prompts = [[x] for x in quick_prompts]
|
1108 |
|
|
|
1545 |
label="GPU Memory (GB) / GPU推理保留内存(GB)",
|
1546 |
minimum=6,
|
1547 |
maximum=128,
|
1548 |
+
value=6,
|
1549 |
step=0.1,
|
1550 |
info="Set this to a larger value if you encounter OOM errors. Larger values cause slower speed. / 如果出现OOM错误,请将此值设置得更大。值越大,速度越慢。"
|
1551 |
)
|