inoculatemedia commited on
Commit
90ee3b7
·
verified ·
1 Parent(s): 30e045c

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +13 -13
app.py CHANGED
@@ -216,11 +216,11 @@ def load_models():
216
 
217
  # 加载模型
218
  try:
219
- text_encoder = LlamaModel.from_pretrained("hunyuanvideo-community/HunyuanVideo-I2V", subfolder='text_encoder', torch_dtype=dtype).to(model_device)
220
- text_encoder_2 = CLIPTextModel.from_pretrained("hunyuanvideo-community/HunyuanVideo-I2V", subfolder='text_encoder_2', torch_dtype=dtype).to(model_device)
221
- tokenizer = LlamaTokenizerFast.from_pretrained("hunyuanvideo-community/HunyuanVideo-I2V", subfolder='tokenizer')
222
- tokenizer_2 = CLIPTokenizer.from_pretrained("hunyuanvideo-community/HunyuanVideo-I2V", subfolder='tokenizer_2')
223
- vae = AutoencoderKLHunyuanVideo.from_pretrained("hunyuanvideo-community/HunyuanVideo-I2V", subfolder='vae', torch_dtype=dtype).to(model_device)
224
 
225
  feature_extractor = SiglipImageProcessor.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='feature_extractor')
226
  image_encoder = SiglipVisionModel.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='image_encoder', torch_dtype=dtype).to(model_device)
@@ -237,11 +237,11 @@ def load_models():
237
  transformer_dtype = torch.float32
238
  cpu_fallback_mode = True
239
 
240
- text_encoder = LlamaModel.from_pretrained("hunyuanvideo-community/HunyuanVideo-I2VHunyuanVideo-I2VHunyuanVideo-I2VHunyuanVideo-I2VHunyuanVideo-I2V", subfolder='text_encoder', torch_dtype=dtype).to('cpu')
241
- text_encoder_2 = CLIPTextModel.from_pretrained("hunyuanvideo-community/HunyuanVideo-I2VHunyuanVideo-I2VHunyuanVideo-I2VHunyuanVideo-I2V", subfolder='text_encoder_2', torch_dtype=dtype).to('cpu')
242
- tokenizer = LlamaTokenizerFast.from_pretrained("hunyuanvideo-community/HunyuanVideo-I2VHunyuanVideo-I2VHunyuanVideo-I2V", subfolder='tokenizer')
243
- tokenizer_2 = CLIPTokenizer.from_pretrained("hunyuanvideo-community/HunyuanVideo-I2VHunyuanVideo-I2V", subfolder='tokenizer_2')
244
- vae = AutoencoderKLHunyuanVideo.from_pretrained("hunyuanvideo-community/HunyuanVideo-I2V", subfolder='vae', torch_dtype=dtype).to('cpu')
245
 
246
  feature_extractor = SiglipImageProcessor.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='feature_extractor')
247
  image_encoder = SiglipVisionModel.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='image_encoder', torch_dtype=dtype).to('cpu')
@@ -1101,8 +1101,8 @@ def end_process():
1101
 
1102
 
1103
  quick_prompts = [
1104
- 'The camera smoothly orbits around the center of the scene, keeping the center point fixed and always in view. The characters are doing some simple body movements.',
1105
-
1106
  ]
1107
  quick_prompts = [[x] for x in quick_prompts]
1108
 
@@ -1545,7 +1545,7 @@ with block:
1545
  label="GPU Memory (GB) / GPU推理保留内存(GB)",
1546
  minimum=6,
1547
  maximum=128,
1548
- value=8,
1549
  step=0.1,
1550
  info="Set this to a larger value if you encounter OOM errors. Larger values cause slower speed. / 如果出现OOM错误,请将此值设置得更大。值越大,速度越慢。"
1551
  )
 
216
 
217
  # 加载模型
218
  try:
219
+ text_encoder = LlamaModel.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='text_encoder', torch_dtype=dtype).to(model_device)
220
+ text_encoder_2 = CLIPTextModel.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='text_encoder_2', torch_dtype=dtype).to(model_device)
221
+ tokenizer = LlamaTokenizerFast.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='tokenizer')
222
+ tokenizer_2 = CLIPTokenizer.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='tokenizer_2')
223
+ vae = AutoencoderKLHunyuanVideo.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='vae', torch_dtype=dtype).to(model_device)
224
 
225
  feature_extractor = SiglipImageProcessor.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='feature_extractor')
226
  image_encoder = SiglipVisionModel.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='image_encoder', torch_dtype=dtype).to(model_device)
 
237
  transformer_dtype = torch.float32
238
  cpu_fallback_mode = True
239
 
240
+ text_encoder = LlamaModel.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='text_encoder', torch_dtype=dtype).to('cpu')
241
+ text_encoder_2 = CLIPTextModel.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='text_encoder_2', torch_dtype=dtype).to('cpu')
242
+ tokenizer = LlamaTokenizerFast.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='tokenizer')
243
+ tokenizer_2 = CLIPTokenizer.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='tokenizer_2')
244
+ vae = AutoencoderKLHunyuanVideo.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='vae', torch_dtype=dtype).to('cpu')
245
 
246
  feature_extractor = SiglipImageProcessor.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='feature_extractor')
247
  image_encoder = SiglipVisionModel.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='image_encoder', torch_dtype=dtype).to('cpu')
 
1101
 
1102
 
1103
  quick_prompts = [
1104
+ 'The girl dances gracefully, with clear movements, full of charm.',
1105
+ 'A character doing some simple body movements.',
1106
  ]
1107
  quick_prompts = [[x] for x in quick_prompts]
1108
 
 
1545
  label="GPU Memory (GB) / GPU推理保留内存(GB)",
1546
  minimum=6,
1547
  maximum=128,
1548
+ value=6,
1549
  step=0.1,
1550
  info="Set this to a larger value if you encounter OOM errors. Larger values cause slower speed. / 如果出现OOM错误,请将此值设置得更大。值越大,速度越慢。"
1551
  )