inoculatemedia commited on
Commit
30e045c
·
verified ·
1 Parent(s): f6e5f0e

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +10 -10
app.py CHANGED
@@ -216,11 +216,11 @@ def load_models():
216
 
217
  # 加载模型
218
  try:
219
- text_encoder = LlamaModel.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='text_encoder', torch_dtype=dtype).to(model_device)
220
- text_encoder_2 = CLIPTextModel.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='text_encoder_2', torch_dtype=dtype).to(model_device)
221
- tokenizer = LlamaTokenizerFast.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='tokenizer')
222
- tokenizer_2 = CLIPTokenizer.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='tokenizer_2')
223
- vae = AutoencoderKLHunyuanVideo.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='vae', torch_dtype=dtype).to(model_device)
224
 
225
  feature_extractor = SiglipImageProcessor.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='feature_extractor')
226
  image_encoder = SiglipVisionModel.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='image_encoder', torch_dtype=dtype).to(model_device)
@@ -237,11 +237,11 @@ def load_models():
237
  transformer_dtype = torch.float32
238
  cpu_fallback_mode = True
239
 
240
- text_encoder = LlamaModel.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='text_encoder', torch_dtype=dtype).to('cpu')
241
- text_encoder_2 = CLIPTextModel.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='text_encoder_2', torch_dtype=dtype).to('cpu')
242
- tokenizer = LlamaTokenizerFast.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='tokenizer')
243
- tokenizer_2 = CLIPTokenizer.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='tokenizer_2')
244
- vae = AutoencoderKLHunyuanVideo.from_pretrained("hunyuanvideo-community/HunyuanVideo", subfolder='vae', torch_dtype=dtype).to('cpu')
245
 
246
  feature_extractor = SiglipImageProcessor.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='feature_extractor')
247
  image_encoder = SiglipVisionModel.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='image_encoder', torch_dtype=dtype).to('cpu')
 
216
 
217
  # 加载模型
218
  try:
219
+ text_encoder = LlamaModel.from_pretrained("hunyuanvideo-community/HunyuanVideo-I2V", subfolder='text_encoder', torch_dtype=dtype).to(model_device)
220
+ text_encoder_2 = CLIPTextModel.from_pretrained("hunyuanvideo-community/HunyuanVideo-I2V", subfolder='text_encoder_2', torch_dtype=dtype).to(model_device)
221
+ tokenizer = LlamaTokenizerFast.from_pretrained("hunyuanvideo-community/HunyuanVideo-I2V", subfolder='tokenizer')
222
+ tokenizer_2 = CLIPTokenizer.from_pretrained("hunyuanvideo-community/HunyuanVideo-I2V", subfolder='tokenizer_2')
223
+ vae = AutoencoderKLHunyuanVideo.from_pretrained("hunyuanvideo-community/HunyuanVideo-I2V", subfolder='vae', torch_dtype=dtype).to(model_device)
224
 
225
  feature_extractor = SiglipImageProcessor.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='feature_extractor')
226
  image_encoder = SiglipVisionModel.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='image_encoder', torch_dtype=dtype).to(model_device)
 
237
  transformer_dtype = torch.float32
238
  cpu_fallback_mode = True
239
 
240
+ text_encoder = LlamaModel.from_pretrained("hunyuanvideo-community/HunyuanVideo-I2VHunyuanVideo-I2VHunyuanVideo-I2VHunyuanVideo-I2VHunyuanVideo-I2V", subfolder='text_encoder', torch_dtype=dtype).to('cpu')
241
+ text_encoder_2 = CLIPTextModel.from_pretrained("hunyuanvideo-community/HunyuanVideo-I2VHunyuanVideo-I2VHunyuanVideo-I2VHunyuanVideo-I2V", subfolder='text_encoder_2', torch_dtype=dtype).to('cpu')
242
+ tokenizer = LlamaTokenizerFast.from_pretrained("hunyuanvideo-community/HunyuanVideo-I2VHunyuanVideo-I2VHunyuanVideo-I2V", subfolder='tokenizer')
243
+ tokenizer_2 = CLIPTokenizer.from_pretrained("hunyuanvideo-community/HunyuanVideo-I2VHunyuanVideo-I2V", subfolder='tokenizer_2')
244
+ vae = AutoencoderKLHunyuanVideo.from_pretrained("hunyuanvideo-community/HunyuanVideo-I2V", subfolder='vae', torch_dtype=dtype).to('cpu')
245
 
246
  feature_extractor = SiglipImageProcessor.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='feature_extractor')
247
  image_encoder = SiglipVisionModel.from_pretrained("lllyasviel/flux_redux_bfl", subfolder='image_encoder', torch_dtype=dtype).to('cpu')