Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -243,6 +243,7 @@ def get_video_duration(video_path):
|
|
243 |
@torch.inference_mode()
|
244 |
@torch.no_grad()
|
245 |
def synthesize_video_with_audio(video_file, caption, cot):
|
|
|
246 |
video_path = video_file
|
247 |
if caption is None:
|
248 |
caption = ''
|
@@ -255,7 +256,6 @@ def synthesize_video_with_audio(video_file, caption, cot):
|
|
255 |
preprocesser = VGGSound(duration_sec=duration_sec)
|
256 |
data = preprocesser.sample(video_path, caption, cot)
|
257 |
|
258 |
-
yield "⏳ Extracting Features…", None
|
259 |
|
260 |
preprocessed_data = {}
|
261 |
metaclip_global_text_features, metaclip_text_features = feature_extractor.encode_text(data['caption'])
|
|
|
243 |
@torch.inference_mode()
|
244 |
@torch.no_grad()
|
245 |
def synthesize_video_with_audio(video_file, caption, cot):
|
246 |
+
yield "⏳ Extracting Features…", None
|
247 |
video_path = video_file
|
248 |
if caption is None:
|
249 |
caption = ''
|
|
|
256 |
preprocesser = VGGSound(duration_sec=duration_sec)
|
257 |
data = preprocesser.sample(video_path, caption, cot)
|
258 |
|
|
|
259 |
|
260 |
preprocessed_data = {}
|
261 |
metaclip_global_text_features, metaclip_text_features = feature_extractor.encode_text(data['caption'])
|