fffiloni commited on
Commit
f7f6486
·
verified ·
1 Parent(s): 9a622ff

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +28 -20
app.py CHANGED
@@ -7,6 +7,9 @@ import json
7
  import tempfile
8
  from huggingface_hub import snapshot_download
9
 
 
 
 
10
  # Download All Required Models using `snapshot_download`
11
 
12
  # Download Wan2.1-I2V-14B-480P model
@@ -79,7 +82,7 @@ GPU_TO_VRAM_PARAMS = {
79
  "NVIDIA A100-SXM4-40GB": 11000000000,
80
  "NVIDIA A100-SXM4-80GB": 22000000000,
81
  "NVIDIA L4": 5000000000,
82
- "NVIDIA L40S": 5000000000
83
  }
84
  USED_VRAM_PARAMS = GPU_TO_VRAM_PARAMS[gpu_name]
85
  print("Using", USED_VRAM_PARAMS, "for num_persistent_param_in_dit")
@@ -110,46 +113,51 @@ def create_temp_input_json(prompt: str, cond_image_path: str, cond_audio_path: s
110
  return temp_json_path
111
 
112
 
113
- def infer(prompt, cond_image_path, cond_audio_path):
114
-
115
  input_json_path = create_temp_input_json(prompt, cond_image_path, cond_audio_path)
116
 
117
- cmd = [
118
- "python3", "generate_multitalk.py",
119
  "--ckpt_dir", "weights/Wan2.1-I2V-14B-480P",
120
  "--wav2vec_dir", "weights/chinese-wav2vec2-base",
121
  "--input_json", input_json_path,
122
- "--sample_steps", "12",
123
- #"--motion_frame", "2",
124
- "--num_persistent_param_in_dit", str(USED_VRAM_PARAMS),
125
  "--mode", "streaming",
126
  "--use_teacache",
127
- "--save_file", "multi_long_mediumvram_exp"
128
  ]
129
 
130
- # Optional: log file
131
- log_file_path = "inference.log"
132
-
133
- # Run and stream logs in real-time
134
- with open(log_file_path, "w") as log_file:
 
 
 
 
 
 
 
 
 
135
  process = subprocess.Popen(
136
  cmd,
137
  stdout=subprocess.PIPE,
138
  stderr=subprocess.STDOUT,
139
  text=True,
140
- bufsize=1 # Line-buffered
141
  )
142
-
143
  for line in process.stdout:
144
- print(line, end="") # Print to console in real-time
145
- log_file.write(line) # Save to log file
146
-
147
  process.wait()
148
 
149
  if process.returncode != 0:
150
  raise RuntimeError("Inference failed. Check inference.log for details.")
151
 
152
- return "multi_long_mediumvram_exp.mp4"
153
 
154
 
155
  with gr.Blocks(title="MultiTalk Inference") as demo:
 
7
  import tempfile
8
  from huggingface_hub import snapshot_download
9
 
10
+ num_gpus = torch.cuda.device_count()
11
+ print(f"GPU AVAILABLE: {num_gpus}")
12
+
13
  # Download All Required Models using `snapshot_download`
14
 
15
  # Download Wan2.1-I2V-14B-480P model
 
82
  "NVIDIA A100-SXM4-40GB": 11000000000,
83
  "NVIDIA A100-SXM4-80GB": 22000000000,
84
  "NVIDIA L4": 5000000000,
85
+ "NVIDIA L40S": 11000000000
86
  }
87
  USED_VRAM_PARAMS = GPU_TO_VRAM_PARAMS[gpu_name]
88
  print("Using", USED_VRAM_PARAMS, "for num_persistent_param_in_dit")
 
113
  return temp_json_path
114
 
115
 
116
+ def infer(prompt, cond_image_path, cond_audio_path):
117
+ # Prepare input JSON
118
  input_json_path = create_temp_input_json(prompt, cond_image_path, cond_audio_path)
119
 
120
+ # Base args
121
+ common_args = [
122
  "--ckpt_dir", "weights/Wan2.1-I2V-14B-480P",
123
  "--wav2vec_dir", "weights/chinese-wav2vec2-base",
124
  "--input_json", input_json_path,
125
+ "--sample_steps", "6",
 
 
126
  "--mode", "streaming",
127
  "--use_teacache",
128
+ "--save_file", "multi_long_multigpu_exp"
129
  ]
130
 
131
+ if num_gpus > 1:
132
+ cmd = [
133
+ "torchrun",
134
+ f"--nproc_per_node={num_gpus}",
135
+ "--standalone",
136
+ "generate_multitalk.py",
137
+ "--dit_fsdp", "--t5_fsdp",
138
+ "--ulysses_size", str(num_gpus),
139
+ ] + common_args
140
+ else:
141
+ cmd = ["python3", "generate_multitalk.py"] + common_args
142
+
143
+ # Log to file and stream
144
+ with open("inference.log", "w") as log_file:
145
  process = subprocess.Popen(
146
  cmd,
147
  stdout=subprocess.PIPE,
148
  stderr=subprocess.STDOUT,
149
  text=True,
150
+ bufsize=1
151
  )
 
152
  for line in process.stdout:
153
+ print(line, end="")
154
+ log_file.write(line)
 
155
  process.wait()
156
 
157
  if process.returncode != 0:
158
  raise RuntimeError("Inference failed. Check inference.log for details.")
159
 
160
+ return "multi_long_multigpu_exp.mp4"
161
 
162
 
163
  with gr.Blocks(title="MultiTalk Inference") as demo: