Spaces:
Runtime error
Runtime error
HeliosZhao
commited on
Commit
•
bd0ad83
1
Parent(s):
c57315b
update
Browse files- app.py +21 -14
- inference.py +1 -0
app.py
CHANGED
@@ -106,6 +106,13 @@ with gr.Blocks(css='style.css') as demo:
|
|
106 |
</div>
|
107 |
""")
|
108 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
109 |
|
110 |
with gr.Row():
|
111 |
with gr.Column():
|
@@ -136,14 +143,14 @@ with gr.Blocks(css='style.css') as demo:
|
|
136 |
placeholder='Example: "A panda is surfing"')
|
137 |
video_length = gr.Slider(label='Video length',
|
138 |
minimum=4,
|
139 |
-
maximum=
|
140 |
step=1,
|
141 |
-
value=
|
142 |
fps = gr.Slider(label='FPS',
|
143 |
minimum=1,
|
144 |
-
maximum=
|
145 |
step=1,
|
146 |
-
value=
|
147 |
seed = gr.Slider(label='Seed',
|
148 |
minimum=0,
|
149 |
maximum=100000,
|
@@ -207,8 +214,8 @@ with gr.Blocks(css='style.css') as demo:
|
|
207 |
[
|
208 |
'Make-A-Protagonist/ikun',
|
209 |
'A man is playing basketball on the beach, anime style.',
|
210 |
-
|
211 |
-
|
212 |
33,
|
213 |
50,
|
214 |
12.5,
|
@@ -224,8 +231,8 @@ with gr.Blocks(css='style.css') as demo:
|
|
224 |
[
|
225 |
'Make-A-Protagonist/huaqiang',
|
226 |
'Elon Musk walking down the street.',
|
227 |
-
|
228 |
-
|
229 |
33,
|
230 |
50,
|
231 |
12.5,
|
@@ -241,8 +248,8 @@ with gr.Blocks(css='style.css') as demo:
|
|
241 |
[
|
242 |
'Make-A-Protagonist/yanzi',
|
243 |
'A panda walking down the snowy street.',
|
244 |
-
|
245 |
-
|
246 |
33,
|
247 |
50,
|
248 |
12.5,
|
@@ -258,8 +265,8 @@ with gr.Blocks(css='style.css') as demo:
|
|
258 |
[
|
259 |
'Make-A-Protagonist/car-turn',
|
260 |
'A car moving in the desert.',
|
261 |
-
|
262 |
-
|
263 |
33,
|
264 |
50,
|
265 |
12.5,
|
@@ -275,8 +282,8 @@ with gr.Blocks(css='style.css') as demo:
|
|
275 |
[
|
276 |
'Make-A-Protagonist/car-turn',
|
277 |
'A Suzuki Jimny driving down a mountain road in the rain.',
|
278 |
-
|
279 |
-
|
280 |
33,
|
281 |
50,
|
282 |
12.5,
|
|
|
106 |
</div>
|
107 |
""")
|
108 |
|
109 |
+
gr.HTML("""
|
110 |
+
<p>We provide a <a href="https://github.com/Make-A-Protagonist/Make-A-Protagonist/blob/main/docs/demo_guidance.md"> Demo Guidance </a> to help users to choose hyperparameters when editing videos.
|
111 |
+
<p>You may duplicate the space and upgrade GPU for better performance and faster inference without waiting in the queue.
|
112 |
+
<p>Alternatively, try our GitHub <a href=https://github.com/Make-A-Protagonist/Make-A-Protagonist> code </a> on your GPU.
|
113 |
+
</p>""")
|
114 |
+
|
115 |
+
|
116 |
|
117 |
with gr.Row():
|
118 |
with gr.Column():
|
|
|
143 |
placeholder='Example: "A panda is surfing"')
|
144 |
video_length = gr.Slider(label='Video length',
|
145 |
minimum=4,
|
146 |
+
maximum=6,
|
147 |
step=1,
|
148 |
+
value=6)
|
149 |
fps = gr.Slider(label='FPS',
|
150 |
minimum=1,
|
151 |
+
maximum=6,
|
152 |
step=1,
|
153 |
+
value=3)
|
154 |
seed = gr.Slider(label='Seed',
|
155 |
minimum=0,
|
156 |
maximum=100000,
|
|
|
214 |
[
|
215 |
'Make-A-Protagonist/ikun',
|
216 |
'A man is playing basketball on the beach, anime style.',
|
217 |
+
6,
|
218 |
+
3,
|
219 |
33,
|
220 |
50,
|
221 |
12.5,
|
|
|
231 |
[
|
232 |
'Make-A-Protagonist/huaqiang',
|
233 |
'Elon Musk walking down the street.',
|
234 |
+
6,
|
235 |
+
3,
|
236 |
33,
|
237 |
50,
|
238 |
12.5,
|
|
|
248 |
[
|
249 |
'Make-A-Protagonist/yanzi',
|
250 |
'A panda walking down the snowy street.',
|
251 |
+
6,
|
252 |
+
3,
|
253 |
33,
|
254 |
50,
|
255 |
12.5,
|
|
|
265 |
[
|
266 |
'Make-A-Protagonist/car-turn',
|
267 |
'A car moving in the desert.',
|
268 |
+
6,
|
269 |
+
3,
|
270 |
33,
|
271 |
50,
|
272 |
12.5,
|
|
|
282 |
[
|
283 |
'Make-A-Protagonist/car-turn',
|
284 |
'A Suzuki Jimny driving down a mountain road in the rain.',
|
285 |
+
6,
|
286 |
+
3,
|
287 |
33,
|
288 |
50,
|
289 |
12.5,
|
inference.py
CHANGED
@@ -242,6 +242,7 @@ class InferencePipeline:
|
|
242 |
sample_indices = self.sample_indices
|
243 |
image_embed = [self.train_dataset.img_embeddings[idx] for idx in sample_indices]
|
244 |
image_embed = torch.stack(image_embed, dim=0).to(device=ddim_inv_latent.device, dtype=ddim_inv_latent.dtype) # F, 768 for UnCLIP-small # F,C
|
|
|
245 |
ref_image = None
|
246 |
|
247 |
# ipdb.set_trace()
|
|
|
242 |
sample_indices = self.sample_indices
|
243 |
image_embed = [self.train_dataset.img_embeddings[idx] for idx in sample_indices]
|
244 |
image_embed = torch.stack(image_embed, dim=0).to(device=ddim_inv_latent.device, dtype=ddim_inv_latent.dtype) # F, 768 for UnCLIP-small # F,C
|
245 |
+
image_embed = image_embed[:video_length]
|
246 |
ref_image = None
|
247 |
|
248 |
# ipdb.set_trace()
|