hanzla commited on
Commit
455cd10
·
1 Parent(s): 62a5a51

adapters added

Browse files
Files changed (1) hide show
  1. app.py +11 -13
app.py CHANGED
@@ -21,24 +21,20 @@ adapter_options = {
21
  device = "cuda"
22
  adapter = MotionAdapter.from_pretrained("guoyww/animatediff-motion-adapter-v1-5-2", torch_dtype=torch.float16)
23
  model_id = "SG161222/Realistic_Vision_V5.1_noVAE"
24
- pipe = AnimateDiffPipeline.from_pretrained(model_id, motion_adapter=adapter, torch_dtype=torch.float16).to(device)
25
- pipe.load_lora_weights(
26
- "guoyww/animatediff-motion-lora-zoom-out", adapter_name="zoom-out",
27
- )
28
- pipe.load_lora_weights(
29
- "guoyww/animatediff-motion-lora-zoom-out", adapter_name="pan-left",
30
- )
31
- scheduler = DDIMScheduler.from_pretrained(
32
  model_id,
33
  subfolder="scheduler",
34
  clip_sample=False,
35
  timestep_spacing="linspace",
36
  beta_schedule="linear",
37
  steps_offset=1,
38
- )
39
- pipe.scheduler = scheduler
40
- @spaces.GPU
41
- def generate_video(prompt, guidance_scale, num_inference_steps, adapter_choices):
42
  pipe.to(device)
43
 
44
  # Set adapters based on user selection
@@ -53,7 +49,7 @@ def generate_video(prompt, guidance_scale, num_inference_steps, adapter_choices)
53
 
54
  output = pipe(
55
  prompt=prompt,
56
- negative_prompt="bad quality, worse quality",
57
  num_frames=16,
58
  guidance_scale=guidance_scale,
59
  num_inference_steps=num_inference_steps,
@@ -66,9 +62,11 @@ def generate_video(prompt, guidance_scale, num_inference_steps, adapter_choices)
66
 
67
 
68
  iface = gr.Interface(
 
69
  fn=generate_video,
70
  inputs=[
71
  gr.Textbox(label="Enter your prompt"),
 
72
  gr.Slider(minimum=0.5, maximum=10, value=7.5, label="Guidance Scale"),
73
  gr.Slider(minimum=4, maximum=24, step=4, value=4, label="Inference Steps"),
74
  gr.CheckboxGroup(adapter_options.keys(), label="Adapter Choice",type='value'),
 
21
  device = "cuda"
22
  adapter = MotionAdapter.from_pretrained("guoyww/animatediff-motion-adapter-v1-5-2", torch_dtype=torch.float16)
23
  model_id = "SG161222/Realistic_Vision_V5.1_noVAE"
24
+
25
+ @spaces.GPU
26
+ def generate_video(prompt,negative_prompt, guidance_scale, num_inference_steps, adapter_choices):
27
+ pipe = AnimateDiffPipeline.from_pretrained(model_id, motion_adapter=adapter, torch_dtype=torch.float16).to(device)
28
+ scheduler = DDIMScheduler.from_pretrained(
 
 
 
29
  model_id,
30
  subfolder="scheduler",
31
  clip_sample=False,
32
  timestep_spacing="linspace",
33
  beta_schedule="linear",
34
  steps_offset=1,
35
+ )
36
+ pipe.scheduler = scheduler
37
+
 
38
  pipe.to(device)
39
 
40
  # Set adapters based on user selection
 
49
 
50
  output = pipe(
51
  prompt=prompt,
52
+ negative_prompt=negative_prompt,
53
  num_frames=16,
54
  guidance_scale=guidance_scale,
55
  num_inference_steps=num_inference_steps,
 
62
 
63
 
64
  iface = gr.Interface(
65
+ theme=gr.themes.Monochrome(primary_hue="red", secondary_hue="pink"),
66
  fn=generate_video,
67
  inputs=[
68
  gr.Textbox(label="Enter your prompt"),
69
+ gr.Textbox(label="Negative Prompt"),
70
  gr.Slider(minimum=0.5, maximum=10, value=7.5, label="Guidance Scale"),
71
  gr.Slider(minimum=4, maximum=24, step=4, value=4, label="Inference Steps"),
72
  gr.CheckboxGroup(adapter_options.keys(), label="Adapter Choice",type='value'),