Fiqa commited on
Commit
dbf999c
·
verified ·
1 Parent(s): 948eab4

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -110,11 +110,11 @@ def generate_caption_and_image(image, f, p, d):
110
  # return generated_image, generated_image1
111
  messages = [{"role": "user", "content": [{"type": "image"},{"type": "text", "text": "If I had to write a haiku for this one, it would be: "}]}]
112
  input_text = processor.apply_chat_template(messages, add_generation_prompt=True)
113
- inputs = processor(image,input_text,add_special_tokens=False,return_tensors="pt").to(device)
114
 
115
  output = model.generate(**inputs, max_new_tokens=30)
116
  caption =processor.decode(output[0])
117
- image = pipe(prompt,height=1024,width=1024,guidance_scale=3.5,num_inference_steps=50,max_sequence_length=512,generator=torch.Generator("cpu").manual_seed(0)).images[0]
118
  return image
119
  return None
120
  # Gradio UI
 
110
  # return generated_image, generated_image1
111
  messages = [{"role": "user", "content": [{"type": "image"},{"type": "text", "text": "If I had to write a haiku for this one, it would be: "}]}]
112
  input_text = processor.apply_chat_template(messages, add_generation_prompt=True)
113
+ inputs = processor(img,input_text,add_special_tokens=False,return_tensors="pt").to(device)
114
 
115
  output = model.generate(**inputs, max_new_tokens=30)
116
  caption =processor.decode(output[0])
117
+ image = pipe(caption,height=1024,width=1024,guidance_scale=3.5,num_inference_steps=50,max_sequence_length=512,generator=torch.Generator("cpu").manual_seed(0)).images[0]
118
  return image
119
  return None
120
  # Gradio UI