909ahmed commited on
Commit
f2be81d
·
verified ·
1 Parent(s): 2fc3055

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -5
app.py CHANGED
@@ -8,10 +8,11 @@ model, preprocess = clip.load("ViT-B/32", device=device)
8
 
9
  def process_image_and_text(image, text):
10
 
11
- text_list = text.tolist()
 
12
  image = preprocess(image).unsqueeze(0).to(device)
13
 
14
- text_tokens = clip.tokenize(text_list).to(device)
15
 
16
  with torch.no_grad():
17
  image_features = model.encode_image(image)
@@ -20,7 +21,7 @@ def process_image_and_text(image, text):
20
  logits_per_image, logits_per_text = model(image, text_tokens)
21
  probs = logits_per_image.softmax(dim=-1)
22
 
23
- return probs
24
 
25
- demo = gr.Interface(fn=process_image_and_text, inputs=[gr.inputs.Image(type="pil"), gr.inputs.Textbox()], outputs="text")
26
- demo.launch()
 
8
 
9
  def process_image_and_text(image, text):
10
 
11
+ text = text.split(",")
12
+ image = Image.fromarray(image)
13
  image = preprocess(image).unsqueeze(0).to(device)
14
 
15
+ text_tokens = clip.tokenize(text).to(device)
16
 
17
  with torch.no_grad():
18
  image_features = model.encode_image(image)
 
21
  logits_per_image, logits_per_text = model(image, text_tokens)
22
  probs = logits_per_image.softmax(dim=-1)
23
 
24
+ return probs.cpu().numpy()[0]
25
 
26
+ demo = gr.Interface(fn=process_image_and_text, inputs=['image', 'text'], outputs="text")
27
+ demo.launch()