909ahmed commited on
Commit
fd2c1e0
·
1 Parent(s): 59575f1

code added

Browse files
Files changed (2) hide show
  1. app.py +24 -0
  2. requirements.txt +4 -0
app.py ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ import torch
3
+ import clip
4
+ from PIL import Image
5
+
6
+ device = "cuda" if torch.cuda.is_available() else "cpu"
7
+ model, preprocess = clip.load("ViT-B/32", device=device)
8
+
9
+ def clip(image, text):
10
+
11
+ image = preprocess(image).unsqueeze(0).to(device)
12
+ text = clip.tokenize([text]).to(device)
13
+
14
+ with torch.no_grad():
15
+ image_features = model.encode_image(image)
16
+ text_features = model.encode_text(text)
17
+
18
+ logits_per_image, logits_per_text = model(image, text)
19
+ probs = logits_per_image.softmax(dim=-1).cpu().numpy()
20
+
21
+ return probs[0]
22
+
23
+ demo = gr.Interface(fn=clip, inputs=["text", "image"], outputs="text")
24
+ demo.launch()
requirements.txt ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ pytorch
2
+ torchvision
3
+ # ftfy regex tqdm
4
+ git+https://github.com/openai/CLIP.git