Spaces:
Runtime error
Runtime error
Deactivated git large coco
Browse files
app.py
CHANGED
@@ -14,8 +14,8 @@ preprocessor_blip_large = AutoProcessor.from_pretrained("Salesforce/blip-image-c
|
|
14 |
model_blip_large = BlipForConditionalGeneration.from_pretrained("Salesforce/blip-image-captioning-large")
|
15 |
|
16 |
# Load the GIT coco model
|
17 |
-
preprocessor_git_large_coco = AutoProcessor.from_pretrained("microsoft/git-large-coco")
|
18 |
-
model_git_large_coco = AutoModelForCausalLM.from_pretrained("microsoft/git-large-coco")
|
19 |
|
20 |
# Load the CLIP model
|
21 |
model_oc_coca, _, transform_oc_coca = open_clip.create_model_and_transforms(
|
@@ -136,7 +136,7 @@ def generate_captions(
|
|
136 |
caption_blip_large = generate_caption(preprocessor_blip_large, model_blip_large, image).strip()
|
137 |
|
138 |
# Generate captions for the image using the GIT coco model
|
139 |
-
caption_git_large_coco = generate_caption(preprocessor_git_large_coco, model_git_large_coco, image).strip()
|
140 |
|
141 |
# Generate captions for the image using the CLIP model
|
142 |
caption_oc_coca = generate_captions_clip(model_oc_coca, transform_oc_coca, image).strip()
|
@@ -157,7 +157,7 @@ iface = gr.Interface(
|
|
157 |
outputs=[
|
158 |
gr.outputs.Textbox(label="Blip base"),
|
159 |
gr.outputs.Textbox(label="Blip large"),
|
160 |
-
gr.outputs.Textbox(label="GIT large coco"),
|
161 |
gr.outputs.Textbox(label="CLIP"),
|
162 |
],
|
163 |
title="Image Captioning",
|
|
|
14 |
model_blip_large = BlipForConditionalGeneration.from_pretrained("Salesforce/blip-image-captioning-large")
|
15 |
|
16 |
# Load the GIT coco model
|
17 |
+
# preprocessor_git_large_coco = AutoProcessor.from_pretrained("microsoft/git-large-coco")
|
18 |
+
# model_git_large_coco = AutoModelForCausalLM.from_pretrained("microsoft/git-large-coco")
|
19 |
|
20 |
# Load the CLIP model
|
21 |
model_oc_coca, _, transform_oc_coca = open_clip.create_model_and_transforms(
|
|
|
136 |
caption_blip_large = generate_caption(preprocessor_blip_large, model_blip_large, image).strip()
|
137 |
|
138 |
# Generate captions for the image using the GIT coco model
|
139 |
+
# caption_git_large_coco = generate_caption(preprocessor_git_large_coco, model_git_large_coco, image).strip()
|
140 |
|
141 |
# Generate captions for the image using the CLIP model
|
142 |
caption_oc_coca = generate_captions_clip(model_oc_coca, transform_oc_coca, image).strip()
|
|
|
157 |
outputs=[
|
158 |
gr.outputs.Textbox(label="Blip base"),
|
159 |
gr.outputs.Textbox(label="Blip large"),
|
160 |
+
# gr.outputs.Textbox(label="GIT large coco"),
|
161 |
gr.outputs.Textbox(label="CLIP"),
|
162 |
],
|
163 |
title="Image Captioning",
|