Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -122,7 +122,7 @@ example_images = [
|
|
122 |
iface = gr.Interface(
|
123 |
fn=predict_percept,
|
124 |
inputs=gr.Image(type="pil", label="Upload an Image"),
|
125 |
-
outputs=[gr.Textbox(label="Emotion"), gr.Textbox(label="Memorability Score"), gr.Textbox(label="IQA Score")]
|
126 |
title="PerceptCLIP",
|
127 |
description="This is an official demo of PerceptCLIP from the paper: [Don’t Judge Before You CLIP: A Unified Approach for Perceptual Tasks](https://arxiv.org/pdf/2503.13260). For each specific task, we fine-tune CLIP with LoRA and an MLP head. Our models achieve state-of-the-art performance. \nThis demo shows results from three models, each corresponding to a different task - visual emotion analysis, memorability prediction, and image quality assessment.",
|
128 |
examples=example_images
|
|
|
122 |
iface = gr.Interface(
|
123 |
fn=predict_percept,
|
124 |
inputs=gr.Image(type="pil", label="Upload an Image"),
|
125 |
+
outputs=[gr.Textbox(label="Emotion"), gr.Textbox(label="Memorability Score"), gr.Textbox(label="IQA Score")],
|
126 |
title="PerceptCLIP",
|
127 |
description="This is an official demo of PerceptCLIP from the paper: [Don’t Judge Before You CLIP: A Unified Approach for Perceptual Tasks](https://arxiv.org/pdf/2503.13260). For each specific task, we fine-tune CLIP with LoRA and an MLP head. Our models achieve state-of-the-art performance. \nThis demo shows results from three models, each corresponding to a different task - visual emotion analysis, memorability prediction, and image quality assessment.",
|
128 |
examples=example_images
|