Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -44,7 +44,7 @@ document_qa_pipeline = pipeline("question-answering", model="deepset/roberta-bas
|
|
44 |
image_classification_pipeline = pipeline("image-classification", model="facebook/detr-resnet-50")
|
45 |
object_detection_pipeline = pipeline("object-detection", model="facebook/detr-resnet-50")
|
46 |
video_classification_pipeline = pipeline("video-classification", model="facebook/timesformer-base-finetuned-k400")
|
47 |
-
text_to_3d_pipeline
|
48 |
keypoint_detection_pipeline = pipeline("keypoint-detection", model="facebook/detr-resnet-50")
|
49 |
translation_pipeline = pipeline("translation_en_to_fr", model="Helsinki-NLP/opus-mt-en-fr")
|
50 |
summarization_pipeline = pipeline("summarization", model="facebook/bart-large-cnn")
|
@@ -113,10 +113,6 @@ def video_classification(video):
|
|
113 |
result = video_classification_pipeline(video)
|
114 |
return result
|
115 |
|
116 |
-
def text_to_3d(text):
|
117 |
-
result = text_to_3d_pipeline(text)
|
118 |
-
return result["image"]
|
119 |
-
|
120 |
def keypoint_detection(image):
|
121 |
result = keypoint_detection_pipeline(image)
|
122 |
return result
|
@@ -203,12 +199,7 @@ with gr.Blocks() as demo:
|
|
203 |
|
204 |
video_classification_generate.click(video_classification, inputs=video_classification_video, outputs=video_classification_output)
|
205 |
|
206 |
-
|
207 |
-
text_to_3d_text = gr.Textbox(label="Text")
|
208 |
-
text_to_3d_generate = gr.Button("Generate 3D")
|
209 |
-
text_to_3d_output = gr.Image(label="3D Model")
|
210 |
-
|
211 |
-
text_to_3d_generate.click(text_to_3d, inputs=text_to_3d_text, outputs=text_to_3d_output)
|
212 |
|
213 |
with gr.Tab("Keypoint Detection"):
|
214 |
keypoint_detection_image = gr.Image(label="Upload Image")
|
@@ -238,5 +229,12 @@ with gr.Blocks() as demo:
|
|
238 |
|
239 |
text_to_audio_generate.click(text_to_audio, inputs=text_to_audio_text, outputs=text_to_audio_output)
|
240 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
241 |
# Launch the Gradio interface
|
242 |
demo.launch()
|
|
|
44 |
image_classification_pipeline = pipeline("image-classification", model="facebook/detr-resnet-50")
|
45 |
object_detection_pipeline = pipeline("object-detection", model="facebook/detr-resnet-50")
|
46 |
video_classification_pipeline = pipeline("video-classification", model="facebook/timesformer-base-finetuned-k400")
|
47 |
+
# Removed text_to_3d_pipeline as it was causing issues
|
48 |
keypoint_detection_pipeline = pipeline("keypoint-detection", model="facebook/detr-resnet-50")
|
49 |
translation_pipeline = pipeline("translation_en_to_fr", model="Helsinki-NLP/opus-mt-en-fr")
|
50 |
summarization_pipeline = pipeline("summarization", model="facebook/bart-large-cnn")
|
|
|
113 |
result = video_classification_pipeline(video)
|
114 |
return result
|
115 |
|
|
|
|
|
|
|
|
|
116 |
def keypoint_detection(image):
|
117 |
result = keypoint_detection_pipeline(image)
|
118 |
return result
|
|
|
199 |
|
200 |
video_classification_generate.click(video_classification, inputs=video_classification_video, outputs=video_classification_output)
|
201 |
|
202 |
+
# Removed Text-to-3D tab as it was causing issues
|
|
|
|
|
|
|
|
|
|
|
203 |
|
204 |
with gr.Tab("Keypoint Detection"):
|
205 |
keypoint_detection_image = gr.Image(label="Upload Image")
|
|
|
229 |
|
230 |
text_to_audio_generate.click(text_to_audio, inputs=text_to_audio_text, outputs=text_to_audio_output)
|
231 |
|
232 |
+
with gr.Tab("Audio Classification"):
|
233 |
+
audio_classification_audio = gr.Audio(label="Upload Audio")
|
234 |
+
audio_classification_generate = gr.Button("Classify")
|
235 |
+
audio_classification_output = gr.Textbox(label="Classification Result")
|
236 |
+
|
237 |
+
audio_classification_generate.click(audio_classification, inputs=audio_classification_audio, outputs=audio_classification_output)
|
238 |
+
|
239 |
# Launch the Gradio interface
|
240 |
demo.launch()
|