Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -240,7 +240,7 @@ css = """
|
|
240 |
|
241 |
# Create the Gradio Interface
|
242 |
with gr.Blocks(css=css, theme="bethecloud/storj_theme") as demo:
|
243 |
-
gr.Markdown("# **VisionScope-
|
244 |
with gr.Row():
|
245 |
with gr.Column():
|
246 |
with gr.Tabs():
|
@@ -274,7 +274,7 @@ with gr.Blocks(css=css, theme="bethecloud/storj_theme") as demo:
|
|
274 |
value="SkyCaptioner-V1"
|
275 |
)
|
276 |
|
277 |
-
gr.Markdown("**Model Info 💻** | [Report Bug](https://huggingface.co/
|
278 |
gr.Markdown("> [SkyCaptioner-V1](https://huggingface.co/Skywork/SkyCaptioner-V1): structural video captioning model designed to generate high-quality, structural descriptions for video data. It integrates specialized sub-expert models.")
|
279 |
gr.Markdown("> [SpaceThinker-Qwen2.5VL-3B](https://huggingface.co/remyxai/SpaceThinker-Qwen2.5VL-3B): thinking/reasoning multimodal/vision-language model (VLM) trained to enhance spatial reasoning.")
|
280 |
gr.Markdown("> [coreOCR-7B-050325-preview](https://huggingface.co/prithivMLmods/coreOCR-7B-050325-preview): model is a fine-tuned version of qwen/qwen2-vl-7b, optimized for document-level optical character recognition (ocr), long-context vision-language understanding.")
|
|
|
240 |
|
241 |
# Create the Gradio Interface
|
242 |
with gr.Blocks(css=css, theme="bethecloud/storj_theme") as demo:
|
243 |
+
gr.Markdown("# **[VisionScope R2](https://huggingface.co/collections/prithivMLmods/multimodal-implementations-67c9982ea04b39f0608badb0)**")
|
244 |
with gr.Row():
|
245 |
with gr.Column():
|
246 |
with gr.Tabs():
|
|
|
274 |
value="SkyCaptioner-V1"
|
275 |
)
|
276 |
|
277 |
+
gr.Markdown("**Model Info 💻** | [Report Bug](https://huggingface.co/spaces/prithivMLmods/VisionScope-R2/discussions)")
|
278 |
gr.Markdown("> [SkyCaptioner-V1](https://huggingface.co/Skywork/SkyCaptioner-V1): structural video captioning model designed to generate high-quality, structural descriptions for video data. It integrates specialized sub-expert models.")
|
279 |
gr.Markdown("> [SpaceThinker-Qwen2.5VL-3B](https://huggingface.co/remyxai/SpaceThinker-Qwen2.5VL-3B): thinking/reasoning multimodal/vision-language model (VLM) trained to enhance spatial reasoning.")
|
280 |
gr.Markdown("> [coreOCR-7B-050325-preview](https://huggingface.co/prithivMLmods/coreOCR-7B-050325-preview): model is a fine-tuned version of qwen/qwen2-vl-7b, optimized for document-level optical character recognition (ocr), long-context vision-language understanding.")
|