prithivMLmods commited on
Commit
cc1a62d
·
verified ·
1 Parent(s): 400b49f

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -240,7 +240,7 @@ css = """
240
 
241
  # Create the Gradio Interface
242
  with gr.Blocks(css=css, theme="bethecloud/storj_theme") as demo:
243
- gr.Markdown("# **VisionScope-R2**")
244
  with gr.Row():
245
  with gr.Column():
246
  with gr.Tabs():
@@ -274,7 +274,7 @@ with gr.Blocks(css=css, theme="bethecloud/storj_theme") as demo:
274
  value="SkyCaptioner-V1"
275
  )
276
 
277
- gr.Markdown("**Model Info 💻** | [Report Bug](https://huggingface.co/collections/prithivMLmods/multimodal-implementations-67c9982ea04b39f0608badb0)")
278
  gr.Markdown("> [SkyCaptioner-V1](https://huggingface.co/Skywork/SkyCaptioner-V1): structural video captioning model designed to generate high-quality, structural descriptions for video data. It integrates specialized sub-expert models.")
279
  gr.Markdown("> [SpaceThinker-Qwen2.5VL-3B](https://huggingface.co/remyxai/SpaceThinker-Qwen2.5VL-3B): thinking/reasoning multimodal/vision-language model (VLM) trained to enhance spatial reasoning.")
280
  gr.Markdown("> [coreOCR-7B-050325-preview](https://huggingface.co/prithivMLmods/coreOCR-7B-050325-preview): model is a fine-tuned version of qwen/qwen2-vl-7b, optimized for document-level optical character recognition (ocr), long-context vision-language understanding.")
 
240
 
241
  # Create the Gradio Interface
242
  with gr.Blocks(css=css, theme="bethecloud/storj_theme") as demo:
243
+ gr.Markdown("# **[VisionScope R2](https://huggingface.co/collections/prithivMLmods/multimodal-implementations-67c9982ea04b39f0608badb0)**")
244
  with gr.Row():
245
  with gr.Column():
246
  with gr.Tabs():
 
274
  value="SkyCaptioner-V1"
275
  )
276
 
277
+ gr.Markdown("**Model Info 💻** | [Report Bug](https://huggingface.co/spaces/prithivMLmods/VisionScope-R2/discussions)")
278
  gr.Markdown("> [SkyCaptioner-V1](https://huggingface.co/Skywork/SkyCaptioner-V1): structural video captioning model designed to generate high-quality, structural descriptions for video data. It integrates specialized sub-expert models.")
279
  gr.Markdown("> [SpaceThinker-Qwen2.5VL-3B](https://huggingface.co/remyxai/SpaceThinker-Qwen2.5VL-3B): thinking/reasoning multimodal/vision-language model (VLM) trained to enhance spatial reasoning.")
280
  gr.Markdown("> [coreOCR-7B-050325-preview](https://huggingface.co/prithivMLmods/coreOCR-7B-050325-preview): model is a fine-tuned version of qwen/qwen2-vl-7b, optimized for document-level optical character recognition (ocr), long-context vision-language understanding.")