Spaces:
Running
Running
Update content.py
Browse files- content.py +9 -3
content.py
CHANGED
@@ -13,9 +13,9 @@ Bottom_logo = f'''<img src="data:image/jpeg;base64,{bottom_logo}" style="width:2
|
|
13 |
intro_md = f'''
|
14 |
# {benchname} Leaderboard
|
15 |
|
16 |
-
* [Dataset](https://huggingface.co/datasets/maum-ai/KOFFVQA_Data)
|
17 |
-
* [Evaluation Code](https://github.com/maum-ai/KOFFVQA)
|
18 |
-
* Report (
|
19 |
|
20 |
{benchname}π is a Free-Form VQA benchmark dataset designed to evaluate Vision-Language Models (VLMs) in Korean language environments. Unlike traditional multiple-choice or predefined answer formats, KOFFVQA challenges models to generate open-ended, natural-language answers to visually grounded questions. This allows for a more comprehensive assessment of a model's ability to understand and generate nuanced Korean responses.
|
21 |
|
@@ -35,6 +35,8 @@ This benchmark includes a total of 275 Korean questions across 10 tasks. The que
|
|
35 |
|
36 |
## News
|
37 |
|
|
|
|
|
38 |
* **2025-01-21**: [Evaluation code](https://github.com/maum-ai/KOFFVQA) and [dataset](https://huggingface.co/datasets/maum-ai/KOFFVQA_Data) release
|
39 |
|
40 |
* **2024-12-06**: Leaderboard Release!
|
@@ -47,4 +49,8 @@ submit_md = f'''
|
|
47 |
|
48 |
We are not accepting model addition requests at the moment. Once the request system is established, we will start accepting requests.
|
49 |
|
|
|
|
|
|
|
|
|
50 |
'''.strip()
|
|
|
13 |
intro_md = f'''
|
14 |
# {benchname} Leaderboard
|
15 |
|
16 |
+
* [π Dataset](https://huggingface.co/datasets/maum-ai/KOFFVQA_Data)
|
17 |
+
* [π§ͺ Evaluation Code](https://github.com/maum-ai/KOFFVQA)
|
18 |
+
* [π Report] (https://arxiv.org/abs/2503.23730)
|
19 |
|
20 |
{benchname}π is a Free-Form VQA benchmark dataset designed to evaluate Vision-Language Models (VLMs) in Korean language environments. Unlike traditional multiple-choice or predefined answer formats, KOFFVQA challenges models to generate open-ended, natural-language answers to visually grounded questions. This allows for a more comprehensive assessment of a model's ability to understand and generate nuanced Korean responses.
|
21 |
|
|
|
35 |
|
36 |
## News
|
37 |
|
38 |
+
* **2025-04-01** : Our paper [KOFFVQA: An Objectively Evaluated Free-form VQA Benchmark for Large Vision-Language Models in the Korean Language](https://arxiv.org/abs/2503.23730) has released and accepted to CVPRW 2025, Workshop on Benchmarking and Expanding AI Multimodal Approaches(BEAM 2025) π
|
39 |
+
|
40 |
* **2025-01-21**: [Evaluation code](https://github.com/maum-ai/KOFFVQA) and [dataset](https://huggingface.co/datasets/maum-ai/KOFFVQA_Data) release
|
41 |
|
42 |
* **2024-12-06**: Leaderboard Release!
|
|
|
49 |
|
50 |
We are not accepting model addition requests at the moment. Once the request system is established, we will start accepting requests.
|
51 |
|
52 |
+
π Curious how your VLM performs in Korean? Use our [Evaluation code](https://github.com/maum-ai/KOFFVQA) to run it on KOFFVQA and check the score.
|
53 |
+
|
54 |
+
π§ββοΈ We currently use google/gemma-2-9b-it as the judge model, so there's no need to worry about API keys or usage fees.
|
55 |
+
|
56 |
'''.strip()
|