Updated the dataset info
Browse files
utils.py
CHANGED
@@ -55,20 +55,19 @@ TABLE_INTRODUCTION = """
|
|
55 |
|
56 |
LEADERBOARD_INFO = """
|
57 |
## Dataset Summary
|
58 |
-
- **Questions and
|
59 |
-
- **Sources:**
|
60 |
-
- **
|
61 |
-
- **
|
62 |
-
- **
|
63 |
-
- **SciBench:** Science questions from college exams.
|
64 |
"""
|
65 |
|
66 |
CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results"
|
67 |
CITATION_BUTTON_TEXT = r"""
|
68 |
-
@article{
|
69 |
-
title={
|
70 |
-
author={
|
71 |
-
journal={arXiv preprint arXiv:2406.
|
72 |
year={2024}
|
73 |
}
|
74 |
"""
|
|
|
55 |
|
56 |
LEADERBOARD_INFO = """
|
57 |
## Dataset Summary
|
58 |
+
- **Questions and Labels:** The task is to decide whether the provided explanation fully explains the joke (good) or does not fully explain the joke (bad).
|
59 |
+
- **Sources:**
|
60 |
+
- **Jokes:** We construct our dataset by including RZB jokes from "Best Annual Threads" between 2018 and 2021 that have been previously crawled (https://github.com/Leymore/ruozhiba). In addition, we directly collect all threads in the "Moderator's Recommendation" section from RZB.
|
61 |
+
- **Explanations:** We source the explanations from GPT-4o and ERNIE-4-turbo.
|
62 |
+
- **Annotations:** We manually annotate the generated explanations as either "fully explain the joke" (good) or "partially explain or not explain the joke" (bad). The gold label is determined by the majority vote among five native Chinese speakers.
|
|
|
63 |
"""
|
64 |
|
65 |
CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results"
|
66 |
CITATION_BUTTON_TEXT = r"""
|
67 |
+
@article{he2024chumor,
|
68 |
+
title={Chumor 1.0: A Truly Funny and Challenging Chinese Humor Understanding Dataset from Ruo Zhi Ba},
|
69 |
+
author={He, Ruiqi and He, Yushu and Bai, Longju and Liu, Jiarui and Sun, Zhenjie and Tang, Zenghao and Wang, He and Xia, Hanchen and Deng, Naihao},
|
70 |
+
journal={arXiv preprint arXiv:2406.12754},
|
71 |
year={2024}
|
72 |
}
|
73 |
"""
|