Upload README.md
Browse files
README.md
CHANGED
@@ -23,19 +23,21 @@ license: cc-by-nc-4.0
|
|
23 |
**Model Architecture**
|
24 |
Poly-platypus-ko is an auto-regressive language model based on the LLaMA2 transformer architecture.
|
25 |
|
26 |
-
**Base Model**
|
27 |
[Polyglot-ko-12.8b](https://huggingface.co/EleutherAI/polyglot-ko-12.8b)
|
28 |
|
29 |
**Fine-tuning method**
|
30 |
Same as [KO-Platypus2](https://github.com/Marker-Inc-Korea/CoT-llama2).
|
31 |
|
32 |
-
**Training Dataset**
|
33 |
I use [KOpen-platypus dataset](https://huggingface.co/datasets/kyujinpy/KOpen-platypus).
|
34 |
I use A100 GPU 40GB and COLAB, when trianing.
|
35 |
|
36 |
# **Model Bechmark1**
|
37 |
|
38 |
-
##
|
|
|
|
|
39 |
| Model | Ko-ARC | Ko-HellaSwag | Ko-MMLU | Ko-TruthfulQA | Ko-CommonGen V2 |
|
40 |
| --- | --- | --- | --- | --- | --- |
|
41 |
| Poly-platypus-ko-12.8b(ours) | NaN | NaN | NaN | NaN | NaN |
|
|
|
23 |
**Model Architecture**
|
24 |
Poly-platypus-ko is an auto-regressive language model based on the LLaMA2 transformer architecture.
|
25 |
|
26 |
+
**Base Model**
|
27 |
[Polyglot-ko-12.8b](https://huggingface.co/EleutherAI/polyglot-ko-12.8b)
|
28 |
|
29 |
**Fine-tuning method**
|
30 |
Same as [KO-Platypus2](https://github.com/Marker-Inc-Korea/CoT-llama2).
|
31 |
|
32 |
+
**Training Dataset**
|
33 |
I use [KOpen-platypus dataset](https://huggingface.co/datasets/kyujinpy/KOpen-platypus).
|
34 |
I use A100 GPU 40GB and COLAB, when trianing.
|
35 |
|
36 |
# **Model Bechmark1**
|
37 |
|
38 |
+
## KO-LLM leaderboard
|
39 |
+
- Follow up as [Open KO-LLM LeaderBoard](https://huggingface.co/spaces/upstage/open-ko-llm-leaderboard).
|
40 |
+
|
41 |
| Model | Ko-ARC | Ko-HellaSwag | Ko-MMLU | Ko-TruthfulQA | Ko-CommonGen V2 |
|
42 |
| --- | --- | --- | --- | --- | --- |
|
43 |
| Poly-platypus-ko-12.8b(ours) | NaN | NaN | NaN | NaN | NaN |
|