Upload README.md
Browse files
README.md
CHANGED
@@ -15,24 +15,28 @@ license: cc-by-nc-4.0
|
|
15 |
## Model Details
|
16 |
|
17 |
**Model Developers** Kyujin Han (kyujinpy)
|
18 |
-
|
19 |
**Input** Models input text only.
|
20 |
-
|
21 |
**Output** Models generate text only.
|
22 |
-
|
23 |
-
**Model Architecture**
|
24 |
Poly-platypus-ko is an auto-regressive language model based on the LLaMA2 transformer architecture.
|
25 |
-
|
|
|
|
|
|
|
|
|
26 |
**Base Model**
|
27 |
[Polyglot-ko-12.8b](https://huggingface.co/EleutherAI/polyglot-ko-12.8b)
|
28 |
-
|
29 |
**Fine-tuning method**
|
30 |
Same as [KO-Platypus2](https://github.com/Marker-Inc-Korea/CoT-llama2).
|
31 |
-
|
32 |
**Training Dataset**
|
33 |
I use [KOpen-platypus dataset](https://huggingface.co/datasets/kyujinpy/KOpen-platypus).
|
34 |
I use A100 GPU 40GB and COLAB, when trianing.
|
35 |
-
|
36 |
---
|
37 |
# **Model Bechmark1**
|
38 |
|
@@ -45,9 +49,8 @@ I use A100 GPU 40GB and COLAB, when trianing.
|
|
45 |
| [KoT-platypus2-7B](https://huggingface.co/kyujinpy/KoT-platypus2-7B) | 45.62 | 38.05 | 49.63 | 34.68 | 37.69 | 68.08 |
|
46 |
| [KO-platypus2-7B-EX](https://huggingface.co/kyujinpy/KO-Platypus2-7B-ex) | 45.41 | 39.08 | 50.86 | 34.60 | 37.94 | 64.55 |
|
47 |
| [42MARU/polyglot-ko-12.8b-instruct](https://huggingface.co/42MARU/polyglot-ko-12.8b-instruct) | 43.89 | 36.35 | 51.59 | 26.38 | 45.16 | 59.98 |
|
48 |
-
| [FINDA-FIT/llama-p](https://huggingface.co/FINDA-FIT/llama-p) | 43.63 | 39.59 | 50.74 | 33.85 | 38.09 | 55.87 |
|
49 |
-
|
50 |
-
> Compare with Top 5 SOTA models. (update: 10/01)
|
51 |
|
52 |
---
|
53 |
# **Model Benchmark2**
|
|
|
15 |
## Model Details
|
16 |
|
17 |
**Model Developers** Kyujin Han (kyujinpy)
|
18 |
+
|
19 |
**Input** Models input text only.
|
20 |
+
|
21 |
**Output** Models generate text only.
|
22 |
+
|
23 |
+
**Model Architecture**
|
24 |
Poly-platypus-ko is an auto-regressive language model based on the LLaMA2 transformer architecture.
|
25 |
+
|
26 |
+
**Repo Link**
|
27 |
+
Github KO-platypus2: [KO-platypus2](https://github.com/Marker-Inc-Korea/KO-Platypus)
|
28 |
+
Github Poly-platypus-ko: [Poly-platypus-ko](https://github.com/KyujinHan/Poly-platypus-ko)
|
29 |
+
|
30 |
**Base Model**
|
31 |
[Polyglot-ko-12.8b](https://huggingface.co/EleutherAI/polyglot-ko-12.8b)
|
32 |
+
|
33 |
**Fine-tuning method**
|
34 |
Same as [KO-Platypus2](https://github.com/Marker-Inc-Korea/CoT-llama2).
|
35 |
+
|
36 |
**Training Dataset**
|
37 |
I use [KOpen-platypus dataset](https://huggingface.co/datasets/kyujinpy/KOpen-platypus).
|
38 |
I use A100 GPU 40GB and COLAB, when trianing.
|
39 |
+
|
40 |
---
|
41 |
# **Model Bechmark1**
|
42 |
|
|
|
49 |
| [KoT-platypus2-7B](https://huggingface.co/kyujinpy/KoT-platypus2-7B) | 45.62 | 38.05 | 49.63 | 34.68 | 37.69 | 68.08 |
|
50 |
| [KO-platypus2-7B-EX](https://huggingface.co/kyujinpy/KO-Platypus2-7B-ex) | 45.41 | 39.08 | 50.86 | 34.60 | 37.94 | 64.55 |
|
51 |
| [42MARU/polyglot-ko-12.8b-instruct](https://huggingface.co/42MARU/polyglot-ko-12.8b-instruct) | 43.89 | 36.35 | 51.59 | 26.38 | 45.16 | 59.98 |
|
52 |
+
| [FINDA-FIT/llama-p](https://huggingface.co/FINDA-FIT/llama-p) | 43.63 | 39.59 | 50.74 | 33.85 | 38.09 | 55.87 |
|
53 |
+
> Compare with Top 4 SOTA models. (update: 10/01)
|
|
|
54 |
|
55 |
---
|
56 |
# **Model Benchmark2**
|