kyujinpy commited on
Commit
0367065
1 Parent(s): 4a1b062

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -11
README.md CHANGED
@@ -15,24 +15,28 @@ license: cc-by-nc-4.0
15
  ## Model Details
16
 
17
  **Model Developers** Kyujin Han (kyujinpy)
18
-
19
  **Input** Models input text only.
20
-
21
  **Output** Models generate text only.
22
-
23
- **Model Architecture**
24
  Poly-platypus-ko is an auto-regressive language model based on the LLaMA2 transformer architecture.
25
-
 
 
 
 
26
  **Base Model**
27
  [Polyglot-ko-12.8b](https://huggingface.co/EleutherAI/polyglot-ko-12.8b)
28
-
29
  **Fine-tuning method**
30
  Same as [KO-Platypus2](https://github.com/Marker-Inc-Korea/CoT-llama2).
31
-
32
  **Training Dataset**
33
  I use [KOpen-platypus dataset](https://huggingface.co/datasets/kyujinpy/KOpen-platypus).
34
  I use A100 GPU 40GB and COLAB, when trianing.
35
-
36
  ---
37
  # **Model Bechmark1**
38
 
@@ -45,9 +49,8 @@ I use A100 GPU 40GB and COLAB, when trianing.
45
  | [KoT-platypus2-7B](https://huggingface.co/kyujinpy/KoT-platypus2-7B) | 45.62 | 38.05 | 49.63 | 34.68 | 37.69 | 68.08 |
46
  | [KO-platypus2-7B-EX](https://huggingface.co/kyujinpy/KO-Platypus2-7B-ex) | 45.41 | 39.08 | 50.86 | 34.60 | 37.94 | 64.55 |
47
  | [42MARU/polyglot-ko-12.8b-instruct](https://huggingface.co/42MARU/polyglot-ko-12.8b-instruct) | 43.89 | 36.35 | 51.59 | 26.38 | 45.16 | 59.98 |
48
- | [FINDA-FIT/llama-p](https://huggingface.co/FINDA-FIT/llama-p) | 43.63 | 39.59 | 50.74 | 33.85 | 38.09 | 55.87 |
49
- | [lcw99/polyglot-ko-12.8b-chang-instruct-chat](https://huggingface.co/lcw99/polyglot-ko-12.8b-chang-instruct-chat) | 43.06 | 34.90 | 52.72 | 25.98 | 44.44 | 57.28 |
50
- > Compare with Top 5 SOTA models. (update: 10/01)
51
 
52
  ---
53
  # **Model Benchmark2**
 
15
  ## Model Details
16
 
17
  **Model Developers** Kyujin Han (kyujinpy)
18
+
19
  **Input** Models input text only.
20
+
21
  **Output** Models generate text only.
22
+
23
+ **Model Architecture**
24
  Poly-platypus-ko is an auto-regressive language model based on the LLaMA2 transformer architecture.
25
+
26
+ **Repo Link**
27
+ Github KO-platypus2: [KO-platypus2](https://github.com/Marker-Inc-Korea/KO-Platypus)
28
+ Github Poly-platypus-ko: [Poly-platypus-ko](https://github.com/KyujinHan/Poly-platypus-ko)
29
+
30
  **Base Model**
31
  [Polyglot-ko-12.8b](https://huggingface.co/EleutherAI/polyglot-ko-12.8b)
32
+
33
  **Fine-tuning method**
34
  Same as [KO-Platypus2](https://github.com/Marker-Inc-Korea/CoT-llama2).
35
+
36
  **Training Dataset**
37
  I use [KOpen-platypus dataset](https://huggingface.co/datasets/kyujinpy/KOpen-platypus).
38
  I use A100 GPU 40GB and COLAB, when trianing.
39
+
40
  ---
41
  # **Model Bechmark1**
42
 
 
49
  | [KoT-platypus2-7B](https://huggingface.co/kyujinpy/KoT-platypus2-7B) | 45.62 | 38.05 | 49.63 | 34.68 | 37.69 | 68.08 |
50
  | [KO-platypus2-7B-EX](https://huggingface.co/kyujinpy/KO-Platypus2-7B-ex) | 45.41 | 39.08 | 50.86 | 34.60 | 37.94 | 64.55 |
51
  | [42MARU/polyglot-ko-12.8b-instruct](https://huggingface.co/42MARU/polyglot-ko-12.8b-instruct) | 43.89 | 36.35 | 51.59 | 26.38 | 45.16 | 59.98 |
52
+ | [FINDA-FIT/llama-p](https://huggingface.co/FINDA-FIT/llama-p) | 43.63 | 39.59 | 50.74 | 33.85 | 38.09 | 55.87 |
53
+ > Compare with Top 4 SOTA models. (update: 10/01)
 
54
 
55
  ---
56
  # **Model Benchmark2**