kyujinpy commited on
Commit
69e6d59
1 Parent(s): 158d815

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -3
README.md CHANGED
@@ -23,19 +23,21 @@ license: cc-by-nc-4.0
23
  **Model Architecture**
24
  Poly-platypus-ko is an auto-regressive language model based on the LLaMA2 transformer architecture.
25
 
26
- **Base Model**
27
  [Polyglot-ko-12.8b](https://huggingface.co/EleutherAI/polyglot-ko-12.8b)
28
 
29
  **Fine-tuning method**
30
  Same as [KO-Platypus2](https://github.com/Marker-Inc-Korea/CoT-llama2).
31
 
32
- **Training Dataset**
33
  I use [KOpen-platypus dataset](https://huggingface.co/datasets/kyujinpy/KOpen-platypus).
34
  I use A100 GPU 40GB and COLAB, when trianing.
35
 
36
  # **Model Bechmark1**
37
 
38
- ## [KO-LLM leaderboard](https://huggingface.co/spaces/upstage/open-ko-llm-leaderboard)
 
 
39
  | Model | Ko-ARC | Ko-HellaSwag | Ko-MMLU | Ko-TruthfulQA | Ko-CommonGen V2 |
40
  | --- | --- | --- | --- | --- | --- |
41
  | Poly-platypus-ko-12.8b(ours) | NaN | NaN | NaN | NaN | NaN |
 
23
  **Model Architecture**
24
  Poly-platypus-ko is an auto-regressive language model based on the LLaMA2 transformer architecture.
25
 
26
+ **Base Model**
27
  [Polyglot-ko-12.8b](https://huggingface.co/EleutherAI/polyglot-ko-12.8b)
28
 
29
  **Fine-tuning method**
30
  Same as [KO-Platypus2](https://github.com/Marker-Inc-Korea/CoT-llama2).
31
 
32
+ **Training Dataset**
33
  I use [KOpen-platypus dataset](https://huggingface.co/datasets/kyujinpy/KOpen-platypus).
34
  I use A100 GPU 40GB and COLAB, when trianing.
35
 
36
  # **Model Bechmark1**
37
 
38
+ ## KO-LLM leaderboard
39
+ - Follow up as [Open KO-LLM LeaderBoard](https://huggingface.co/spaces/upstage/open-ko-llm-leaderboard).
40
+
41
  | Model | Ko-ARC | Ko-HellaSwag | Ko-MMLU | Ko-TruthfulQA | Ko-CommonGen V2 |
42
  | --- | --- | --- | --- | --- | --- |
43
  | Poly-platypus-ko-12.8b(ours) | NaN | NaN | NaN | NaN | NaN |