ChangheeCho
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -15,14 +15,18 @@ tags:
|
|
15 |
|
16 |
## Model Details
|
17 |
|
18 |
-
**Input:** The model accepts only text input.
|
19 |
-
|
20 |
-
**Output:** The model produces text output exclusively.
|
21 |
-
|
22 |
**Model Architecture:**
|
23 |
|
24 |
urLLM-KO-7B is an auto-regressive language model that leverages an optimized transformer architecture derived from Llama-2-7b.
|
25 |
|
26 |
**Training Corpus**
|
27 |
|
28 |
-
The model was trained using selected datasets from Modu Corpus and Korean Wikipedia (approximately
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
|
16 |
## Model Details
|
17 |
|
|
|
|
|
|
|
|
|
18 |
**Model Architecture:**
|
19 |
|
20 |
urLLM-KO-7B is an auto-regressive language model that leverages an optimized transformer architecture derived from Llama-2-7b.
|
21 |
|
22 |
**Training Corpus**
|
23 |
|
24 |
+
The model was trained using selected datasets from Modu Corpus and Korean Wikipedia (approximately 28GB).
|
25 |
+
|
26 |
+
**Vocab Expansion**
|
27 |
+
|
28 |
+
|
29 |
+
| Model Name | Vocabulary Size | Description |
|
30 |
+
| --- | --- | --- |
|
31 |
+
| Original Solar | 32000 | Sentencepiece BPE |
|
32 |
+
| **Expanded urLLM-KO-7B** | 51385 | Sentencepiece BPE. Added Korean vocab and merges |
|