Add context length to README
Browse files
README.md
CHANGED
@@ -24,9 +24,11 @@ SEA-LION stands for _Southeast Asian Languages In One Network_.
|
|
24 |
|
25 |
## Model Details
|
26 |
|
27 |
-
###
|
28 |
We performed instruction tuning in English and also in ASEAN languages such as Indonesian, Thai and Vietnamese on our [continued pre-trained Llama3 CPT 8B SEA-LIONv2](https://huggingface.co/aisingapore/llama3-8b-cpt-sea-lionv2-base), a decoder model using the Llama3 architecture, to create Llama3 8B SEA-LIONv2 Instruct.
|
29 |
|
|
|
|
|
30 |
### Benchmark Performance
|
31 |
We evaluated Llama3 8B SEA-LIONv2 Instruct on both general language capabilities and instruction-following capabilities.
|
32 |
|
|
|
24 |
|
25 |
## Model Details
|
26 |
|
27 |
+
### Model Description
|
28 |
We performed instruction tuning in English and also in ASEAN languages such as Indonesian, Thai and Vietnamese on our [continued pre-trained Llama3 CPT 8B SEA-LIONv2](https://huggingface.co/aisingapore/llama3-8b-cpt-sea-lionv2-base), a decoder model using the Llama3 architecture, to create Llama3 8B SEA-LIONv2 Instruct.
|
29 |
|
30 |
+
The model has a context length of 8192.
|
31 |
+
|
32 |
### Benchmark Performance
|
33 |
We evaluated Llama3 8B SEA-LIONv2 Instruct on both general language capabilities and instruction-following capabilities.
|
34 |
|