Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: other
|
3 |
+
license_name: exaone
|
4 |
+
license_link: LICENSE
|
5 |
+
language:
|
6 |
+
- en
|
7 |
+
- ko
|
8 |
+
tags:
|
9 |
+
- lg-ai
|
10 |
+
- exaone
|
11 |
+
- exaone-3.5
|
12 |
+
pipeline_tag: text-generation
|
13 |
+
library_name: transformers
|
14 |
+
---
|
15 |
+
|
16 |
+
|
17 |
+
# Updates in EXAONE-3.5
|
18 |
+
|
19 |
+
## Key Changes
|
20 |
+
- **RoPE Scaling Parameter**: Added to support longer `context_length`.
|
21 |
+
- **Memory Optimization**: For the 2.4B model, `tie_word_embeddings` is set to `True` for improved memory efficiency.
|
22 |
+
|
23 |
+
⚠️ Using the original [Llamafy script](https://huggingface.co/maywell/EXAONE-3.0-7.8B-Instruct-Llamafied) as-is may lead to performance degradation.
|
24 |
+
|
25 |
+
To address this, I have updated the script and uploaded the Llamafied version of the model.
|
26 |
+
|
27 |
+
## Special Thanks
|
28 |
+
|
29 |
+
- **[@maywell](https://huggingface.co/maywell)**
|
30 |
+
For updating the code and uploading the model.
|
31 |
+
|
32 |
+
- **LG AI Research**
|
33 |
+
For releasing the original model.
|
34 |
+
Check out the [original release here](https://huggingface.co/collections/LGAI-EXAONE/exaone-35-674d0e1bb3dcd2ab6f39dbb4).
|