MonteXiaofeng
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -2,15 +2,29 @@
|
|
2 |
license: apache-2.0
|
3 |
language:
|
4 |
- en
|
|
|
5 |
base_model:
|
6 |
- meta-llama/Meta-Llama-3.1-8B-Instruct
|
7 |
datasets:
|
8 |
- BAAI/IndustryInstruction_Finance-Economics
|
|
|
9 |
---
|
10 |
|
11 |
-
This model is
|
12 |
|
13 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
14 |
|
|
|
|
|
15 |
|
16 |
-
|
|
|
|
|
|
2 |
license: apache-2.0
|
3 |
language:
|
4 |
- en
|
5 |
+
- zh
|
6 |
base_model:
|
7 |
- meta-llama/Meta-Llama-3.1-8B-Instruct
|
8 |
datasets:
|
9 |
- BAAI/IndustryInstruction_Finance-Economics
|
10 |
+
- BAAI/IndustryInstruction
|
11 |
---
|
12 |
|
13 |
+
This model is finetuned on the model llama3.1-8b-instruct using the dataset [BAAI/IndustryInstruction_Finance-Economics](https://huggingface.co/datasets/BAAI/IndustryInstruction_Finance-Economics) dataset, dataset details can jump to the repo: [BAAI/IndustryInstruction](https://huggingface.co/datasets/BAAI/IndustryInstruction)
|
14 |
|
15 |
+
## training params
|
16 |
+
```
|
17 |
+
learning_rate=1e-5
|
18 |
+
lr_scheduler_type=cosine
|
19 |
+
max_length=2048
|
20 |
+
warmup_ratio=0.05
|
21 |
+
batch_size=64
|
22 |
+
epoch=10
|
23 |
+
```
|
24 |
|
25 |
+
select best ckpt by the evaluation loss
|
26 |
+
## evaluation
|
27 |
|
28 |
+
The following is an evaluation on the FinerBen dataset metrci. Since there are too many samples in the dataset, I randomly selected 500 samples from each dataset for evaluation.
|
29 |
+
|
30 |
+
![image/png](https://cdn-uploads.huggingface.co/production/uploads/642f6c64f945a8a5c9ee5b5d/shSgSkQ7nQqiBAl6IwBy5.png)
|