Update README.md
Browse files
README.md
CHANGED
@@ -21,17 +21,21 @@ It achieves the following results on the evaluation set:
|
|
21 |
|
22 |
## Model description
|
23 |
|
24 |
-
|
25 |
|
26 |
## Intended uses & limitations
|
27 |
|
28 |
-
|
|
|
|
|
29 |
## Training and evaluation data
|
30 |
|
31 |
-
|
32 |
|
33 |
## Training procedure
|
34 |
|
|
|
|
|
35 |
### Training hyperparameters
|
36 |
|
37 |
The following hyperparameters were used during training:
|
@@ -59,4 +63,7 @@ The following hyperparameters were used during training:
|
|
59 |
- Transformers 4.35.2
|
60 |
- Pytorch 2.1.0+cu118
|
61 |
- Datasets 2.15.0
|
62 |
-
- Tokenizers 0.15.0
|
|
|
|
|
|
|
|
21 |
|
22 |
## Model description
|
23 |
|
24 |
+
Model has been fine-tuned from Llama 2 7B model for legal summarization tasks
|
25 |
|
26 |
## Intended uses & limitations
|
27 |
|
28 |
+
The model has been fine-tuned with legal summarization text for summarization tasks. Can produce repeating text when creating longer outputs.
|
29 |
+
Tested only with english and the bill_summary_us dataset.
|
30 |
+
|
31 |
## Training and evaluation data
|
32 |
|
33 |
+
Training and evaluation data is from the bill_summary_us. An around 1500 row split from the dataset was used with further split of 80:20 for training and evaluation.
|
34 |
|
35 |
## Training procedure
|
36 |
|
37 |
+
SFTTrainer from Hugging Face's TRL library used for fine-tuning process.
|
38 |
+
|
39 |
### Training hyperparameters
|
40 |
|
41 |
The following hyperparameters were used during training:
|
|
|
63 |
- Transformers 4.35.2
|
64 |
- Pytorch 2.1.0+cu118
|
65 |
- Datasets 2.15.0
|
66 |
+
- Tokenizers 0.15.0
|
67 |
+
|
68 |
+
### License
|
69 |
+
Llama 2 is licensed under the LLAMA 2 Community License, Copyright © Meta Platforms, Inc. All Rights Reserved.
|