frankmorales2020 commited on
Commit
4ea78b8
·
verified ·
1 Parent(s): 9a0df79

Model save

Browse files
Files changed (1) hide show
  1. README.md +3 -36
README.md CHANGED
@@ -20,15 +20,9 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.1](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1) on the generator dataset.
22
 
23
- Dataset: [dair-ai/emotion](https://huggingface.co/datasets/dair-ai/emotion)
24
-
25
  ## Model description
26
 
27
- Article: https://ai.plainenglish.io/fine-tuning-the-mistral-7b-instruct-v0-1-model-with-the-emotion-dataset-c84c50b553dc
28
-
29
- Fine tunning: https://github.com/frank-morales2020/MLxDL/blob/main/FineTuning_Mistral_7b_hfdeployment_dataset_Emotion.ipynb
30
-
31
- Evaluation: https://github.com/frank-morales2020/MLxDL/blob/main/FineTunning_Testing_For_EmotionQADataset.ipynb
32
 
33
  ## Intended uses & limitations
34
 
@@ -36,34 +30,7 @@ More information needed
36
 
37
  ## Training and evaluation data
38
 
39
- Evaluation: https://github.com/frank-morales2020/MLxDL/blob/main/FineTunning_Testing_For_EmotionQADataset.ipynb
40
-
41
-
42
- *************
43
-
44
- The following hyperparameters were used during training:
45
- learning_rate: 0.0002 train_batch_size: 3 eval_batch_size: 8 seed: 42 gradient_accumulation_steps: 2 total_train_batch_size: 6 optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 lr_scheduler_type: constant lr_scheduler_warmup_ratio: 0.03
46
- num_epochs: 1
47
-
48
- Accuracy (Eval dataset and predict) for a sample of 2000: 59.45%
49
-
50
- *************
51
-
52
- The following hyperparameters were used during training:
53
- learning_rate: 0.0002 train_batch_size: 3 eval_batch_size: 8 seed: 42 gradient_accumulation_steps: 2 total_train_batch_size: 6 optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 lr_scheduler_type: constant lr_scheduler_warmup_ratio: 0.03
54
- num_epochs: 25
55
-
56
- Accuracy (Eval dataset and predict) for a sample of 2000: 79.95%
57
-
58
- *************
59
-
60
- The following hyperparameters were used during training:
61
- learning_rate: 0.0002 train_batch_size: 3 eval_batch_size: 8 seed: 42 gradient_accumulation_steps: 2 total_train_batch_size: 6 optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 lr_scheduler_type: constant lr_scheduler_warmup_ratio: 0.03
62
- num_epochs: 40
63
-
64
- Accuracy (Eval dataset and predict) for a sample of 2000: 80.70%
65
-
66
- *************
67
 
68
  ## Training procedure
69
 
@@ -79,7 +46,7 @@ The following hyperparameters were used during training:
79
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
80
  - lr_scheduler_type: constant
81
  - lr_scheduler_warmup_ratio: 0.03
82
- - num_epochs: 40
83
 
84
  ### Training results
85
 
 
20
 
21
  This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.1](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1) on the generator dataset.
22
 
 
 
23
  ## Model description
24
 
25
+ More information needed
 
 
 
 
26
 
27
  ## Intended uses & limitations
28
 
 
30
 
31
  ## Training and evaluation data
32
 
33
+ More information needed
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
34
 
35
  ## Training procedure
36
 
 
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: constant
48
  - lr_scheduler_warmup_ratio: 0.03
49
+ - num_epochs: 100
50
 
51
  ### Training results
52