Isotonic commited on
Commit
314fd33
1 Parent(s): ba1a9d3

End of training

Browse files
README.md CHANGED
@@ -3,9 +3,23 @@ license: apache-2.0
3
  base_model: google/t5-small-lm-adapt
4
  tags:
5
  - generated_from_trainer
 
 
 
 
6
  model-index:
7
  - name: plan_t5
8
- results: []
 
 
 
 
 
 
 
 
 
 
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -13,7 +27,14 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # plan_t5
15
 
16
- This model is a fine-tuned version of [google/t5-small-lm-adapt](https://huggingface.co/google/t5-small-lm-adapt) on an unknown dataset.
 
 
 
 
 
 
 
17
 
18
  ## Model description
19
 
 
3
  base_model: google/t5-small-lm-adapt
4
  tags:
5
  - generated_from_trainer
6
+ datasets:
7
+ - Isotonic/planner_dataset
8
+ metrics:
9
+ - rouge
10
  model-index:
11
  - name: plan_t5
12
+ results:
13
+ - task:
14
+ name: Summarization
15
+ type: summarization
16
+ dataset:
17
+ name: Isotonic/planner_dataset
18
+ type: Isotonic/planner_dataset
19
+ metrics:
20
+ - name: Rouge1
21
+ type: rouge
22
+ value: 58.1207
23
  ---
24
 
25
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
27
 
28
  # plan_t5
29
 
30
+ This model is a fine-tuned version of [google/t5-small-lm-adapt](https://huggingface.co/google/t5-small-lm-adapt) on the Isotonic/planner_dataset dataset.
31
+ It achieves the following results on the evaluation set:
32
+ - Loss: 1.4444
33
+ - Rouge1: 58.1207
34
+ - Rouge2: 24.3398
35
+ - Rougel: 58.1271
36
+ - Rougelsum: 58.1292
37
+ - Gen Len: 7.9747
38
 
39
  ## Model description
40
 
all_results.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "eval_gen_len": 7.9747384536871655,
4
+ "eval_loss": 1.4443535804748535,
5
+ "eval_rouge1": 58.1207,
6
+ "eval_rouge2": 24.3398,
7
+ "eval_rougeL": 58.1271,
8
+ "eval_rougeLsum": 58.1292,
9
+ "eval_runtime": 237.0503,
10
+ "eval_samples": 7838,
11
+ "eval_samples_per_second": 33.065,
12
+ "eval_steps_per_second": 4.134,
13
+ "train_loss": 1.643041113804542,
14
+ "train_runtime": 148.4875,
15
+ "train_samples": 31349,
16
+ "train_samples_per_second": 211.122,
17
+ "train_steps_per_second": 26.393
18
+ }
eval_results.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "eval_gen_len": 7.9747384536871655,
4
+ "eval_loss": 1.4443535804748535,
5
+ "eval_rouge1": 58.1207,
6
+ "eval_rouge2": 24.3398,
7
+ "eval_rougeL": 58.1271,
8
+ "eval_rougeLsum": 58.1292,
9
+ "eval_runtime": 237.0503,
10
+ "eval_samples": 7838,
11
+ "eval_samples_per_second": 33.065,
12
+ "eval_steps_per_second": 4.134
13
+ }
generation_config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "_from_model_config": true,
3
  "decoder_start_token_id": 0,
4
  "eos_token_id": 1,
5
  "pad_token_id": 0,
 
1
  {
 
2
  "decoder_start_token_id": 0,
3
  "eos_token_id": 1,
4
  "pad_token_id": 0,
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "train_loss": 1.643041113804542,
4
+ "train_runtime": 148.4875,
5
+ "train_samples": 31349,
6
+ "train_samples_per_second": 211.122,
7
+ "train_steps_per_second": 26.393
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 1.0,
5
+ "eval_steps": 500,
6
+ "global_step": 3919,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.13,
13
+ "grad_norm": 1.9183778762817383,
14
+ "learning_rate": 3.188775510204082e-05,
15
+ "loss": 2.8223,
16
+ "step": 500
17
+ },
18
+ {
19
+ "epoch": 0.26,
20
+ "grad_norm": 2.046781539916992,
21
+ "learning_rate": 4.655502392344498e-05,
22
+ "loss": 1.5043,
23
+ "step": 1000
24
+ },
25
+ {
26
+ "epoch": 0.38,
27
+ "grad_norm": 1.9903734922409058,
28
+ "learning_rate": 3.8580542264752795e-05,
29
+ "loss": 1.4817,
30
+ "step": 1500
31
+ },
32
+ {
33
+ "epoch": 0.51,
34
+ "grad_norm": 1.7653305530548096,
35
+ "learning_rate": 3.060606060606061e-05,
36
+ "loss": 1.4671,
37
+ "step": 2000
38
+ },
39
+ {
40
+ "epoch": 0.64,
41
+ "grad_norm": 1.9060845375061035,
42
+ "learning_rate": 2.2631578947368423e-05,
43
+ "loss": 1.4637,
44
+ "step": 2500
45
+ },
46
+ {
47
+ "epoch": 0.77,
48
+ "grad_norm": 1.784713625907898,
49
+ "learning_rate": 1.4657097288676236e-05,
50
+ "loss": 1.4588,
51
+ "step": 3000
52
+ },
53
+ {
54
+ "epoch": 0.89,
55
+ "grad_norm": 1.737605094909668,
56
+ "learning_rate": 6.6826156299840516e-06,
57
+ "loss": 1.4615,
58
+ "step": 3500
59
+ },
60
+ {
61
+ "epoch": 1.0,
62
+ "step": 3919,
63
+ "total_flos": 102290970089472.0,
64
+ "train_loss": 1.643041113804542,
65
+ "train_runtime": 148.4875,
66
+ "train_samples_per_second": 211.122,
67
+ "train_steps_per_second": 26.393
68
+ }
69
+ ],
70
+ "logging_steps": 500,
71
+ "max_steps": 3919,
72
+ "num_input_tokens_seen": 0,
73
+ "num_train_epochs": 1,
74
+ "save_steps": 500,
75
+ "total_flos": 102290970089472.0,
76
+ "train_batch_size": 8,
77
+ "trial_name": null,
78
+ "trial_params": null
79
+ }