junthbasnet commited on
Commit
f089482
·
verified ·
1 Parent(s): ada8f6b

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,51 +1,7 @@
1
  {
2
  "_name_or_path": "t5-small",
3
- "adapters": {
4
- "adapters": {
5
- "imdb_pfeiffer": "9076f36a74755ac4"
6
- },
7
- "config_map": {
8
- "9076f36a74755ac4": {
9
- "adapter_residual_before_ln": false,
10
- "cross_adapter": false,
11
- "dropout": 0.0,
12
- "factorized_phm_W": true,
13
- "factorized_phm_rule": false,
14
- "hypercomplex_nonlinearity": "glorot-uniform",
15
- "init_weights": "bert",
16
- "inv_adapter": null,
17
- "inv_adapter_reduction_factor": null,
18
- "is_parallel": false,
19
- "learn_phm": true,
20
- "leave_out": [],
21
- "ln_after": false,
22
- "ln_before": false,
23
- "mh_adapter": false,
24
- "non_linearity": "relu",
25
- "original_ln_after": true,
26
- "original_ln_before": true,
27
- "output_adapter": true,
28
- "phm_bias": true,
29
- "phm_c_init": "normal",
30
- "phm_dim": 4,
31
- "phm_init_range": 0.0001,
32
- "phm_layer": false,
33
- "phm_rank": 1,
34
- "reduction_factor": 16,
35
- "residual_before_ln": true,
36
- "scaling": 1.0,
37
- "shared_W_phm": false,
38
- "shared_phm_rule": true,
39
- "stochastic_depth": 0.0,
40
- "use_gating": false
41
- }
42
- },
43
- "fusion_config_map": {},
44
- "fusion_name_map": {},
45
- "fusions": {}
46
- },
47
  "architectures": [
48
- "T5AdapterModel"
49
  ],
50
  "classifier_dropout": 0.0,
51
  "d_ff": 2048,
@@ -56,11 +12,9 @@
56
  "dropout_rate": 0.1,
57
  "eos_token_id": 1,
58
  "feed_forward_proj": "relu",
59
- "id2label": null,
60
  "initializer_factor": 1.0,
61
  "is_encoder_decoder": true,
62
  "is_gated_act": false,
63
- "label2id": null,
64
  "layer_norm_epsilon": 1e-06,
65
  "model_type": "t5",
66
  "n_positions": 512,
@@ -69,28 +23,6 @@
69
  "num_layers": 6,
70
  "output_past": true,
71
  "pad_token_id": 0,
72
- "prediction_heads": {
73
- "default": {
74
- "activation_function": null,
75
- "bias": false,
76
- "head_type": "seq2seq_lm",
77
- "label2id": null,
78
- "layer_norm": false,
79
- "layers": 1,
80
- "shift_labels": false,
81
- "vocab_size": 32128
82
- },
83
- "imdb_pfeiffer": {
84
- "activation_function": null,
85
- "bias": false,
86
- "head_type": "seq2seq_lm",
87
- "label2id": null,
88
- "layer_norm": false,
89
- "layers": 1,
90
- "shift_labels": false,
91
- "vocab_size": 32128
92
- }
93
- },
94
  "relative_attention_max_distance": 128,
95
  "relative_attention_num_buckets": 32,
96
  "task_specific_params": {
 
1
  {
2
  "_name_or_path": "t5-small",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  "architectures": [
4
+ "T5ForConditionalGeneration"
5
  ],
6
  "classifier_dropout": 0.0,
7
  "d_ff": 2048,
 
12
  "dropout_rate": 0.1,
13
  "eos_token_id": 1,
14
  "feed_forward_proj": "relu",
 
15
  "initializer_factor": 1.0,
16
  "is_encoder_decoder": true,
17
  "is_gated_act": false,
 
18
  "layer_norm_epsilon": 1e-06,
19
  "model_type": "t5",
20
  "n_positions": 512,
 
23
  "num_layers": 6,
24
  "output_past": true,
25
  "pad_token_id": 0,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  "relative_attention_max_distance": 128,
27
  "relative_attention_num_buckets": 32,
28
  "task_specific_params": {
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7e40f01a6989af3dbd8712678119ae884e164d8a136116cdec86327977573f8f
3
- size 243649520
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5deb8eaaf50dcd203f4c4b710030eb2ad7895a09ff0cf0ae55bb0617751e020c
3
+ size 243231448
runs/Apr18_10-28-36_448e4276cffe/events.out.tfevents.1744972755.448e4276cffe.40881.10 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:960a678826707a000c8e35e69827198e1fd3af6babf38a1c4900ae362133db3d
3
+ size 405
runs/Apr18_11-09-55_448e4276cffe/events.out.tfevents.1744974597.448e4276cffe.40881.11 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7e39dfad3b1e20733a403b69cca9492f8bc6df1b6e45be1b3610a697d4abeef
3
+ size 6212
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7583de184e2beb0dc73eedd63571ec8daafb059f31aff0e63af2f61b42b1cc7f
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ca501d2ba86ccc9a09ec9d382dddcf4c54f22483c66f9961156eb3af0262aa4
3
  size 5368