tristayqc commited on
Commit
d7870f8
·
verified ·
1 Parent(s): b158d1e

Training in progress, step 500

Browse files
config.json CHANGED
@@ -1,76 +1,34 @@
1
  {
2
- "_name_or_path": "facebook/bart-base",
3
- "activation_dropout": 0.1,
4
- "activation_function": "gelu",
5
- "add_bias_logits": false,
6
- "add_final_layer_norm": false,
7
  "architectures": [
8
- "BartForCausalLM"
9
  ],
10
- "attention_dropout": 0.1,
11
- "bos_token_id": 0,
12
- "classif_dropout": 0.1,
13
- "classifier_dropout": 0.0,
14
- "d_model": 768,
15
- "decoder_attention_heads": 12,
16
- "decoder_ffn_dim": 3072,
17
- "decoder_layerdrop": 0.0,
18
- "decoder_layers": 6,
19
- "decoder_start_token_id": 2,
20
- "dropout": 0.1,
21
- "early_stopping": true,
22
- "encoder_attention_heads": 12,
23
- "encoder_ffn_dim": 3072,
24
- "encoder_layerdrop": 0.0,
25
- "encoder_layers": 6,
26
  "eos_token_id": 2,
27
- "forced_bos_token_id": 0,
28
- "forced_eos_token_id": 2,
29
  "gradient_checkpointing": false,
30
- "id2label": {
31
- "0": "LABEL_0",
32
- "1": "LABEL_1",
33
- "2": "LABEL_2"
34
- },
35
- "init_std": 0.02,
36
- "is_decoder": true,
37
- "is_encoder_decoder": false,
38
- "label2id": {
39
- "LABEL_0": 0,
40
- "LABEL_1": 1,
41
- "LABEL_2": 2
42
- },
43
- "max_position_embeddings": 1024,
44
- "model_type": "bart",
45
- "no_repeat_ngram_size": 3,
46
- "normalize_before": false,
47
- "normalize_embedding": true,
48
- "num_beams": 4,
49
- "num_hidden_layers": 6,
50
- "pad_token_id": 1,
51
- "scale_embedding": false,
52
- "task_specific_params": {
53
- "summarization": {
54
- "length_penalty": 1.0,
55
- "max_length": 128,
56
- "min_length": 12,
57
- "num_beams": 4
58
- },
59
- "summarization_cnn": {
60
- "length_penalty": 2.0,
61
- "max_length": 142,
62
- "min_length": 56,
63
- "num_beams": 4
64
- },
65
- "summarization_xsum": {
66
- "length_penalty": 1.0,
67
- "max_length": 62,
68
- "min_length": 11,
69
- "num_beams": 6
70
- }
71
- },
72
  "torch_dtype": "float32",
73
  "transformers_version": "4.40.1",
 
 
74
  "use_cache": true,
75
- "vocab_size": 50265
76
  }
 
1
  {
2
+ "_name_or_path": "google/bigbird-roberta-base",
 
 
 
 
3
  "architectures": [
4
+ "BigBirdForCausalLM"
5
  ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "attention_type": "block_sparse",
8
+ "block_size": 64,
9
+ "bos_token_id": 1,
10
+ "classifier_dropout": null,
 
 
 
 
 
 
 
 
 
 
 
11
  "eos_token_id": 2,
 
 
12
  "gradient_checkpointing": false,
13
+ "hidden_act": "gelu_new",
14
+ "hidden_dropout_prob": 0.1,
15
+ "hidden_size": 768,
16
+ "initializer_range": 0.02,
17
+ "intermediate_size": 3072,
18
+ "layer_norm_eps": 1e-12,
19
+ "max_position_embeddings": 4096,
20
+ "model_type": "big_bird",
21
+ "num_attention_heads": 12,
22
+ "num_hidden_layers": 12,
23
+ "num_random_blocks": 3,
24
+ "pad_token_id": 0,
25
+ "position_embedding_type": "absolute",
26
+ "rescale_embeddings": false,
27
+ "sep_token_id": 66,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
  "torch_dtype": "float32",
29
  "transformers_version": "4.40.1",
30
+ "type_vocab_size": 2,
31
+ "use_bias": true,
32
  "use_cache": true,
33
+ "vocab_size": 50358
34
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:91a1c1f931cc7789d455a9636688f9f4ed43401b4a0284863d2c861df8faed9e
3
- size 384433400
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ddd5310a00fb766fedd32a0e2f7ce0f523d244ff36fdfdea5fa37cb54e16ca5
3
+ size 512469040
runs/Apr25_02-36-08_bf3541846459/events.out.tfevents.1714012569.bf3541846459.310.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ddae4937225a855175d62e771e1587b05364490354b09487a1254f5237b8ed83
3
+ size 5123
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9b061a21382b0709bc111d6528f426d5d136a239fa2c43af1476a2a81e5a0d5d
3
  size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:082c1c4670749193a8202d4efc048053bd772e623c1371b200a579b65d254a65
3
  size 4984