dima806 commited on
Commit
e911d41
1 Parent(s): 93f2eb7

Upload . with huggingface_hub

Browse files
checkpoint-2129/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-cased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "id2label": {
12
+ "0": "NOT ENGAGING TITLE",
13
+ "1": "ENGAGING TITLE"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "max_position_embeddings": 512,
17
+ "model_type": "distilbert",
18
+ "n_heads": 12,
19
+ "n_layers": 6,
20
+ "output_past": true,
21
+ "pad_token_id": 0,
22
+ "problem_type": "single_label_classification",
23
+ "qa_dropout": 0.1,
24
+ "seq_classif_dropout": 0.2,
25
+ "sinusoidal_pos_embds": false,
26
+ "tie_weights_": true,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.28.1",
29
+ "vocab_size": 28996
30
+ }
checkpoint-2129/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12661c1b807a71f76496315e6c8c2418a04937bb4977a83f7fe257954e4624c4
3
+ size 526325317
checkpoint-2129/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4620318968f84e74e73d471a46d932c4ca087e46e6d28a3420b8b61ec7b0729
3
+ size 263167661
checkpoint-2129/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5adb7e529eb113cdd52ec4ee8c4d96b2d8ff061fadcb01323dab8b224cefc3f4
3
+ size 14575
checkpoint-2129/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:923a8812e6482c25919577d51110e470174396622d73814cd19f46a83bb83da1
3
+ size 627
checkpoint-2129/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-2129/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61144ad5a14db849074143ef18b12aaa5ca89078c4d68f0089580887e1baf58c
3
+ size 3579
config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-cased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "id2label": {
12
+ "0": "NOT ENGAGING TITLE",
13
+ "1": "ENGAGING TITLE"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "max_position_embeddings": 512,
17
+ "model_type": "distilbert",
18
+ "n_heads": 12,
19
+ "n_layers": 6,
20
+ "output_past": true,
21
+ "pad_token_id": 0,
22
+ "problem_type": "single_label_classification",
23
+ "qa_dropout": 0.1,
24
+ "seq_classif_dropout": 0.2,
25
+ "sinusoidal_pos_embds": false,
26
+ "tie_weights_": true,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.28.1",
29
+ "vocab_size": 28996
30
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4620318968f84e74e73d471a46d932c4ca087e46e6d28a3420b8b61ec7b0729
3
+ size 263167661
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61144ad5a14db849074143ef18b12aaa5ca89078c4d68f0089580887e1baf58c
3
+ size 3579
vocab.txt ADDED
The diff for this file is too large to render. See raw diff