dima806 commited on
Commit
1063c31
1 Parent(s): 2a7fc52

Upload . with huggingface_hub

Browse files
checkpoint-11196/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-cased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "id2label": {
12
+ "0": "NOT ENGAGING TITLE",
13
+ "1": "ENGAGING TITLE"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "max_position_embeddings": 512,
17
+ "model_type": "distilbert",
18
+ "n_heads": 12,
19
+ "n_layers": 6,
20
+ "output_past": true,
21
+ "pad_token_id": 0,
22
+ "problem_type": "single_label_classification",
23
+ "qa_dropout": 0.1,
24
+ "seq_classif_dropout": 0.2,
25
+ "sinusoidal_pos_embds": false,
26
+ "tie_weights_": true,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.28.1",
29
+ "vocab_size": 28996
30
+ }
checkpoint-11196/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0c9f48b1a4e4608d360fef74f46f75b0aa9c1408f659624500761363011418f
3
+ size 526325317
checkpoint-11196/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf34c91b711fe89de62c190bda74c8a3981edb0d601f9828baa229dfaeb6fdad
3
+ size 263167661
checkpoint-11196/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91056ce5e13444a0d1ca76aa0cc76709fadb8a2fdc9ea7369d3cd17b2f64ea07
3
+ size 14575
checkpoint-11196/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ebc0177eaee28d14b65284376d2a406bc9bd65295d6134e2508db48974221823
3
+ size 627
checkpoint-11196/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-11196/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35e9afce1e3c79b8e48cf5f8ebfabc8d8c664c71a26a1f06b3d10a19ff9b5ca5
3
+ size 3579
checkpoint-16794/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-cased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "id2label": {
12
+ "0": "NOT ENGAGING TITLE",
13
+ "1": "ENGAGING TITLE"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "max_position_embeddings": 512,
17
+ "model_type": "distilbert",
18
+ "n_heads": 12,
19
+ "n_layers": 6,
20
+ "output_past": true,
21
+ "pad_token_id": 0,
22
+ "problem_type": "single_label_classification",
23
+ "qa_dropout": 0.1,
24
+ "seq_classif_dropout": 0.2,
25
+ "sinusoidal_pos_embds": false,
26
+ "tie_weights_": true,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.28.1",
29
+ "vocab_size": 28996
30
+ }
checkpoint-16794/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0e68772d5c1c3369e3179ee552bfb96b61820028127d33d1a00566c68bb44b6
3
+ size 526325317
checkpoint-16794/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fec2c4521bc2433939d2191f771616cd5ca4aff4b5cbd6bdbe7ea451e9935735
3
+ size 263167661
checkpoint-16794/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a7af44f20dc082b05521639914a4b98a4bf731de83e1df333785b547a5b48ee
3
+ size 14575
checkpoint-16794/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b140457cb27dc42f3682b45f7560a4ae790a8d62991205d65bd3e7085f5e11b1
3
+ size 627
checkpoint-16794/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-16794/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35e9afce1e3c79b8e48cf5f8ebfabc8d8c664c71a26a1f06b3d10a19ff9b5ca5
3
+ size 3579
checkpoint-22392/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-cased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "id2label": {
12
+ "0": "NOT ENGAGING TITLE",
13
+ "1": "ENGAGING TITLE"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "max_position_embeddings": 512,
17
+ "model_type": "distilbert",
18
+ "n_heads": 12,
19
+ "n_layers": 6,
20
+ "output_past": true,
21
+ "pad_token_id": 0,
22
+ "problem_type": "single_label_classification",
23
+ "qa_dropout": 0.1,
24
+ "seq_classif_dropout": 0.2,
25
+ "sinusoidal_pos_embds": false,
26
+ "tie_weights_": true,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.28.1",
29
+ "vocab_size": 28996
30
+ }
checkpoint-22392/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abe86fe80b81495adc1eb44268d2f8bded373846fc374d274885e33ac61415d2
3
+ size 526325317
checkpoint-22392/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e9464ce73a956f3b2a081ee659115a564d9267962e841e5529f2827e377c422
3
+ size 263167661
checkpoint-22392/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02d319aa65c310a0bc008fe46e294cb2a48d1aa9f59cb9c47df8988910c17a97
3
+ size 14575
checkpoint-22392/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e6d609329b46546905ca2d48fa798d02fa521a392d8c7de1263c0af435ccb9d
3
+ size 627
checkpoint-22392/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-22392/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35e9afce1e3c79b8e48cf5f8ebfabc8d8c664c71a26a1f06b3d10a19ff9b5ca5
3
+ size 3579
checkpoint-27990/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-cased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "id2label": {
12
+ "0": "NOT ENGAGING TITLE",
13
+ "1": "ENGAGING TITLE"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "max_position_embeddings": 512,
17
+ "model_type": "distilbert",
18
+ "n_heads": 12,
19
+ "n_layers": 6,
20
+ "output_past": true,
21
+ "pad_token_id": 0,
22
+ "problem_type": "single_label_classification",
23
+ "qa_dropout": 0.1,
24
+ "seq_classif_dropout": 0.2,
25
+ "sinusoidal_pos_embds": false,
26
+ "tie_weights_": true,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.28.1",
29
+ "vocab_size": 28996
30
+ }
checkpoint-27990/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33e7d1c1d8f1b2e5065f0fb635e833876a4b806a1e2b04d8d249f6bcad0099bd
3
+ size 526325317
checkpoint-27990/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc3f218cc7fe17f99cff2073e6775659bdb6527bfc03dd0726de66bd8c418e3c
3
+ size 263167661
checkpoint-27990/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37fe28ae80f2d632b6b717bd04c1b2493dbdd3fc0d70fd0703774c7a1b08e7af
3
+ size 14575
checkpoint-27990/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a5822604a4c55b529289283e2c1b6e955971da3151777903b5c85cf5addd934
3
+ size 627
checkpoint-27990/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-27990/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35e9afce1e3c79b8e48cf5f8ebfabc8d8c664c71a26a1f06b3d10a19ff9b5ca5
3
+ size 3579
checkpoint-5598/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-cased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "id2label": {
12
+ "0": "NOT ENGAGING TITLE",
13
+ "1": "ENGAGING TITLE"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "max_position_embeddings": 512,
17
+ "model_type": "distilbert",
18
+ "n_heads": 12,
19
+ "n_layers": 6,
20
+ "output_past": true,
21
+ "pad_token_id": 0,
22
+ "problem_type": "single_label_classification",
23
+ "qa_dropout": 0.1,
24
+ "seq_classif_dropout": 0.2,
25
+ "sinusoidal_pos_embds": false,
26
+ "tie_weights_": true,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.28.1",
29
+ "vocab_size": 28996
30
+ }
checkpoint-5598/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87b8d7458577e71c061d093ff84e0b8021d170a326badb2953211da1bfb84ca8
3
+ size 526325317
checkpoint-5598/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cced252aba8e5e27ca69f2fdf77fdbddeed3045556bec4c2f25214addc61007
3
+ size 263167661
checkpoint-5598/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69e1d6b6beab1f972a4809e3bb3b779c7547f78b224abe50015f44b6b6b529b2
3
+ size 14575
checkpoint-5598/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c3b4eba5f14146ea51d5bd86209eae0316ecd2b59d049f3c804b2e99cc8e066
3
+ size 627
checkpoint-5598/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-5598/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35e9afce1e3c79b8e48cf5f8ebfabc8d8c664c71a26a1f06b3d10a19ff9b5ca5
3
+ size 3579
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b372977d6ab72af91850d33e0b08944fd024457512c01fcd811870bde650d898
3
  size 263167661
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fec2c4521bc2433939d2191f771616cd5ca4aff4b5cbd6bdbe7ea451e9935735
3
  size 263167661
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b1e3143372676e416a3afc8577502b2fd1d7e0d0a66edcca2d99681e9727436b
3
  size 3579
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35e9afce1e3c79b8e48cf5f8ebfabc8d8c664c71a26a1f06b3d10a19ff9b5ca5
3
  size 3579