Upload . with huggingface_hub
Browse files- checkpoint-11196/config.json +30 -0
- checkpoint-11196/optimizer.pt +3 -0
- checkpoint-11196/pytorch_model.bin +3 -0
- checkpoint-11196/rng_state.pth +3 -0
- checkpoint-11196/scheduler.pt +3 -0
- checkpoint-11196/trainer_state.json +0 -0
- checkpoint-11196/training_args.bin +3 -0
- checkpoint-16794/config.json +30 -0
- checkpoint-16794/optimizer.pt +3 -0
- checkpoint-16794/pytorch_model.bin +3 -0
- checkpoint-16794/rng_state.pth +3 -0
- checkpoint-16794/scheduler.pt +3 -0
- checkpoint-16794/trainer_state.json +0 -0
- checkpoint-16794/training_args.bin +3 -0
- checkpoint-22392/config.json +30 -0
- checkpoint-22392/optimizer.pt +3 -0
- checkpoint-22392/pytorch_model.bin +3 -0
- checkpoint-22392/rng_state.pth +3 -0
- checkpoint-22392/scheduler.pt +3 -0
- checkpoint-22392/trainer_state.json +0 -0
- checkpoint-22392/training_args.bin +3 -0
- checkpoint-27990/config.json +30 -0
- checkpoint-27990/optimizer.pt +3 -0
- checkpoint-27990/pytorch_model.bin +3 -0
- checkpoint-27990/rng_state.pth +3 -0
- checkpoint-27990/scheduler.pt +3 -0
- checkpoint-27990/trainer_state.json +0 -0
- checkpoint-27990/training_args.bin +3 -0
- checkpoint-5598/config.json +30 -0
- checkpoint-5598/optimizer.pt +3 -0
- checkpoint-5598/pytorch_model.bin +3 -0
- checkpoint-5598/rng_state.pth +3 -0
- checkpoint-5598/scheduler.pt +3 -0
- checkpoint-5598/trainer_state.json +0 -0
- checkpoint-5598/training_args.bin +3 -0
- pytorch_model.bin +1 -1
- training_args.bin +1 -1
checkpoint-11196/config.json
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "distilbert-base-cased",
|
3 |
+
"activation": "gelu",
|
4 |
+
"architectures": [
|
5 |
+
"DistilBertForSequenceClassification"
|
6 |
+
],
|
7 |
+
"attention_dropout": 0.1,
|
8 |
+
"dim": 768,
|
9 |
+
"dropout": 0.1,
|
10 |
+
"hidden_dim": 3072,
|
11 |
+
"id2label": {
|
12 |
+
"0": "NOT ENGAGING TITLE",
|
13 |
+
"1": "ENGAGING TITLE"
|
14 |
+
},
|
15 |
+
"initializer_range": 0.02,
|
16 |
+
"max_position_embeddings": 512,
|
17 |
+
"model_type": "distilbert",
|
18 |
+
"n_heads": 12,
|
19 |
+
"n_layers": 6,
|
20 |
+
"output_past": true,
|
21 |
+
"pad_token_id": 0,
|
22 |
+
"problem_type": "single_label_classification",
|
23 |
+
"qa_dropout": 0.1,
|
24 |
+
"seq_classif_dropout": 0.2,
|
25 |
+
"sinusoidal_pos_embds": false,
|
26 |
+
"tie_weights_": true,
|
27 |
+
"torch_dtype": "float32",
|
28 |
+
"transformers_version": "4.28.1",
|
29 |
+
"vocab_size": 28996
|
30 |
+
}
|
checkpoint-11196/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c0c9f48b1a4e4608d360fef74f46f75b0aa9c1408f659624500761363011418f
|
3 |
+
size 526325317
|
checkpoint-11196/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cf34c91b711fe89de62c190bda74c8a3981edb0d601f9828baa229dfaeb6fdad
|
3 |
+
size 263167661
|
checkpoint-11196/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:91056ce5e13444a0d1ca76aa0cc76709fadb8a2fdc9ea7369d3cd17b2f64ea07
|
3 |
+
size 14575
|
checkpoint-11196/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ebc0177eaee28d14b65284376d2a406bc9bd65295d6134e2508db48974221823
|
3 |
+
size 627
|
checkpoint-11196/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-11196/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:35e9afce1e3c79b8e48cf5f8ebfabc8d8c664c71a26a1f06b3d10a19ff9b5ca5
|
3 |
+
size 3579
|
checkpoint-16794/config.json
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "distilbert-base-cased",
|
3 |
+
"activation": "gelu",
|
4 |
+
"architectures": [
|
5 |
+
"DistilBertForSequenceClassification"
|
6 |
+
],
|
7 |
+
"attention_dropout": 0.1,
|
8 |
+
"dim": 768,
|
9 |
+
"dropout": 0.1,
|
10 |
+
"hidden_dim": 3072,
|
11 |
+
"id2label": {
|
12 |
+
"0": "NOT ENGAGING TITLE",
|
13 |
+
"1": "ENGAGING TITLE"
|
14 |
+
},
|
15 |
+
"initializer_range": 0.02,
|
16 |
+
"max_position_embeddings": 512,
|
17 |
+
"model_type": "distilbert",
|
18 |
+
"n_heads": 12,
|
19 |
+
"n_layers": 6,
|
20 |
+
"output_past": true,
|
21 |
+
"pad_token_id": 0,
|
22 |
+
"problem_type": "single_label_classification",
|
23 |
+
"qa_dropout": 0.1,
|
24 |
+
"seq_classif_dropout": 0.2,
|
25 |
+
"sinusoidal_pos_embds": false,
|
26 |
+
"tie_weights_": true,
|
27 |
+
"torch_dtype": "float32",
|
28 |
+
"transformers_version": "4.28.1",
|
29 |
+
"vocab_size": 28996
|
30 |
+
}
|
checkpoint-16794/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f0e68772d5c1c3369e3179ee552bfb96b61820028127d33d1a00566c68bb44b6
|
3 |
+
size 526325317
|
checkpoint-16794/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fec2c4521bc2433939d2191f771616cd5ca4aff4b5cbd6bdbe7ea451e9935735
|
3 |
+
size 263167661
|
checkpoint-16794/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6a7af44f20dc082b05521639914a4b98a4bf731de83e1df333785b547a5b48ee
|
3 |
+
size 14575
|
checkpoint-16794/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b140457cb27dc42f3682b45f7560a4ae790a8d62991205d65bd3e7085f5e11b1
|
3 |
+
size 627
|
checkpoint-16794/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-16794/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:35e9afce1e3c79b8e48cf5f8ebfabc8d8c664c71a26a1f06b3d10a19ff9b5ca5
|
3 |
+
size 3579
|
checkpoint-22392/config.json
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "distilbert-base-cased",
|
3 |
+
"activation": "gelu",
|
4 |
+
"architectures": [
|
5 |
+
"DistilBertForSequenceClassification"
|
6 |
+
],
|
7 |
+
"attention_dropout": 0.1,
|
8 |
+
"dim": 768,
|
9 |
+
"dropout": 0.1,
|
10 |
+
"hidden_dim": 3072,
|
11 |
+
"id2label": {
|
12 |
+
"0": "NOT ENGAGING TITLE",
|
13 |
+
"1": "ENGAGING TITLE"
|
14 |
+
},
|
15 |
+
"initializer_range": 0.02,
|
16 |
+
"max_position_embeddings": 512,
|
17 |
+
"model_type": "distilbert",
|
18 |
+
"n_heads": 12,
|
19 |
+
"n_layers": 6,
|
20 |
+
"output_past": true,
|
21 |
+
"pad_token_id": 0,
|
22 |
+
"problem_type": "single_label_classification",
|
23 |
+
"qa_dropout": 0.1,
|
24 |
+
"seq_classif_dropout": 0.2,
|
25 |
+
"sinusoidal_pos_embds": false,
|
26 |
+
"tie_weights_": true,
|
27 |
+
"torch_dtype": "float32",
|
28 |
+
"transformers_version": "4.28.1",
|
29 |
+
"vocab_size": 28996
|
30 |
+
}
|
checkpoint-22392/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:abe86fe80b81495adc1eb44268d2f8bded373846fc374d274885e33ac61415d2
|
3 |
+
size 526325317
|
checkpoint-22392/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8e9464ce73a956f3b2a081ee659115a564d9267962e841e5529f2827e377c422
|
3 |
+
size 263167661
|
checkpoint-22392/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:02d319aa65c310a0bc008fe46e294cb2a48d1aa9f59cb9c47df8988910c17a97
|
3 |
+
size 14575
|
checkpoint-22392/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9e6d609329b46546905ca2d48fa798d02fa521a392d8c7de1263c0af435ccb9d
|
3 |
+
size 627
|
checkpoint-22392/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-22392/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:35e9afce1e3c79b8e48cf5f8ebfabc8d8c664c71a26a1f06b3d10a19ff9b5ca5
|
3 |
+
size 3579
|
checkpoint-27990/config.json
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "distilbert-base-cased",
|
3 |
+
"activation": "gelu",
|
4 |
+
"architectures": [
|
5 |
+
"DistilBertForSequenceClassification"
|
6 |
+
],
|
7 |
+
"attention_dropout": 0.1,
|
8 |
+
"dim": 768,
|
9 |
+
"dropout": 0.1,
|
10 |
+
"hidden_dim": 3072,
|
11 |
+
"id2label": {
|
12 |
+
"0": "NOT ENGAGING TITLE",
|
13 |
+
"1": "ENGAGING TITLE"
|
14 |
+
},
|
15 |
+
"initializer_range": 0.02,
|
16 |
+
"max_position_embeddings": 512,
|
17 |
+
"model_type": "distilbert",
|
18 |
+
"n_heads": 12,
|
19 |
+
"n_layers": 6,
|
20 |
+
"output_past": true,
|
21 |
+
"pad_token_id": 0,
|
22 |
+
"problem_type": "single_label_classification",
|
23 |
+
"qa_dropout": 0.1,
|
24 |
+
"seq_classif_dropout": 0.2,
|
25 |
+
"sinusoidal_pos_embds": false,
|
26 |
+
"tie_weights_": true,
|
27 |
+
"torch_dtype": "float32",
|
28 |
+
"transformers_version": "4.28.1",
|
29 |
+
"vocab_size": 28996
|
30 |
+
}
|
checkpoint-27990/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:33e7d1c1d8f1b2e5065f0fb635e833876a4b806a1e2b04d8d249f6bcad0099bd
|
3 |
+
size 526325317
|
checkpoint-27990/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dc3f218cc7fe17f99cff2073e6775659bdb6527bfc03dd0726de66bd8c418e3c
|
3 |
+
size 263167661
|
checkpoint-27990/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:37fe28ae80f2d632b6b717bd04c1b2493dbdd3fc0d70fd0703774c7a1b08e7af
|
3 |
+
size 14575
|
checkpoint-27990/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3a5822604a4c55b529289283e2c1b6e955971da3151777903b5c85cf5addd934
|
3 |
+
size 627
|
checkpoint-27990/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-27990/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:35e9afce1e3c79b8e48cf5f8ebfabc8d8c664c71a26a1f06b3d10a19ff9b5ca5
|
3 |
+
size 3579
|
checkpoint-5598/config.json
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "distilbert-base-cased",
|
3 |
+
"activation": "gelu",
|
4 |
+
"architectures": [
|
5 |
+
"DistilBertForSequenceClassification"
|
6 |
+
],
|
7 |
+
"attention_dropout": 0.1,
|
8 |
+
"dim": 768,
|
9 |
+
"dropout": 0.1,
|
10 |
+
"hidden_dim": 3072,
|
11 |
+
"id2label": {
|
12 |
+
"0": "NOT ENGAGING TITLE",
|
13 |
+
"1": "ENGAGING TITLE"
|
14 |
+
},
|
15 |
+
"initializer_range": 0.02,
|
16 |
+
"max_position_embeddings": 512,
|
17 |
+
"model_type": "distilbert",
|
18 |
+
"n_heads": 12,
|
19 |
+
"n_layers": 6,
|
20 |
+
"output_past": true,
|
21 |
+
"pad_token_id": 0,
|
22 |
+
"problem_type": "single_label_classification",
|
23 |
+
"qa_dropout": 0.1,
|
24 |
+
"seq_classif_dropout": 0.2,
|
25 |
+
"sinusoidal_pos_embds": false,
|
26 |
+
"tie_weights_": true,
|
27 |
+
"torch_dtype": "float32",
|
28 |
+
"transformers_version": "4.28.1",
|
29 |
+
"vocab_size": 28996
|
30 |
+
}
|
checkpoint-5598/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:87b8d7458577e71c061d093ff84e0b8021d170a326badb2953211da1bfb84ca8
|
3 |
+
size 526325317
|
checkpoint-5598/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8cced252aba8e5e27ca69f2fdf77fdbddeed3045556bec4c2f25214addc61007
|
3 |
+
size 263167661
|
checkpoint-5598/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:69e1d6b6beab1f972a4809e3bb3b779c7547f78b224abe50015f44b6b6b529b2
|
3 |
+
size 14575
|
checkpoint-5598/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4c3b4eba5f14146ea51d5bd86209eae0316ecd2b59d049f3c804b2e99cc8e066
|
3 |
+
size 627
|
checkpoint-5598/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-5598/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:35e9afce1e3c79b8e48cf5f8ebfabc8d8c664c71a26a1f06b3d10a19ff9b5ca5
|
3 |
+
size 3579
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 263167661
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fec2c4521bc2433939d2191f771616cd5ca4aff4b5cbd6bdbe7ea451e9935735
|
3 |
size 263167661
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3579
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:35e9afce1e3c79b8e48cf5f8ebfabc8d8c664c71a26a1f06b3d10a19ff9b5ca5
|
3 |
size 3579
|