Training in progress, step 10
Browse files- config.json +21 -0
- model.safetensors +3 -0
- runs/Oct21_18-14-39_torch-flash-large-0-5/events.out.tfevents.1729534479.torch-flash-large-0-5.1138.0 +3 -0
- runs/Oct21_18-19-47_torch-flash-large-0-5/events.out.tfevents.1729534796.torch-flash-large-0-5.1279.0 +3 -0
- runs/Oct21_18-19-47_torch-flash-large-0-5/events.out.tfevents.1729534993.torch-flash-large-0-5.1279.1 +3 -0
- runs/Oct21_18-26-28_torch-flash-large-0-5/events.out.tfevents.1729535188.torch-flash-large-0-5.1486.0 +3 -0
- runs/Oct21_18-27-11_torch-flash-large-0-5/events.out.tfevents.1729535232.torch-flash-large-0-5.1630.0 +3 -0
- runs/Oct21_18-29-01_torch-flash-large-0-5/events.out.tfevents.1729535341.torch-flash-large-0-5.1771.0 +3 -0
- runs/Oct21_18-29-22_torch-flash-large-0-5/events.out.tfevents.1729535362.torch-flash-large-0-5.1912.0 +3 -0
- runs/Oct21_18-30-28_torch-flash-large-0-5/events.out.tfevents.1729535429.torch-flash-large-0-5.2277.0 +3 -0
- runs/Oct21_18-32-23_torch-flash-large-0-5/events.out.tfevents.1729535544.torch-flash-large-0-5.2502.0 +3 -0
- runs/Oct21_18-32-51_torch-flash-large-0-5/events.out.tfevents.1729535572.torch-flash-large-0-5.2744.0 +3 -0
- runs/Oct21_18-39-13_torch-flash-large-0-5/events.out.tfevents.1729535955.torch-flash-large-0-5.3295.0 +3 -0
- runs/Oct21_18-40-54_torch-flash-large-0-5/events.out.tfevents.1729536058.torch-flash-large-0-5.3859.0 +3 -0
- training_args.bin +3 -0
config.json
ADDED
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "tattabio/gLM2_650M",
|
3 |
+
"architectures": [
|
4 |
+
"gLM2ForMaskedLM"
|
5 |
+
],
|
6 |
+
"auto_map": {
|
7 |
+
"AutoConfig": "configuration_glm2.gLM2Config",
|
8 |
+
"AutoModel": "modeling_glm2.gLM2Model",
|
9 |
+
"AutoModelForMaskedLM": "modeling_glm2.gLM2ForMaskedLM"
|
10 |
+
},
|
11 |
+
"depth": 33,
|
12 |
+
"dim": 1280,
|
13 |
+
"ffn_dim_multiplier": null,
|
14 |
+
"heads": 20,
|
15 |
+
"model_type": "gLM2",
|
16 |
+
"norm_eps": 1e-05,
|
17 |
+
"swiglu_multiple_of": 256,
|
18 |
+
"torch_dtype": "float32",
|
19 |
+
"transformers_version": "4.45.2",
|
20 |
+
"vocab_size": 37
|
21 |
+
}
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b9cd9d352414fa8c4ffe85d9f6f78b782b292d0166e609387da5b2e30e8e2863
|
3 |
+
size 2682482800
|
runs/Oct21_18-14-39_torch-flash-large-0-5/events.out.tfevents.1729534479.torch-flash-large-0-5.1138.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5c75778e00f22ad171e4c50670f883557f21d3f7a7ea471fd74d55125d80f838
|
3 |
+
size 4729
|
runs/Oct21_18-19-47_torch-flash-large-0-5/events.out.tfevents.1729534796.torch-flash-large-0-5.1279.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3400e4f2b50279cfd8953d46f340a1cfbba82196854f61d70e5585efc374e8fb
|
3 |
+
size 9418
|
runs/Oct21_18-19-47_torch-flash-large-0-5/events.out.tfevents.1729534993.torch-flash-large-0-5.1279.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:654c3f4d7cd1a791855bed6b03a6faf82a7545f3bf11f898c434de6619fe79f1
|
3 |
+
size 4729
|
runs/Oct21_18-26-28_torch-flash-large-0-5/events.out.tfevents.1729535188.torch-flash-large-0-5.1486.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:88b00e3dc528e2acd6e50d583bf71b6dfc170809e3f0c28c0ba866b49a6c51d5
|
3 |
+
size 4729
|
runs/Oct21_18-27-11_torch-flash-large-0-5/events.out.tfevents.1729535232.torch-flash-large-0-5.1630.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:03dafc7647842a16c03e8bdd13a2d1b905523dbbde16a6f165a1db6f1eb41c6f
|
3 |
+
size 4729
|
runs/Oct21_18-29-01_torch-flash-large-0-5/events.out.tfevents.1729535341.torch-flash-large-0-5.1771.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:680cb3e1319745501042848ba8b7c64825dc2f97893315d357a32b11535d0fba
|
3 |
+
size 4729
|
runs/Oct21_18-29-22_torch-flash-large-0-5/events.out.tfevents.1729535362.torch-flash-large-0-5.1912.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:778f3c68af5bc214d53bb18fc4d15ad0c6a7ceb802a3fb6b9f554b3eaf8d547d
|
3 |
+
size 4729
|
runs/Oct21_18-30-28_torch-flash-large-0-5/events.out.tfevents.1729535429.torch-flash-large-0-5.2277.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ccc08256e563bf662ff17a3f7219924aae3db241c94f5202ba8c606669af325f
|
3 |
+
size 4729
|
runs/Oct21_18-32-23_torch-flash-large-0-5/events.out.tfevents.1729535544.torch-flash-large-0-5.2502.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7094220ea702a01c4f2ef23e5feb553ec041c78177c5d58f6ec6ac06a587b496
|
3 |
+
size 4729
|
runs/Oct21_18-32-51_torch-flash-large-0-5/events.out.tfevents.1729535572.torch-flash-large-0-5.2744.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e4327600f4bc930f38eb26498c72b88c8ffef9616adf0037c6c96cf781f10333
|
3 |
+
size 4729
|
runs/Oct21_18-39-13_torch-flash-large-0-5/events.out.tfevents.1729535955.torch-flash-large-0-5.3295.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:89dab5ce6340fdd804c1dca0eb4f6b29c722aa800d3f89b1a3e51f3ae2cd91f6
|
3 |
+
size 4729
|
runs/Oct21_18-40-54_torch-flash-large-0-5/events.out.tfevents.1729536058.torch-flash-large-0-5.3859.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bbd0c2b6f4e727a0785d6219bfebd735b639754f77f340267493f9bc73d38b87
|
3 |
+
size 5076
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3e01b400000af8ed88c37cdab915c988efe8d45bf123e24ed5ab99bdd9558353
|
3 |
+
size 5240
|