stas commited on
Commit
850afc3
·
1 Parent(s): 883eb1b
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. exp12-1/000-module.3.input_layernorm/events.out.tfevents.1638234485.r7i4n1.3380055.4 +3 -0
  2. exp12-1/000-module.3.mlp.dense_4h_to_h/events.out.tfevents.1638234487.r7i4n1.3380055.12 +3 -0
  3. exp12-1/000-module.3.mlp.dense_h_to_4h/events.out.tfevents.1638234486.r7i4n1.3380055.11 +3 -0
  4. exp12-1/000-module.3.mlp/events.out.tfevents.1638234487.r7i4n1.3380055.13 +3 -0
  5. exp12-1/000-module.3.post_attention_layernorm/events.out.tfevents.1638234486.r7i4n1.3380055.10 +3 -0
  6. exp12-1/000-module.3.self_attention.attention_dropout/events.out.tfevents.1638234486.r7i4n1.3380055.7 +3 -0
  7. exp12-1/000-module.3.self_attention.dense/events.out.tfevents.1638234486.r7i4n1.3380055.8 +3 -0
  8. exp12-1/000-module.3.self_attention.query_key_value/events.out.tfevents.1638234486.r7i4n1.3380055.5 +3 -0
  9. exp12-1/000-module.3.self_attention.scale_mask_softmax/events.out.tfevents.1638234486.r7i4n1.3380055.6 +3 -0
  10. exp12-1/000-module.3.self_attention/events.out.tfevents.1638234486.r7i4n1.3380055.9 +3 -0
  11. exp12-1/000-module.3/events.out.tfevents.1638234487.r7i4n1.3380055.14 +3 -0
  12. exp12-1/000-module.4.input_layernorm/events.out.tfevents.1638234487.r7i4n1.3380055.15 +3 -0
  13. exp12-1/000-module.4.mlp.dense_4h_to_h/events.out.tfevents.1638234487.r7i4n1.3380055.23 +3 -0
  14. exp12-1/000-module.4.mlp.dense_h_to_4h/events.out.tfevents.1638234487.r7i4n1.3380055.22 +3 -0
  15. exp12-1/000-module.4.mlp/events.out.tfevents.1638234487.r7i4n1.3380055.24 +3 -0
  16. exp12-1/000-module.4.post_attention_layernorm/events.out.tfevents.1638234487.r7i4n1.3380055.21 +3 -0
  17. exp12-1/000-module.4.self_attention.attention_dropout/events.out.tfevents.1638234487.r7i4n1.3380055.18 +3 -0
  18. exp12-1/000-module.4.self_attention.dense/events.out.tfevents.1638234487.r7i4n1.3380055.19 +3 -0
  19. exp12-1/000-module.4.self_attention.query_key_value/events.out.tfevents.1638234487.r7i4n1.3380055.16 +3 -0
  20. exp12-1/000-module.4.self_attention.scale_mask_softmax/events.out.tfevents.1638234487.r7i4n1.3380055.17 +3 -0
  21. exp12-1/000-module.4.self_attention/events.out.tfevents.1638234487.r7i4n1.3380055.20 +3 -0
  22. exp12-1/000-module.4/events.out.tfevents.1638234487.r7i4n1.3380055.25 +3 -0
  23. exp12-1/000-module.tied_modules.embed.embedding_dropout/events.out.tfevents.1638234485.r7i4n1.3380055.2 +3 -0
  24. exp12-1/000-module.tied_modules.embed.position_embeddings/events.out.tfevents.1638234485.r7i4n1.3380055.1 +3 -0
  25. exp12-1/000-module.tied_modules.embed.word_embeddings/events.out.tfevents.1638234485.r7i4n1.3380055.0 +3 -0
  26. exp12-1/000-module.tied_modules.embed/events.out.tfevents.1638234485.r7i4n1.3380055.3 +3 -0
  27. exp12-1/000-module/events.out.tfevents.1638234487.r7i4n1.3380055.26 +3 -0
  28. exp12-1/056-module.17.input_layernorm/events.out.tfevents.1638234497.r7i5n6.1000222.0 +3 -0
  29. exp12-1/056-module.17.mlp.dense_4h_to_h/events.out.tfevents.1638234498.r7i5n6.1000222.8 +3 -0
  30. exp12-1/056-module.17.mlp.dense_h_to_4h/events.out.tfevents.1638234498.r7i5n6.1000222.7 +3 -0
  31. exp12-1/056-module.17.mlp/events.out.tfevents.1638234498.r7i5n6.1000222.9 +3 -0
  32. exp12-1/056-module.17.post_attention_layernorm/events.out.tfevents.1638234498.r7i5n6.1000222.6 +3 -0
  33. exp12-1/056-module.17.self_attention.attention_dropout/events.out.tfevents.1638234498.r7i5n6.1000222.3 +3 -0
  34. exp12-1/056-module.17.self_attention.dense/events.out.tfevents.1638234498.r7i5n6.1000222.4 +3 -0
  35. exp12-1/056-module.17.self_attention.query_key_value/events.out.tfevents.1638234498.r7i5n6.1000222.1 +3 -0
  36. exp12-1/056-module.17.self_attention.scale_mask_softmax/events.out.tfevents.1638234498.r7i5n6.1000222.2 +3 -0
  37. exp12-1/056-module.17.self_attention/events.out.tfevents.1638234498.r7i5n6.1000222.5 +3 -0
  38. exp12-1/056-module.17/events.out.tfevents.1638234498.r7i5n6.1000222.10 +3 -0
  39. exp12-1/056-module.18.input_layernorm/events.out.tfevents.1638234498.r7i5n6.1000222.11 +3 -0
  40. exp12-1/056-module.18.mlp.dense_4h_to_h/events.out.tfevents.1638234499.r7i5n6.1000222.19 +3 -0
  41. exp12-1/056-module.18.mlp.dense_h_to_4h/events.out.tfevents.1638234498.r7i5n6.1000222.18 +3 -0
  42. exp12-1/056-module.18.mlp/events.out.tfevents.1638234499.r7i5n6.1000222.20 +3 -0
  43. exp12-1/056-module.18.post_attention_layernorm/events.out.tfevents.1638234498.r7i5n6.1000222.17 +3 -0
  44. exp12-1/056-module.18.self_attention.attention_dropout/events.out.tfevents.1638234498.r7i5n6.1000222.14 +3 -0
  45. exp12-1/056-module.18.self_attention.dense/events.out.tfevents.1638234498.r7i5n6.1000222.15 +3 -0
  46. exp12-1/056-module.18.self_attention.query_key_value/events.out.tfevents.1638234498.r7i5n6.1000222.12 +3 -0
  47. exp12-1/056-module.18.self_attention.scale_mask_softmax/events.out.tfevents.1638234498.r7i5n6.1000222.13 +3 -0
  48. exp12-1/056-module.18.self_attention/events.out.tfevents.1638234498.r7i5n6.1000222.16 +3 -0
  49. exp12-1/056-module.18/events.out.tfevents.1638234499.r7i5n6.1000222.21 +3 -0
  50. exp12-1/056-module/events.out.tfevents.1638234499.r7i5n6.1000222.22 +3 -0
exp12-1/000-module.3.input_layernorm/events.out.tfevents.1638234485.r7i4n1.3380055.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee87864c04c11073676e441238e7c24a765692e8f4661f86e11854c10461ee01
3
+ size 83263528
exp12-1/000-module.3.mlp.dense_4h_to_h/events.out.tfevents.1638234487.r7i4n1.3380055.12 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66111d565d4814fe816fa6a0391478e7a782c7cbb8f13c80734901a3f576908c
3
+ size 84049960
exp12-1/000-module.3.mlp.dense_h_to_4h/events.out.tfevents.1638234486.r7i4n1.3380055.11 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b43ebf4a03af3110a8e1072d6d8ff630951710f9ccf653f4a30133795ae6d6dc
3
+ size 84049960
exp12-1/000-module.3.mlp/events.out.tfevents.1638234487.r7i4n1.3380055.13 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e3c1a4ebd4ffb2d3980810b9ce63db8bf342b9ec091afc35d1b42ad1a353b0e
3
+ size 49250344
exp12-1/000-module.3.post_attention_layernorm/events.out.tfevents.1638234486.r7i4n1.3380055.10 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b681938878c1fa4575190b3545f8c88cb7f32f1458527057435556f921508601
3
+ size 83263528
exp12-1/000-module.3.self_attention.attention_dropout/events.out.tfevents.1638234486.r7i4n1.3380055.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:480920f205e29c1a1bfe2b92bd2aeda3232ef893e3e510d2ae019bdc17c5ac23
3
+ size 36274216
exp12-1/000-module.3.self_attention.dense/events.out.tfevents.1638234486.r7i4n1.3380055.8 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f8a498a847171260878d0dd9aa375f547d6b6ef5ba2b817658804ab6c8eaa31
3
+ size 84049960
exp12-1/000-module.3.self_attention.query_key_value/events.out.tfevents.1638234486.r7i4n1.3380055.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f40aa024966e524b6aa18b531ec4605edb368d6c9251fab76c2405525d7b255e
3
+ size 77758504
exp12-1/000-module.3.self_attention.scale_mask_softmax/events.out.tfevents.1638234486.r7i4n1.3380055.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f30ef6c52fc4b1c866729c1302f54210f62bcd5a9a28472c1cfa651e5496379
3
+ size 36274216
exp12-1/000-module.3.self_attention/events.out.tfevents.1638234486.r7i4n1.3380055.9 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97a298c057c19570569b67ad6fba4f99314c2228b27a6cc7c00ec182108e41e4
3
+ size 49250344
exp12-1/000-module.3/events.out.tfevents.1638234487.r7i4n1.3380055.14 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:093eaec96287fd9d032a035c0c549a3c95ae76e5f1c6639b900b9ee38dd1414e
3
+ size 48463912
exp12-1/000-module.4.input_layernorm/events.out.tfevents.1638234487.r7i4n1.3380055.15 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2791b9b723387cbd1764e3a1cc46052cc8cd905391d94eea2b7b1fc1d65e82a9
3
+ size 83263528
exp12-1/000-module.4.mlp.dense_4h_to_h/events.out.tfevents.1638234487.r7i4n1.3380055.23 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcf2d161e9d8e50a48d99d860eee98782a062bec23262dfad7e9a75ffd6aeb25
3
+ size 84049960
exp12-1/000-module.4.mlp.dense_h_to_4h/events.out.tfevents.1638234487.r7i4n1.3380055.22 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03b38271ae083e54e7333ffc9c301681f68fc7503d587abf8465ce4a9b3b9b15
3
+ size 84049960
exp12-1/000-module.4.mlp/events.out.tfevents.1638234487.r7i4n1.3380055.24 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0322b8ea0339d2b9950eac30e7bcda37ac02d31193f099b08e2ed130ca9150ab
3
+ size 49250344
exp12-1/000-module.4.post_attention_layernorm/events.out.tfevents.1638234487.r7i4n1.3380055.21 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef94bfc9bd23bf5ed8ea138c5035d40a5e82f00365704f90ef6d28b782da478f
3
+ size 83263528
exp12-1/000-module.4.self_attention.attention_dropout/events.out.tfevents.1638234487.r7i4n1.3380055.18 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aedbed966768c86f59c163696cce6a196baacc34aad9c94006c65614604f1f2e
3
+ size 36274216
exp12-1/000-module.4.self_attention.dense/events.out.tfevents.1638234487.r7i4n1.3380055.19 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65c145079425dcabbadc5d844593881e68d9e649f0d33fb41102860f70e0fc53
3
+ size 84049960
exp12-1/000-module.4.self_attention.query_key_value/events.out.tfevents.1638234487.r7i4n1.3380055.16 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e116b99e30d53169e24a936d5bef1a00e5aacb91f993a708db9385ab7f55ef8
3
+ size 77758504
exp12-1/000-module.4.self_attention.scale_mask_softmax/events.out.tfevents.1638234487.r7i4n1.3380055.17 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18f6b836d34a03c8de6a71efc9c4e7ded856c11bf5be5ea8c095dee5d480329d
3
+ size 36274216
exp12-1/000-module.4.self_attention/events.out.tfevents.1638234487.r7i4n1.3380055.20 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c41df05b06b8143ec37f342f4e390c59aaaf2d2579e476a80af79caf96cd296b
3
+ size 49250344
exp12-1/000-module.4/events.out.tfevents.1638234487.r7i4n1.3380055.25 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63ee26fcc3e5d06e654bb60656d0eca9173381927a4b3cf39c1d1cbe128d2390
3
+ size 48463912
exp12-1/000-module.tied_modules.embed.embedding_dropout/events.out.tfevents.1638234485.r7i4n1.3380055.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e23cbbdbd5d5a1eb44fd03906e6c529ac297c048de3da02c502b1b233429886b
3
+ size 24281128
exp12-1/000-module.tied_modules.embed.position_embeddings/events.out.tfevents.1638234485.r7i4n1.3380055.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d92d19bbd8900e596d15d40507081362122d8edf5d4cff71e7a6c52240bf6e47
3
+ size 29982760
exp12-1/000-module.tied_modules.embed.word_embeddings/events.out.tfevents.1638234485.r7i4n1.3380055.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca1e9f74d5132f37cbc173a5f9330c5014230ad8d41609f92f504930030c6358
3
+ size 29982760
exp12-1/000-module.tied_modules.embed/events.out.tfevents.1638234485.r7i4n1.3380055.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:646f8ce33d68ef92c64b14532c1a077594f3ae304314706ed9aab8c73159a0c7
3
+ size 18186280
exp12-1/000-module/events.out.tfevents.1638234487.r7i4n1.3380055.26 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d98f711b037ea33b1970099f4a1adacce9be7b636104390970ad2bbfbf51bb0f
3
+ size 18186280
exp12-1/056-module.17.input_layernorm/events.out.tfevents.1638234497.r7i5n6.1000222.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5db5d187927d36ad3d1b6d557cf72d9f0deb44cbce4ab0b917d24d3c72ca8e29
3
+ size 83263528
exp12-1/056-module.17.mlp.dense_4h_to_h/events.out.tfevents.1638234498.r7i5n6.1000222.8 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3058901fb22e6e2fdf36709530b22533c66b3cd886204d94b30c0a400405b78
3
+ size 84049960
exp12-1/056-module.17.mlp.dense_h_to_4h/events.out.tfevents.1638234498.r7i5n6.1000222.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22033d2e60a37e4432340a0b7eecf18ecf7f56bf463273c57bece70c6f2a39b9
3
+ size 84049960
exp12-1/056-module.17.mlp/events.out.tfevents.1638234498.r7i5n6.1000222.9 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9005a093c50462c93c79d588cf7af7bd534f10f3b973ddc1ee8cdf653da3a2e
3
+ size 49250344
exp12-1/056-module.17.post_attention_layernorm/events.out.tfevents.1638234498.r7i5n6.1000222.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1dc35c6b58d42fc0add9aa2658fecd18888ccc4984500f702697d2baf4c29a8
3
+ size 83263528
exp12-1/056-module.17.self_attention.attention_dropout/events.out.tfevents.1638234498.r7i5n6.1000222.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5868efee5fec0255ea708f9a0398b1fd9044ad0ddc544ff1b6525b2cd7dc54b9
3
+ size 36274216
exp12-1/056-module.17.self_attention.dense/events.out.tfevents.1638234498.r7i5n6.1000222.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8022536a204bb933d39a81038f7e8f9c3590880499bc7d597c578ecfe43b7826
3
+ size 84049960
exp12-1/056-module.17.self_attention.query_key_value/events.out.tfevents.1638234498.r7i5n6.1000222.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4bec436b3b7b1e19601313878722a119b2a54df22af5ba6b334355ea8f4b51c6
3
+ size 77758504
exp12-1/056-module.17.self_attention.scale_mask_softmax/events.out.tfevents.1638234498.r7i5n6.1000222.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ead3f195588ab028415b7377726b7025f125c2e4c8293470433fcbfe8c4953f6
3
+ size 36274216
exp12-1/056-module.17.self_attention/events.out.tfevents.1638234498.r7i5n6.1000222.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8edce22155278265adb1e00d5c3664128319d6fafb95e36bf6bc6006d8af6335
3
+ size 49250344
exp12-1/056-module.17/events.out.tfevents.1638234498.r7i5n6.1000222.10 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:937af3adb4dc08d677c6a1156b9062c42838591bce6dd2d96efc3c096eb744ba
3
+ size 48463912
exp12-1/056-module.18.input_layernorm/events.out.tfevents.1638234498.r7i5n6.1000222.11 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c772fe64103ed937b5300491eaa9df8bb53a73ea5dc47aed198d9044abbb488f
3
+ size 83263528
exp12-1/056-module.18.mlp.dense_4h_to_h/events.out.tfevents.1638234499.r7i5n6.1000222.19 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e065dab3048bc8a48cfd19f775fc04567c733ddfbf02dd1ee23bc6010267a984
3
+ size 84049960
exp12-1/056-module.18.mlp.dense_h_to_4h/events.out.tfevents.1638234498.r7i5n6.1000222.18 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:465dbb54c21d319f7bc3c4db3a4483f3044be3879732682f2124d60f114c2e59
3
+ size 84049960
exp12-1/056-module.18.mlp/events.out.tfevents.1638234499.r7i5n6.1000222.20 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:512e73e5433df576708fc867ff5fae34d0d1eb00316cd428467ff3f833402306
3
+ size 49250344
exp12-1/056-module.18.post_attention_layernorm/events.out.tfevents.1638234498.r7i5n6.1000222.17 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06dc0faf1abfaef74eec666f85aeffa88c90f4d51cbae970f7f74c6cebc0d9ff
3
+ size 83263528
exp12-1/056-module.18.self_attention.attention_dropout/events.out.tfevents.1638234498.r7i5n6.1000222.14 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ddc62ce7ae93d747b7c225420c6fcab620353849be53e9c19dd8e4f7aa99cb8
3
+ size 36274216
exp12-1/056-module.18.self_attention.dense/events.out.tfevents.1638234498.r7i5n6.1000222.15 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f89afe50ae35544ff9e3c7a91a6b09d742187371f0cb41426eaed1f4276b5186
3
+ size 84049960
exp12-1/056-module.18.self_attention.query_key_value/events.out.tfevents.1638234498.r7i5n6.1000222.12 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a79782070ea0d313c1afb3f1cca5e154e3916da65fa255434c74989c2612dac4
3
+ size 77758504
exp12-1/056-module.18.self_attention.scale_mask_softmax/events.out.tfevents.1638234498.r7i5n6.1000222.13 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df00d64aaca737ece2c8e8070c499ec035864c4f3041d102334fdbb89242fe3c
3
+ size 36274216
exp12-1/056-module.18.self_attention/events.out.tfevents.1638234498.r7i5n6.1000222.16 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:793bbd5ea435faefb7b4598bb2ee097a139cce9266b43edf37f5515d5a94f5ad
3
+ size 49250344
exp12-1/056-module.18/events.out.tfevents.1638234499.r7i5n6.1000222.21 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:483d6cba47801aa66af95fb8fb9c00783201fa0c597670e74e8cb6150c4a09a0
3
+ size 48463912
exp12-1/056-module/events.out.tfevents.1638234499.r7i5n6.1000222.22 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:086baedeab33f4b419670c4ae782734a9a2026a93d08de3ae015a69af719a0e3
3
+ size 24281128