stas commited on
Commit
1dab139
·
1 Parent(s): f2c9d24
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. exp12/000-module.3.input_layernorm/events.out.tfevents.1638184552.r7i2n3.3196863.4 +3 -0
  2. exp12/000-module.3.mlp.dense_4h_to_h/events.out.tfevents.1638184553.r7i2n3.3196863.12 +3 -0
  3. exp12/000-module.3.mlp.dense_h_to_4h/events.out.tfevents.1638184553.r7i2n3.3196863.11 +3 -0
  4. exp12/000-module.3.mlp/events.out.tfevents.1638184553.r7i2n3.3196863.13 +3 -0
  5. exp12/000-module.3.post_attention_layernorm/events.out.tfevents.1638184552.r7i2n3.3196863.10 +3 -0
  6. exp12/000-module.3.self_attention.attention_dropout/events.out.tfevents.1638184552.r7i2n3.3196863.7 +3 -0
  7. exp12/000-module.3.self_attention.dense/events.out.tfevents.1638184552.r7i2n3.3196863.8 +3 -0
  8. exp12/000-module.3.self_attention.query_key_value/events.out.tfevents.1638184552.r7i2n3.3196863.5 +3 -0
  9. exp12/000-module.3.self_attention.scale_mask_softmax/events.out.tfevents.1638184552.r7i2n3.3196863.6 +3 -0
  10. exp12/000-module.3.self_attention/events.out.tfevents.1638184552.r7i2n3.3196863.9 +3 -0
  11. exp12/000-module.3/events.out.tfevents.1638184553.r7i2n3.3196863.14 +3 -0
  12. exp12/000-module.4.input_layernorm/events.out.tfevents.1638184553.r7i2n3.3196863.15 +3 -0
  13. exp12/000-module.4.mlp.dense_4h_to_h/events.out.tfevents.1638184553.r7i2n3.3196863.23 +3 -0
  14. exp12/000-module.4.mlp.dense_h_to_4h/events.out.tfevents.1638184553.r7i2n3.3196863.22 +3 -0
  15. exp12/000-module.4.mlp/events.out.tfevents.1638184553.r7i2n3.3196863.24 +3 -0
  16. exp12/000-module.4.post_attention_layernorm/events.out.tfevents.1638184553.r7i2n3.3196863.21 +3 -0
  17. exp12/000-module.4.self_attention.attention_dropout/events.out.tfevents.1638184553.r7i2n3.3196863.18 +3 -0
  18. exp12/000-module.4.self_attention.dense/events.out.tfevents.1638184553.r7i2n3.3196863.19 +3 -0
  19. exp12/000-module.4.self_attention.query_key_value/events.out.tfevents.1638184553.r7i2n3.3196863.16 +3 -0
  20. exp12/000-module.4.self_attention.scale_mask_softmax/events.out.tfevents.1638184553.r7i2n3.3196863.17 +3 -0
  21. exp12/000-module.4.self_attention/events.out.tfevents.1638184553.r7i2n3.3196863.20 +3 -0
  22. exp12/000-module.4/events.out.tfevents.1638184553.r7i2n3.3196863.25 +3 -0
  23. exp12/000-module.tied_modules.embed.embedding_dropout/events.out.tfevents.1638184552.r7i2n3.3196863.2 +3 -0
  24. exp12/000-module.tied_modules.embed.position_embeddings/events.out.tfevents.1638184552.r7i2n3.3196863.1 +3 -0
  25. exp12/000-module.tied_modules.embed.word_embeddings/events.out.tfevents.1638184552.r7i2n3.3196863.0 +3 -0
  26. exp12/000-module.tied_modules.embed/events.out.tfevents.1638184552.r7i2n3.3196863.3 +3 -0
  27. exp12/000-module/events.out.tfevents.1638184553.r7i2n3.3196863.26 +3 -0
  28. exp12/028-module.17.input_layernorm/events.out.tfevents.1638184564.r7i5n0.901020.0 +3 -0
  29. exp12/028-module.17.mlp.dense_4h_to_h/events.out.tfevents.1638184565.r7i5n0.901020.8 +3 -0
  30. exp12/028-module.17.mlp.dense_h_to_4h/events.out.tfevents.1638184565.r7i5n0.901020.7 +3 -0
  31. exp12/028-module.17.mlp/events.out.tfevents.1638184565.r7i5n0.901020.9 +3 -0
  32. exp12/028-module.17.post_attention_layernorm/events.out.tfevents.1638184565.r7i5n0.901020.6 +3 -0
  33. exp12/028-module.17.self_attention.attention_dropout/events.out.tfevents.1638184564.r7i5n0.901020.3 +3 -0
  34. exp12/028-module.17.self_attention.dense/events.out.tfevents.1638184564.r7i5n0.901020.4 +3 -0
  35. exp12/028-module.17.self_attention.query_key_value/events.out.tfevents.1638184564.r7i5n0.901020.1 +3 -0
  36. exp12/028-module.17.self_attention.scale_mask_softmax/events.out.tfevents.1638184564.r7i5n0.901020.2 +3 -0
  37. exp12/028-module.17.self_attention/events.out.tfevents.1638184564.r7i5n0.901020.5 +3 -0
  38. exp12/028-module.17/events.out.tfevents.1638184565.r7i5n0.901020.10 +3 -0
  39. exp12/028-module.18.input_layernorm/events.out.tfevents.1638184565.r7i5n0.901020.11 +3 -0
  40. exp12/028-module.18.mlp.dense_4h_to_h/events.out.tfevents.1638184565.r7i5n0.901020.19 +3 -0
  41. exp12/028-module.18.mlp.dense_h_to_4h/events.out.tfevents.1638184565.r7i5n0.901020.18 +3 -0
  42. exp12/028-module.18.mlp/events.out.tfevents.1638184565.r7i5n0.901020.20 +3 -0
  43. exp12/028-module.18.post_attention_layernorm/events.out.tfevents.1638184565.r7i5n0.901020.17 +3 -0
  44. exp12/028-module.18.self_attention.attention_dropout/events.out.tfevents.1638184565.r7i5n0.901020.14 +3 -0
  45. exp12/028-module.18.self_attention.dense/events.out.tfevents.1638184565.r7i5n0.901020.15 +3 -0
  46. exp12/028-module.18.self_attention.query_key_value/events.out.tfevents.1638184565.r7i5n0.901020.12 +3 -0
  47. exp12/028-module.18.self_attention.scale_mask_softmax/events.out.tfevents.1638184565.r7i5n0.901020.13 +3 -0
  48. exp12/028-module.18.self_attention/events.out.tfevents.1638184565.r7i5n0.901020.16 +3 -0
  49. exp12/028-module.18/events.out.tfevents.1638184565.r7i5n0.901020.21 +3 -0
  50. exp12/028-module/events.out.tfevents.1638184565.r7i5n0.901020.22 +3 -0
exp12/000-module.3.input_layernorm/events.out.tfevents.1638184552.r7i2n3.3196863.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77b9723a64f3e52b0298b25f0d25721b53bbe5c5ddb950aa7961f11d5a704f79
3
+ size 34964200
exp12/000-module.3.mlp.dense_4h_to_h/events.out.tfevents.1638184553.r7i2n3.3196863.12 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae46e64c74647b36546b691cbe80c61467d2e8084eb259acee73b166bf36b2d0
3
+ size 35294440
exp12/000-module.3.mlp.dense_h_to_4h/events.out.tfevents.1638184553.r7i2n3.3196863.11 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15aa30f307ac598f79260c4e834a2bd58af71ace8508776f5c07d0c848d8b72f
3
+ size 35294440
exp12/000-module.3.mlp/events.out.tfevents.1638184553.r7i2n3.3196863.13 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8d44408acc42c1d0fb59d499e6abba473928fc9965bbfc9f3ed3e755957d45b
3
+ size 20681320
exp12/000-module.3.post_attention_layernorm/events.out.tfevents.1638184552.r7i2n3.3196863.10 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:581026e76b670c797399d0b3319eb64eb63235f5a8fdeaa4c04d55f61946d126
3
+ size 34964200
exp12/000-module.3.self_attention.attention_dropout/events.out.tfevents.1638184552.r7i2n3.3196863.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09ae1a532dc9bea63a0bb9374cca3410024276bfc532cc453686e5001a8485d1
3
+ size 15232360
exp12/000-module.3.self_attention.dense/events.out.tfevents.1638184552.r7i2n3.3196863.8 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:444ef3f2822f129fb6640e0ca5536460f7493e53016af5565c235bbfc8b12a11
3
+ size 35294440
exp12/000-module.3.self_attention.query_key_value/events.out.tfevents.1638184552.r7i2n3.3196863.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f49caad64163a7f0b60a792ae7ba3286e05c1658537e1d596ac8e3958c02cc7
3
+ size 32652520
exp12/000-module.3.self_attention.scale_mask_softmax/events.out.tfevents.1638184552.r7i2n3.3196863.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf4bb9c549740d9f9d3197a1d0e7fed3863ecaa1d91fac914dc8354a15399af6
3
+ size 15232360
exp12/000-module.3.self_attention/events.out.tfevents.1638184552.r7i2n3.3196863.9 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e880989c79c4deb3ac62615697b9351e95173a8ee2a003929dafeb6c9ee25a5a
3
+ size 20681320
exp12/000-module.3/events.out.tfevents.1638184553.r7i2n3.3196863.14 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1642643d1ef105ddc0fce26a931a39059594812a5b0b2775993e390e0e6beb0d
3
+ size 20351080
exp12/000-module.4.input_layernorm/events.out.tfevents.1638184553.r7i2n3.3196863.15 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1691bf962fbd373d02b75839b898ca6dec996720c324776b11348eb48fc490cd
3
+ size 34964200
exp12/000-module.4.mlp.dense_4h_to_h/events.out.tfevents.1638184553.r7i2n3.3196863.23 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dac7b159773687f18efb4e507811ac64cabeba0d1a61cd465ce210676e38badf
3
+ size 35294440
exp12/000-module.4.mlp.dense_h_to_4h/events.out.tfevents.1638184553.r7i2n3.3196863.22 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2334a1b124a34385befa5efade65dae4396a4e5344f2a6cef2b24bf5ab431014
3
+ size 35294440
exp12/000-module.4.mlp/events.out.tfevents.1638184553.r7i2n3.3196863.24 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:439596f079cdc73db8a7195cd9988945afb942596abe3d7841421c51baa9b04a
3
+ size 20681320
exp12/000-module.4.post_attention_layernorm/events.out.tfevents.1638184553.r7i2n3.3196863.21 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8649cbdfdd29c9d005976a65107f19a37a2d8e4487ceb21890106ebfaea304d0
3
+ size 34964200
exp12/000-module.4.self_attention.attention_dropout/events.out.tfevents.1638184553.r7i2n3.3196863.18 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:674d5da2dfeeea59e77fff50405777399716b9cb90e71c749754f0ec92f6f11f
3
+ size 15232360
exp12/000-module.4.self_attention.dense/events.out.tfevents.1638184553.r7i2n3.3196863.19 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c517cae1f4d124e2ff0d414c610d990635d301d01e54f4957a0d40ccf36a6fc0
3
+ size 35294440
exp12/000-module.4.self_attention.query_key_value/events.out.tfevents.1638184553.r7i2n3.3196863.16 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80337eed03b9f5a96847ab640833ca263bd8a1fb747e0b2d73c05b554919a0a7
3
+ size 32652520
exp12/000-module.4.self_attention.scale_mask_softmax/events.out.tfevents.1638184553.r7i2n3.3196863.17 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c13b894b0de1b16c690d96765c7947090d17b3773d288b0656f1d0d786906ca
3
+ size 15232360
exp12/000-module.4.self_attention/events.out.tfevents.1638184553.r7i2n3.3196863.20 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa2df26aa31249986cb43d7e3a360a45ba54fd2cf8ccb05c8dea3024794a922d
3
+ size 20681320
exp12/000-module.4/events.out.tfevents.1638184553.r7i2n3.3196863.25 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed3c13413b1cc722b2513ca3bcece4acda133fc823da9ca9069c5c1377cb9677
3
+ size 20351080
exp12/000-module.tied_modules.embed.embedding_dropout/events.out.tfevents.1638184552.r7i2n3.3196863.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:403e944f860a9754d701324c956687c3280b76e815894bed544ad61bc57d43a6
3
+ size 10196200
exp12/000-module.tied_modules.embed.position_embeddings/events.out.tfevents.1638184552.r7i2n3.3196863.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:add4654c64e4ff0a42809ad60c43635a4134e317ade60194b906d0cba4797f00
3
+ size 12590440
exp12/000-module.tied_modules.embed.word_embeddings/events.out.tfevents.1638184552.r7i2n3.3196863.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19c59c78b5e0b0c4b2e78f722462a5b7cc8c095b885709ccea4eb204c7b2d0c7
3
+ size 12590440
exp12/000-module.tied_modules.embed/events.out.tfevents.1638184552.r7i2n3.3196863.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2cddfcb32efa15d03bf08024d7755f101095b6b4b1fbb48c556d860aebefbf62
3
+ size 7636840
exp12/000-module/events.out.tfevents.1638184553.r7i2n3.3196863.26 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39583d662686a1c8117f25887ef356b03add6100f870e0a5be2affd1e6a2829b
3
+ size 7636840
exp12/028-module.17.input_layernorm/events.out.tfevents.1638184564.r7i5n0.901020.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efc5099f5baa23d7b41f64e5118021811a49e4d0c32a48d1d946441e1708b8b6
3
+ size 34964200
exp12/028-module.17.mlp.dense_4h_to_h/events.out.tfevents.1638184565.r7i5n0.901020.8 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bebbdb5f7532272c41e7fe3b2b34f204c9ad2258e9c129a97a4c7a79fa266f5b
3
+ size 35294440
exp12/028-module.17.mlp.dense_h_to_4h/events.out.tfevents.1638184565.r7i5n0.901020.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10ee637087beff074605ffcdae21253aed01de9fbfc8ed4c7da0a957f3ab1955
3
+ size 35294440
exp12/028-module.17.mlp/events.out.tfevents.1638184565.r7i5n0.901020.9 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9100be613d06c20b801966856b2328a78c47f413a5da70c5553ddcba7b3a1ed
3
+ size 20681320
exp12/028-module.17.post_attention_layernorm/events.out.tfevents.1638184565.r7i5n0.901020.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da3eeb8c386802eecec12065db1ee487fbc920aaeaed982db54eb12d6b1edf0f
3
+ size 34964200
exp12/028-module.17.self_attention.attention_dropout/events.out.tfevents.1638184564.r7i5n0.901020.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63f34ae9c839dca2b7cf7ef906b321649c3f4a9c603df02ec1683a5b2b0c919d
3
+ size 15232360
exp12/028-module.17.self_attention.dense/events.out.tfevents.1638184564.r7i5n0.901020.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc67d4149b1a94954e253d84b8fb81570adab4203f7a14d5cc77a4876f67ba73
3
+ size 35294440
exp12/028-module.17.self_attention.query_key_value/events.out.tfevents.1638184564.r7i5n0.901020.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d09f219af79854e066273469cc98ebe4b47c677a9b350b3e4caef39e85a666b
3
+ size 32652520
exp12/028-module.17.self_attention.scale_mask_softmax/events.out.tfevents.1638184564.r7i5n0.901020.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0d928f604a996cda9dc87f53ab83e12f84ece3d4ae7ab3346d60d2959de881a
3
+ size 15232360
exp12/028-module.17.self_attention/events.out.tfevents.1638184564.r7i5n0.901020.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b60ae0b8d36e246c85a92a8d63c3d4c12777fb5bc73c30b3f669b167cb240cd
3
+ size 20681320
exp12/028-module.17/events.out.tfevents.1638184565.r7i5n0.901020.10 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b4325fc6d3a49209321716d6c072fc94d71f644530e40a867f7a274a633ce88
3
+ size 20351080
exp12/028-module.18.input_layernorm/events.out.tfevents.1638184565.r7i5n0.901020.11 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d21dd61686667d742c4217279044623984ce51d7c40ae3ea16dee9353c1e8a3
3
+ size 34964200
exp12/028-module.18.mlp.dense_4h_to_h/events.out.tfevents.1638184565.r7i5n0.901020.19 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:306a41d60235cd2f637a190f59b7b9ccfbc91b2072cd5a91cfcaa10a2545c976
3
+ size 35294440
exp12/028-module.18.mlp.dense_h_to_4h/events.out.tfevents.1638184565.r7i5n0.901020.18 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6558cc7ad9de56aebaa6bf4844cd0e867c26d79e4a789477dd0f43e2faa4dfa
3
+ size 35294440
exp12/028-module.18.mlp/events.out.tfevents.1638184565.r7i5n0.901020.20 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23da557c4eb2567b8f10b3ab198fb62be0a9b6a5b857605e9597c5aeb970c4eb
3
+ size 20681320
exp12/028-module.18.post_attention_layernorm/events.out.tfevents.1638184565.r7i5n0.901020.17 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:173ab0b0f89563e5dd9bc21e48f3c034629cacf7937b31c67b77742f43f58cec
3
+ size 34964200
exp12/028-module.18.self_attention.attention_dropout/events.out.tfevents.1638184565.r7i5n0.901020.14 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e5f2fec1d01a4ab736a17bb908207dfae77d8ff5bd4853800a0830d7fa50c87
3
+ size 15232360
exp12/028-module.18.self_attention.dense/events.out.tfevents.1638184565.r7i5n0.901020.15 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8497c316e3b72f202bb420a3f56dbac8e78e199cd0c3ad2cdf60ea9f559cee44
3
+ size 35294440
exp12/028-module.18.self_attention.query_key_value/events.out.tfevents.1638184565.r7i5n0.901020.12 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7a4447ba92dcd40d29ef00720e0986f8c14f7011586a07f3e5e92b4bb2efc49
3
+ size 32652520
exp12/028-module.18.self_attention.scale_mask_softmax/events.out.tfevents.1638184565.r7i5n0.901020.13 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6affe0431d439d8caf285c1875ce466f90e2a7e9504ee41516d2e7647dbc7484
3
+ size 15232360
exp12/028-module.18.self_attention/events.out.tfevents.1638184565.r7i5n0.901020.16 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fca4495f2925f36349afeeb834698cdd413adae8da5a915c5bf24a88ed0ceed
3
+ size 20681320
exp12/028-module.18/events.out.tfevents.1638184565.r7i5n0.901020.21 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b64222c2b5fed2950f3f267e9b26995dffcdbf3dda88ee62393cfebb017c89e
3
+ size 20351080
exp12/028-module/events.out.tfevents.1638184565.r7i5n0.901020.22 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21bcc079aa79275e3a15c276121c320d2663c2d151915baa0303540bc23c1b6e
3
+ size 10196200