bigscience-bot commited on
Commit
72d9069
·
1 Parent(s): ee64e24
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. max/000-module.3.input_layernorm/events.out.tfevents.1637040677.r6i3n0.1690606.4 +3 -0
  2. max/000-module.3.mlp.dense_4h_to_h/events.out.tfevents.1637040678.r6i3n0.1690606.12 +3 -0
  3. max/000-module.3.mlp.dense_h_to_4h/events.out.tfevents.1637040678.r6i3n0.1690606.11 +3 -0
  4. max/000-module.3.mlp/events.out.tfevents.1637040678.r6i3n0.1690606.13 +3 -0
  5. max/000-module.3.post_attention_layernorm/events.out.tfevents.1637040678.r6i3n0.1690606.10 +3 -0
  6. max/000-module.3.self_attention.attention_dropout/events.out.tfevents.1637040677.r6i3n0.1690606.7 +3 -0
  7. max/000-module.3.self_attention.dense/events.out.tfevents.1637040677.r6i3n0.1690606.8 +3 -0
  8. max/000-module.3.self_attention.query_key_value/events.out.tfevents.1637040677.r6i3n0.1690606.5 +3 -0
  9. max/000-module.3.self_attention.scale_mask_softmax/events.out.tfevents.1637040677.r6i3n0.1690606.6 +3 -0
  10. max/000-module.3.self_attention/events.out.tfevents.1637040677.r6i3n0.1690606.9 +3 -0
  11. max/000-module.3/events.out.tfevents.1637040678.r6i3n0.1690606.14 +3 -0
  12. max/000-module.4.input_layernorm/events.out.tfevents.1637040678.r6i3n0.1690606.15 +3 -0
  13. max/000-module.4.mlp.dense_4h_to_h/events.out.tfevents.1637040678.r6i3n0.1690606.23 +3 -0
  14. max/000-module.4.mlp.dense_h_to_4h/events.out.tfevents.1637040678.r6i3n0.1690606.22 +3 -0
  15. max/000-module.4.mlp/events.out.tfevents.1637040678.r6i3n0.1690606.24 +3 -0
  16. max/000-module.4.post_attention_layernorm/events.out.tfevents.1637040678.r6i3n0.1690606.21 +3 -0
  17. max/000-module.4.self_attention.attention_dropout/events.out.tfevents.1637040678.r6i3n0.1690606.18 +3 -0
  18. max/000-module.4.self_attention.dense/events.out.tfevents.1637040678.r6i3n0.1690606.19 +3 -0
  19. max/000-module.4.self_attention.query_key_value/events.out.tfevents.1637040678.r6i3n0.1690606.16 +3 -0
  20. max/000-module.4.self_attention.scale_mask_softmax/events.out.tfevents.1637040678.r6i3n0.1690606.17 +3 -0
  21. max/000-module.4.self_attention/events.out.tfevents.1637040678.r6i3n0.1690606.20 +3 -0
  22. max/000-module.4/events.out.tfevents.1637040678.r6i3n0.1690606.25 +3 -0
  23. max/000-module.tied_modules.embed.embedding_dropout/events.out.tfevents.1637040677.r6i3n0.1690606.2 +3 -0
  24. max/000-module.tied_modules.embed.position_embeddings/events.out.tfevents.1637040677.r6i3n0.1690606.1 +3 -0
  25. max/000-module.tied_modules.embed.word_embeddings/events.out.tfevents.1637040677.r6i3n0.1690606.0 +3 -0
  26. max/000-module.tied_modules.embed/events.out.tfevents.1637040677.r6i3n0.1690606.3 +3 -0
  27. max/000-module/events.out.tfevents.1637040678.r6i3n0.1690606.26 +3 -0
  28. max/001-module.3.input_layernorm/events.out.tfevents.1637040677.r6i3n0.1690607.4 +3 -0
  29. max/001-module.3.mlp.dense_4h_to_h/events.out.tfevents.1637040678.r6i3n0.1690607.12 +3 -0
  30. max/001-module.3.mlp.dense_h_to_4h/events.out.tfevents.1637040678.r6i3n0.1690607.11 +3 -0
  31. max/001-module.3.mlp/events.out.tfevents.1637040678.r6i3n0.1690607.13 +3 -0
  32. max/001-module.3.post_attention_layernorm/events.out.tfevents.1637040678.r6i3n0.1690607.10 +3 -0
  33. max/001-module.3.self_attention.attention_dropout/events.out.tfevents.1637040677.r6i3n0.1690607.7 +3 -0
  34. max/001-module.3.self_attention.dense/events.out.tfevents.1637040677.r6i3n0.1690607.8 +3 -0
  35. max/001-module.3.self_attention.query_key_value/events.out.tfevents.1637040677.r6i3n0.1690607.5 +3 -0
  36. max/001-module.3.self_attention.scale_mask_softmax/events.out.tfevents.1637040677.r6i3n0.1690607.6 +3 -0
  37. max/001-module.3.self_attention/events.out.tfevents.1637040677.r6i3n0.1690607.9 +3 -0
  38. max/001-module.3/events.out.tfevents.1637040678.r6i3n0.1690607.14 +3 -0
  39. max/001-module.4.input_layernorm/events.out.tfevents.1637040678.r6i3n0.1690607.15 +3 -0
  40. max/001-module.4.mlp.dense_4h_to_h/events.out.tfevents.1637040678.r6i3n0.1690607.23 +3 -0
  41. max/001-module.4.mlp.dense_h_to_4h/events.out.tfevents.1637040678.r6i3n0.1690607.22 +3 -0
  42. max/001-module.4.mlp/events.out.tfevents.1637040678.r6i3n0.1690607.24 +3 -0
  43. max/001-module.4.post_attention_layernorm/events.out.tfevents.1637040678.r6i3n0.1690607.21 +3 -0
  44. max/001-module.4.self_attention.attention_dropout/events.out.tfevents.1637040678.r6i3n0.1690607.18 +3 -0
  45. max/001-module.4.self_attention.dense/events.out.tfevents.1637040678.r6i3n0.1690607.19 +3 -0
  46. max/001-module.4.self_attention.query_key_value/events.out.tfevents.1637040678.r6i3n0.1690607.16 +3 -0
  47. max/001-module.4.self_attention.scale_mask_softmax/events.out.tfevents.1637040678.r6i3n0.1690607.17 +3 -0
  48. max/001-module.4.self_attention/events.out.tfevents.1637040678.r6i3n0.1690607.20 +3 -0
  49. max/001-module.4/events.out.tfevents.1637040678.r6i3n0.1690607.25 +3 -0
  50. max/001-module.tied_modules.embed.embedding_dropout/events.out.tfevents.1637040677.r6i3n0.1690607.2 +3 -0
max/000-module.3.input_layernorm/events.out.tfevents.1637040677.r6i3n0.1690606.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9997f1c7ae14fdbd42637245c98dd6a7772ecf67d5705414f9d3844dc1d1cb6a
3
+ size 2211880
max/000-module.3.mlp.dense_4h_to_h/events.out.tfevents.1637040678.r6i3n0.1690606.12 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:adf9e077b383398a868b582475b36237a7b037dd6cb576f619be07f89940d5e2
3
+ size 3047464
max/000-module.3.mlp.dense_h_to_4h/events.out.tfevents.1637040678.r6i3n0.1690606.11 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:390ac5468dbcf38046b735e9442d0e0de477123a361ae664c33336028cf86ccc
3
+ size 3047464
max/000-module.3.mlp/events.out.tfevents.1637040678.r6i3n0.1690606.13 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7df1771d6c66766551da12834144c7e2cd4a18fa68c21f9e149c01fe2b1a6531
3
+ size 1572904
max/000-module.3.post_attention_layernorm/events.out.tfevents.1637040678.r6i3n0.1690606.10 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e44b7ec36328e93ee22bd24a503484f5cf38d5e5c55ff079a7fb6b0b76766abf
3
+ size 2211880
max/000-module.3.self_attention.attention_dropout/events.out.tfevents.1637040677.r6i3n0.1690606.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:659be3417d7b10684e3c39065d717fb8a1ffae743df9ab34994df62f04b127e4
3
+ size 737320
max/000-module.3.self_attention.dense/events.out.tfevents.1637040677.r6i3n0.1690606.8 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c5436be10254d3aa17b3334ec9e22c85ebe161768968fea24b95f3de75544f3
3
+ size 3047464
max/000-module.3.self_attention.query_key_value/events.out.tfevents.1637040677.r6i3n0.1690606.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff3075f76325043de9781bb9898366db44da9913d25fe868079e83b267eb8c27
3
+ size 2261032
max/000-module.3.self_attention.scale_mask_softmax/events.out.tfevents.1637040677.r6i3n0.1690606.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f28c87a22bfc8212184d02221308e0ee02b5a629a669e12292617dcd8ffab3d
3
+ size 1507368
max/000-module.3.self_attention/events.out.tfevents.1637040677.r6i3n0.1690606.9 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16e0d7359e970b75777e9a82b0f68501ac6dea7babc443e8016188de3d99b260
3
+ size 2342952
max/000-module.3/events.out.tfevents.1637040678.r6i3n0.1690606.14 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3c0fe3031ea742fb11112cd235eb4e2fe79ab0ed497e9cb96282527c92e287f
3
+ size 737320
max/000-module.4.input_layernorm/events.out.tfevents.1637040678.r6i3n0.1690606.15 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7a57e70cef8f7e01dfed9be712441eca68cb0e2b13e34466c2f3b6bed8d44bd
3
+ size 2211880
max/000-module.4.mlp.dense_4h_to_h/events.out.tfevents.1637040678.r6i3n0.1690606.23 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49bd48a77dace52e26325cd9c3a59ba262684af77967eeae3f0ae01391bf97ce
3
+ size 3047464
max/000-module.4.mlp.dense_h_to_4h/events.out.tfevents.1637040678.r6i3n0.1690606.22 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c0a2cf1a4460bec2dbb2991e98cc6e9359bcafddeb8f4a43a9a55673f4d16f7
3
+ size 3047464
max/000-module.4.mlp/events.out.tfevents.1637040678.r6i3n0.1690606.24 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0847fda3fc9e2730117dd4373c65915f3f4c828c66216446266d1bdff8f14177
3
+ size 1572904
max/000-module.4.post_attention_layernorm/events.out.tfevents.1637040678.r6i3n0.1690606.21 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a9c1fa4d2041aa430080a590c22e0cc0833936b51011ea5e9b55bbead99dd2d
3
+ size 2211880
max/000-module.4.self_attention.attention_dropout/events.out.tfevents.1637040678.r6i3n0.1690606.18 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:191fab9de417a1b47907fcb0b7a35d283d3d7542d45b0a54f78be289ef4f8cda
3
+ size 737320
max/000-module.4.self_attention.dense/events.out.tfevents.1637040678.r6i3n0.1690606.19 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:709f59c3b9ae749dd7b30bc944d973dc2aa654313cd4461fac41b3a04d016f97
3
+ size 3047464
max/000-module.4.self_attention.query_key_value/events.out.tfevents.1637040678.r6i3n0.1690606.16 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13e98373b676c0e0d1a03dd97e15bc508c53618e4ae75ea9ea25c3138684b492
3
+ size 2261032
max/000-module.4.self_attention.scale_mask_softmax/events.out.tfevents.1637040678.r6i3n0.1690606.17 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:653c4aed9c1c447bebc454d3b9ee64b933ea8c8434ff9483722a2484f13377d2
3
+ size 1507368
max/000-module.4.self_attention/events.out.tfevents.1637040678.r6i3n0.1690606.20 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ef310a09ed4a2c08e9e0edee46006f8aabf37eedfae8f6df0719dcc67c32dcf
3
+ size 2342952
max/000-module.4/events.out.tfevents.1637040678.r6i3n0.1690606.25 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8e646cb3961742b9674bdbf3b0525833673bac3d968eda5667c66c207f71180
3
+ size 737320
max/000-module.tied_modules.embed.embedding_dropout/events.out.tfevents.1637040677.r6i3n0.1690606.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6892778779f583689278c16d8262462d010d9f82966dc288e41228614422e67
3
+ size 368680
max/000-module.tied_modules.embed.position_embeddings/events.out.tfevents.1637040677.r6i3n0.1690606.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed9ec3acfc74cd3c6a47ce579adbbc48453843786b747ee945404ad75885e58d
3
+ size 753704
max/000-module.tied_modules.embed.word_embeddings/events.out.tfevents.1637040677.r6i3n0.1690606.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb1323c3bc495a6328188e8df43a4968f39c533c7563faca4d89b9ab24890ff6
3
+ size 753704
max/000-module.tied_modules.embed/events.out.tfevents.1637040677.r6i3n0.1690606.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0eb75973362472eea6bec9c230a2f50018011c064e925f711c83524f2f241ed0
3
+ size 40
max/000-module/events.out.tfevents.1637040678.r6i3n0.1690606.26 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64bd0264658258467ec8ffceaf5353f4f83f26bfe9db934ce6094cfd64137b18
3
+ size 40
max/001-module.3.input_layernorm/events.out.tfevents.1637040677.r6i3n0.1690607.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9ab6282cf03a105e93b7fb46f67b8a4e1bd3c08bc2663759aae5e599deda2e7
3
+ size 2211880
max/001-module.3.mlp.dense_4h_to_h/events.out.tfevents.1637040678.r6i3n0.1690607.12 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2cd17ca191924198aa4248659be50ae4b8a334c4e4b69c6b7c98c51b76e94961
3
+ size 3047464
max/001-module.3.mlp.dense_h_to_4h/events.out.tfevents.1637040678.r6i3n0.1690607.11 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:980c26079cb47206f11a89dc0482c16222a88a202a412af44e4d7ebcb011f6fb
3
+ size 3047464
max/001-module.3.mlp/events.out.tfevents.1637040678.r6i3n0.1690607.13 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c82286214bce5eed38e146dbb1c4ede8d295dcb6d74d2d9c2bad5ba02578f8ee
3
+ size 1572904
max/001-module.3.post_attention_layernorm/events.out.tfevents.1637040678.r6i3n0.1690607.10 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:982eae2730eac2ad7d9bbbb8ba2839790a0bc3f56d2021a8bcea356668ece150
3
+ size 2211880
max/001-module.3.self_attention.attention_dropout/events.out.tfevents.1637040677.r6i3n0.1690607.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:680e0234d86b01cbe5b068ef612a56c18e777a3f40cf359061b5b0405015b511
3
+ size 737320
max/001-module.3.self_attention.dense/events.out.tfevents.1637040677.r6i3n0.1690607.8 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28b2b266cf757e1dbde695f977ffb2d7ddc40cd495bd01713a5296e4c4f9a81e
3
+ size 3047464
max/001-module.3.self_attention.query_key_value/events.out.tfevents.1637040677.r6i3n0.1690607.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31bf552a4f4fbf11fe63d4b2cbd1dabee49b4adee3f20198443b300a82f0e15c
3
+ size 2261032
max/001-module.3.self_attention.scale_mask_softmax/events.out.tfevents.1637040677.r6i3n0.1690607.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d42249027db2c505dd41cd53f3f39814f329ade1ef5a15625548ee948a925bf4
3
+ size 1507368
max/001-module.3.self_attention/events.out.tfevents.1637040677.r6i3n0.1690607.9 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b07833ab97741cebb19d264a6204acf18bbda6b1a4025c3f5ce40a73e29c4a54
3
+ size 2342952
max/001-module.3/events.out.tfevents.1637040678.r6i3n0.1690607.14 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ee95a904d2e669ad25b9601b039eb8d9566f09c1864f1fb86c0f9c9d7165e86
3
+ size 737320
max/001-module.4.input_layernorm/events.out.tfevents.1637040678.r6i3n0.1690607.15 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0df2386bf49a6aa55a354062fef9c4799c89e92dc1ac1fa3b382799772f7c0e4
3
+ size 2211880
max/001-module.4.mlp.dense_4h_to_h/events.out.tfevents.1637040678.r6i3n0.1690607.23 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e605ab063cb258aa17777846ac1d9ee5c6e8702fde59caf2790ed13e5f7e21b9
3
+ size 3047464
max/001-module.4.mlp.dense_h_to_4h/events.out.tfevents.1637040678.r6i3n0.1690607.22 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5dcedaddbfb1909bfa600b7c9481f49134efefa6fa165ed7c869f124fb5c282
3
+ size 3047464
max/001-module.4.mlp/events.out.tfevents.1637040678.r6i3n0.1690607.24 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ffef7491de2c4a2859da2625808c3f317ef4a6b4630c5184a08b10a5dad5dc6
3
+ size 1572904
max/001-module.4.post_attention_layernorm/events.out.tfevents.1637040678.r6i3n0.1690607.21 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60dc27a89bb6cc9b5091458e119afb5eae375bd1be5859dcab617f9ef2586520
3
+ size 2211880
max/001-module.4.self_attention.attention_dropout/events.out.tfevents.1637040678.r6i3n0.1690607.18 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afca7a3eb97f02c42e967fb2be6c2811164a6cf413bf7f83dcea1d9b4350d787
3
+ size 737320
max/001-module.4.self_attention.dense/events.out.tfevents.1637040678.r6i3n0.1690607.19 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a565a39f986bf895c821aaa2b5a19e2dcec92bffe19042479ac606225f354739
3
+ size 3047464
max/001-module.4.self_attention.query_key_value/events.out.tfevents.1637040678.r6i3n0.1690607.16 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:832e5bb0966960b61292533c2e5f4166b2dd1a0a39574103314fac4470420fe5
3
+ size 2261032
max/001-module.4.self_attention.scale_mask_softmax/events.out.tfevents.1637040678.r6i3n0.1690607.17 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3339587cf9a0fc1c92d7b68f8a2685cf9d454fe2546c621db56337d4b45d145
3
+ size 1507368
max/001-module.4.self_attention/events.out.tfevents.1637040678.r6i3n0.1690607.20 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:100cb4c696c0efb8944ba0489c32c497596d1ff6c141e14e33c097ad8ce73d24
3
+ size 2342952
max/001-module.4/events.out.tfevents.1637040678.r6i3n0.1690607.25 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fce9c50c45707564e074d8c96c11b09536bf332e41aa544d80fa8616ac3b18d0
3
+ size 737320
max/001-module.tied_modules.embed.embedding_dropout/events.out.tfevents.1637040677.r6i3n0.1690607.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a5f552b5bf42cd3d78a1132801e5b39115def77d70126a6ed0b748c49ce1376
3
+ size 368680