Training in progress, step 2000
Browse files- model-00001-of-00002.safetensors +1 -1
- model-00002-of-00002.safetensors +1 -1
- trainer_log.jsonl +400 -0
model-00001-of-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4988025760
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bd537c093919489f9a1e66f9c98b5be0ca4df2f0212dcf0cf24cb248eedeb35a
|
3 |
size 4988025760
|
model-00002-of-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 240691728
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b0a6e4517c3c68803e809227a19afe812d00f20c5215c95f45853dba6989b698
|
3 |
size 240691728
|
trainer_log.jsonl
CHANGED
@@ -1610,3 +1610,403 @@
|
|
1610 |
{"current_steps": 1610, "total_steps": 3906, "loss": 1.3526, "learning_rate": 2.5814441273287107e-05, "epoch": 0.41216, "percentage": 41.22, "elapsed_time": "5:55:07", "remaining_time": "8:26:26"}
|
1611 |
{"current_steps": 1611, "total_steps": 3906, "loss": 1.3944, "learning_rate": 2.5798888917752214e-05, "epoch": 0.412416, "percentage": 41.24, "elapsed_time": "5:55:20", "remaining_time": "8:26:13"}
|
1612 |
{"current_steps": 1612, "total_steps": 3906, "loss": 1.374, "learning_rate": 2.5783332732904804e-05, "epoch": 0.412672, "percentage": 41.27, "elapsed_time": "5:55:34", "remaining_time": "8:26:00"}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1610 |
{"current_steps": 1610, "total_steps": 3906, "loss": 1.3526, "learning_rate": 2.5814441273287107e-05, "epoch": 0.41216, "percentage": 41.22, "elapsed_time": "5:55:07", "remaining_time": "8:26:26"}
|
1611 |
{"current_steps": 1611, "total_steps": 3906, "loss": 1.3944, "learning_rate": 2.5798888917752214e-05, "epoch": 0.412416, "percentage": 41.24, "elapsed_time": "5:55:20", "remaining_time": "8:26:13"}
|
1612 |
{"current_steps": 1612, "total_steps": 3906, "loss": 1.374, "learning_rate": 2.5783332732904804e-05, "epoch": 0.412672, "percentage": 41.27, "elapsed_time": "5:55:34", "remaining_time": "8:26:00"}
|
1613 |
+
{"current_steps": 1613, "total_steps": 3906, "loss": 1.3745, "learning_rate": 2.576777272901744e-05, "epoch": 0.412928, "percentage": 41.3, "elapsed_time": "5:55:47", "remaining_time": "8:25:46"}
|
1614 |
+
{"current_steps": 1614, "total_steps": 3906, "loss": 1.3849, "learning_rate": 2.5752208916365212e-05, "epoch": 0.413184, "percentage": 41.32, "elapsed_time": "5:56:00", "remaining_time": "8:25:33"}
|
1615 |
+
{"current_steps": 1615, "total_steps": 3906, "loss": 1.3861, "learning_rate": 2.5736641305225743e-05, "epoch": 0.41344, "percentage": 41.35, "elapsed_time": "5:56:13", "remaining_time": "8:25:20"}
|
1616 |
+
{"current_steps": 1616, "total_steps": 3906, "loss": 1.4244, "learning_rate": 2.5721069905879133e-05, "epoch": 0.413696, "percentage": 41.37, "elapsed_time": "5:56:27", "remaining_time": "8:25:07"}
|
1617 |
+
{"current_steps": 1617, "total_steps": 3906, "loss": 1.3637, "learning_rate": 2.570549472860801e-05, "epoch": 0.413952, "percentage": 41.4, "elapsed_time": "5:56:40", "remaining_time": "8:24:53"}
|
1618 |
+
{"current_steps": 1618, "total_steps": 3906, "loss": 1.3453, "learning_rate": 2.5689915783697482e-05, "epoch": 0.414208, "percentage": 41.42, "elapsed_time": "5:56:53", "remaining_time": "8:24:40"}
|
1619 |
+
{"current_steps": 1619, "total_steps": 3906, "loss": 1.37, "learning_rate": 2.5674333081435143e-05, "epoch": 0.414464, "percentage": 41.45, "elapsed_time": "5:57:06", "remaining_time": "8:24:26"}
|
1620 |
+
{"current_steps": 1620, "total_steps": 3906, "loss": 1.3795, "learning_rate": 2.5658746632111074e-05, "epoch": 0.41472, "percentage": 41.47, "elapsed_time": "5:57:19", "remaining_time": "8:24:13"}
|
1621 |
+
{"current_steps": 1621, "total_steps": 3906, "loss": 1.3535, "learning_rate": 2.5643156446017848e-05, "epoch": 0.414976, "percentage": 41.5, "elapsed_time": "5:57:32", "remaining_time": "8:24:00"}
|
1622 |
+
{"current_steps": 1622, "total_steps": 3906, "loss": 1.3604, "learning_rate": 2.5627562533450472e-05, "epoch": 0.415232, "percentage": 41.53, "elapsed_time": "5:57:45", "remaining_time": "8:23:46"}
|
1623 |
+
{"current_steps": 1623, "total_steps": 3906, "loss": 1.3832, "learning_rate": 2.5611964904706428e-05, "epoch": 0.415488, "percentage": 41.55, "elapsed_time": "5:57:59", "remaining_time": "8:23:33"}
|
1624 |
+
{"current_steps": 1624, "total_steps": 3906, "loss": 1.4134, "learning_rate": 2.5596363570085662e-05, "epoch": 0.415744, "percentage": 41.58, "elapsed_time": "5:58:12", "remaining_time": "8:23:20"}
|
1625 |
+
{"current_steps": 1625, "total_steps": 3906, "loss": 1.3489, "learning_rate": 2.5580758539890553e-05, "epoch": 0.416, "percentage": 41.6, "elapsed_time": "5:58:25", "remaining_time": "8:23:07"}
|
1626 |
+
{"current_steps": 1626, "total_steps": 3906, "loss": 1.3886, "learning_rate": 2.5565149824425934e-05, "epoch": 0.416256, "percentage": 41.63, "elapsed_time": "5:58:38", "remaining_time": "8:22:53"}
|
1627 |
+
{"current_steps": 1627, "total_steps": 3906, "loss": 1.3808, "learning_rate": 2.5549537433999056e-05, "epoch": 0.416512, "percentage": 41.65, "elapsed_time": "5:58:51", "remaining_time": "8:22:40"}
|
1628 |
+
{"current_steps": 1628, "total_steps": 3906, "loss": 1.3681, "learning_rate": 2.5533921378919613e-05, "epoch": 0.416768, "percentage": 41.68, "elapsed_time": "5:59:05", "remaining_time": "8:22:27"}
|
1629 |
+
{"current_steps": 1629, "total_steps": 3906, "loss": 1.384, "learning_rate": 2.55183016694997e-05, "epoch": 0.417024, "percentage": 41.71, "elapsed_time": "5:59:18", "remaining_time": "8:22:13"}
|
1630 |
+
{"current_steps": 1630, "total_steps": 3906, "loss": 1.3844, "learning_rate": 2.5502678316053843e-05, "epoch": 0.41728, "percentage": 41.73, "elapsed_time": "5:59:31", "remaining_time": "8:22:00"}
|
1631 |
+
{"current_steps": 1631, "total_steps": 3906, "loss": 1.3767, "learning_rate": 2.548705132889896e-05, "epoch": 0.417536, "percentage": 41.76, "elapsed_time": "5:59:44", "remaining_time": "8:21:47"}
|
1632 |
+
{"current_steps": 1632, "total_steps": 3906, "loss": 1.38, "learning_rate": 2.5471420718354388e-05, "epoch": 0.417792, "percentage": 41.78, "elapsed_time": "5:59:57", "remaining_time": "8:21:33"}
|
1633 |
+
{"current_steps": 1633, "total_steps": 3906, "loss": 1.3975, "learning_rate": 2.5455786494741837e-05, "epoch": 0.418048, "percentage": 41.81, "elapsed_time": "6:00:10", "remaining_time": "8:21:20"}
|
1634 |
+
{"current_steps": 1634, "total_steps": 3906, "loss": 1.4076, "learning_rate": 2.544014866838541e-05, "epoch": 0.418304, "percentage": 41.83, "elapsed_time": "6:00:24", "remaining_time": "8:21:07"}
|
1635 |
+
{"current_steps": 1635, "total_steps": 3906, "loss": 1.3902, "learning_rate": 2.542450724961159e-05, "epoch": 0.41856, "percentage": 41.86, "elapsed_time": "6:00:37", "remaining_time": "8:20:53"}
|
1636 |
+
{"current_steps": 1636, "total_steps": 3906, "loss": 1.3779, "learning_rate": 2.540886224874923e-05, "epoch": 0.418816, "percentage": 41.88, "elapsed_time": "6:00:50", "remaining_time": "8:20:40"}
|
1637 |
+
{"current_steps": 1637, "total_steps": 3906, "loss": 1.3826, "learning_rate": 2.5393213676129554e-05, "epoch": 0.419072, "percentage": 41.91, "elapsed_time": "6:01:03", "remaining_time": "8:20:27"}
|
1638 |
+
{"current_steps": 1638, "total_steps": 3906, "loss": 1.4204, "learning_rate": 2.5377561542086137e-05, "epoch": 0.419328, "percentage": 41.94, "elapsed_time": "6:01:16", "remaining_time": "8:20:14"}
|
1639 |
+
{"current_steps": 1639, "total_steps": 3906, "loss": 1.4037, "learning_rate": 2.5361905856954914e-05, "epoch": 0.419584, "percentage": 41.96, "elapsed_time": "6:01:30", "remaining_time": "8:20:00"}
|
1640 |
+
{"current_steps": 1640, "total_steps": 3906, "loss": 1.3718, "learning_rate": 2.534624663107416e-05, "epoch": 0.41984, "percentage": 41.99, "elapsed_time": "6:01:43", "remaining_time": "8:19:47"}
|
1641 |
+
{"current_steps": 1641, "total_steps": 3906, "loss": 1.3503, "learning_rate": 2.5330583874784482e-05, "epoch": 0.420096, "percentage": 42.01, "elapsed_time": "6:01:56", "remaining_time": "8:19:34"}
|
1642 |
+
{"current_steps": 1642, "total_steps": 3906, "loss": 1.4129, "learning_rate": 2.5314917598428825e-05, "epoch": 0.420352, "percentage": 42.04, "elapsed_time": "6:02:09", "remaining_time": "8:19:20"}
|
1643 |
+
{"current_steps": 1643, "total_steps": 3906, "loss": 1.3932, "learning_rate": 2.5299247812352473e-05, "epoch": 0.420608, "percentage": 42.06, "elapsed_time": "6:02:22", "remaining_time": "8:19:07"}
|
1644 |
+
{"current_steps": 1644, "total_steps": 3906, "loss": 1.3481, "learning_rate": 2.5283574526902993e-05, "epoch": 0.420864, "percentage": 42.09, "elapsed_time": "6:02:35", "remaining_time": "8:18:54"}
|
1645 |
+
{"current_steps": 1645, "total_steps": 3906, "loss": 1.349, "learning_rate": 2.5267897752430298e-05, "epoch": 0.42112, "percentage": 42.11, "elapsed_time": "6:02:49", "remaining_time": "8:18:40"}
|
1646 |
+
{"current_steps": 1646, "total_steps": 3906, "loss": 1.3434, "learning_rate": 2.5252217499286588e-05, "epoch": 0.421376, "percentage": 42.14, "elapsed_time": "6:03:02", "remaining_time": "8:18:27"}
|
1647 |
+
{"current_steps": 1647, "total_steps": 3906, "loss": 1.4296, "learning_rate": 2.5236533777826352e-05, "epoch": 0.421632, "percentage": 42.17, "elapsed_time": "6:03:15", "remaining_time": "8:18:14"}
|
1648 |
+
{"current_steps": 1648, "total_steps": 3906, "loss": 1.4062, "learning_rate": 2.522084659840638e-05, "epoch": 0.421888, "percentage": 42.19, "elapsed_time": "6:03:28", "remaining_time": "8:18:01"}
|
1649 |
+
{"current_steps": 1649, "total_steps": 3906, "loss": 1.3614, "learning_rate": 2.5205155971385767e-05, "epoch": 0.422144, "percentage": 42.22, "elapsed_time": "6:03:41", "remaining_time": "8:17:47"}
|
1650 |
+
{"current_steps": 1650, "total_steps": 3906, "loss": 1.3762, "learning_rate": 2.518946190712584e-05, "epoch": 0.4224, "percentage": 42.24, "elapsed_time": "6:03:55", "remaining_time": "8:17:34"}
|
1651 |
+
{"current_steps": 1651, "total_steps": 3906, "loss": 1.3922, "learning_rate": 2.517376441599022e-05, "epoch": 0.422656, "percentage": 42.27, "elapsed_time": "6:04:08", "remaining_time": "8:17:21"}
|
1652 |
+
{"current_steps": 1652, "total_steps": 3906, "loss": 1.3904, "learning_rate": 2.5158063508344795e-05, "epoch": 0.422912, "percentage": 42.29, "elapsed_time": "6:04:21", "remaining_time": "8:17:08"}
|
1653 |
+
{"current_steps": 1653, "total_steps": 3906, "loss": 1.362, "learning_rate": 2.51423591945577e-05, "epoch": 0.423168, "percentage": 42.32, "elapsed_time": "6:04:34", "remaining_time": "8:16:54"}
|
1654 |
+
{"current_steps": 1654, "total_steps": 3906, "loss": 1.3682, "learning_rate": 2.5126651484999326e-05, "epoch": 0.423424, "percentage": 42.35, "elapsed_time": "6:04:47", "remaining_time": "8:16:41"}
|
1655 |
+
{"current_steps": 1655, "total_steps": 3906, "loss": 1.3425, "learning_rate": 2.5110940390042307e-05, "epoch": 0.42368, "percentage": 42.37, "elapsed_time": "6:05:01", "remaining_time": "8:16:28"}
|
1656 |
+
{"current_steps": 1656, "total_steps": 3906, "loss": 1.3785, "learning_rate": 2.5095225920061497e-05, "epoch": 0.423936, "percentage": 42.4, "elapsed_time": "6:05:14", "remaining_time": "8:16:14"}
|
1657 |
+
{"current_steps": 1657, "total_steps": 3906, "loss": 1.3701, "learning_rate": 2.5079508085433997e-05, "epoch": 0.424192, "percentage": 42.42, "elapsed_time": "6:05:27", "remaining_time": "8:16:01"}
|
1658 |
+
{"current_steps": 1658, "total_steps": 3906, "loss": 1.404, "learning_rate": 2.5063786896539127e-05, "epoch": 0.424448, "percentage": 42.45, "elapsed_time": "6:05:40", "remaining_time": "8:15:48"}
|
1659 |
+
{"current_steps": 1659, "total_steps": 3906, "loss": 1.3714, "learning_rate": 2.5048062363758402e-05, "epoch": 0.424704, "percentage": 42.47, "elapsed_time": "6:05:53", "remaining_time": "8:15:34"}
|
1660 |
+
{"current_steps": 1660, "total_steps": 3906, "loss": 1.3984, "learning_rate": 2.503233449747558e-05, "epoch": 0.42496, "percentage": 42.5, "elapsed_time": "6:06:06", "remaining_time": "8:15:21"}
|
1661 |
+
{"current_steps": 1661, "total_steps": 3906, "loss": 1.3376, "learning_rate": 2.5016603308076595e-05, "epoch": 0.425216, "percentage": 42.52, "elapsed_time": "6:06:20", "remaining_time": "8:15:08"}
|
1662 |
+
{"current_steps": 1662, "total_steps": 3906, "loss": 1.3666, "learning_rate": 2.500086880594958e-05, "epoch": 0.425472, "percentage": 42.55, "elapsed_time": "6:06:33", "remaining_time": "8:14:55"}
|
1663 |
+
{"current_steps": 1663, "total_steps": 3906, "loss": 1.3714, "learning_rate": 2.4985131001484856e-05, "epoch": 0.425728, "percentage": 42.58, "elapsed_time": "6:06:46", "remaining_time": "8:14:41"}
|
1664 |
+
{"current_steps": 1664, "total_steps": 3906, "loss": 1.3628, "learning_rate": 2.4969389905074933e-05, "epoch": 0.425984, "percentage": 42.6, "elapsed_time": "6:06:59", "remaining_time": "8:14:28"}
|
1665 |
+
{"current_steps": 1665, "total_steps": 3906, "loss": 1.4037, "learning_rate": 2.495364552711448e-05, "epoch": 0.42624, "percentage": 42.63, "elapsed_time": "6:07:12", "remaining_time": "8:14:15"}
|
1666 |
+
{"current_steps": 1666, "total_steps": 3906, "loss": 1.3997, "learning_rate": 2.4937897878000342e-05, "epoch": 0.426496, "percentage": 42.65, "elapsed_time": "6:07:26", "remaining_time": "8:14:01"}
|
1667 |
+
{"current_steps": 1667, "total_steps": 3906, "loss": 1.4175, "learning_rate": 2.4922146968131532e-05, "epoch": 0.426752, "percentage": 42.68, "elapsed_time": "6:07:39", "remaining_time": "8:13:48"}
|
1668 |
+
{"current_steps": 1668, "total_steps": 3906, "loss": 1.3648, "learning_rate": 2.4906392807909197e-05, "epoch": 0.427008, "percentage": 42.7, "elapsed_time": "6:07:52", "remaining_time": "8:13:35"}
|
1669 |
+
{"current_steps": 1669, "total_steps": 3906, "loss": 1.378, "learning_rate": 2.489063540773665e-05, "epoch": 0.427264, "percentage": 42.73, "elapsed_time": "6:08:05", "remaining_time": "8:13:21"}
|
1670 |
+
{"current_steps": 1670, "total_steps": 3906, "loss": 1.3753, "learning_rate": 2.4874874778019325e-05, "epoch": 0.42752, "percentage": 42.75, "elapsed_time": "6:08:18", "remaining_time": "8:13:08"}
|
1671 |
+
{"current_steps": 1671, "total_steps": 3906, "loss": 1.3904, "learning_rate": 2.4859110929164804e-05, "epoch": 0.427776, "percentage": 42.78, "elapsed_time": "6:08:32", "remaining_time": "8:12:55"}
|
1672 |
+
{"current_steps": 1672, "total_steps": 3906, "loss": 1.442, "learning_rate": 2.4843343871582782e-05, "epoch": 0.428032, "percentage": 42.81, "elapsed_time": "6:08:45", "remaining_time": "8:12:42"}
|
1673 |
+
{"current_steps": 1673, "total_steps": 3906, "loss": 1.3966, "learning_rate": 2.4827573615685094e-05, "epoch": 0.428288, "percentage": 42.83, "elapsed_time": "6:08:58", "remaining_time": "8:12:29"}
|
1674 |
+
{"current_steps": 1674, "total_steps": 3906, "loss": 1.3697, "learning_rate": 2.4811800171885657e-05, "epoch": 0.428544, "percentage": 42.86, "elapsed_time": "6:09:11", "remaining_time": "8:12:15"}
|
1675 |
+
{"current_steps": 1675, "total_steps": 3906, "loss": 1.3686, "learning_rate": 2.4796023550600514e-05, "epoch": 0.4288, "percentage": 42.88, "elapsed_time": "6:09:24", "remaining_time": "8:12:02"}
|
1676 |
+
{"current_steps": 1676, "total_steps": 3906, "loss": 1.3925, "learning_rate": 2.47802437622478e-05, "epoch": 0.429056, "percentage": 42.91, "elapsed_time": "6:09:38", "remaining_time": "8:11:49"}
|
1677 |
+
{"current_steps": 1677, "total_steps": 3906, "loss": 1.3974, "learning_rate": 2.476446081724775e-05, "epoch": 0.429312, "percentage": 42.93, "elapsed_time": "6:09:51", "remaining_time": "8:11:35"}
|
1678 |
+
{"current_steps": 1678, "total_steps": 3906, "loss": 1.3221, "learning_rate": 2.4748674726022664e-05, "epoch": 0.429568, "percentage": 42.96, "elapsed_time": "6:10:04", "remaining_time": "8:11:22"}
|
1679 |
+
{"current_steps": 1679, "total_steps": 3906, "loss": 1.3095, "learning_rate": 2.4732885498996946e-05, "epoch": 0.429824, "percentage": 42.99, "elapsed_time": "6:10:17", "remaining_time": "8:11:09"}
|
1680 |
+
{"current_steps": 1680, "total_steps": 3906, "loss": 1.388, "learning_rate": 2.471709314659704e-05, "epoch": 0.43008, "percentage": 43.01, "elapsed_time": "6:10:30", "remaining_time": "8:10:55"}
|
1681 |
+
{"current_steps": 1681, "total_steps": 3906, "loss": 1.415, "learning_rate": 2.4701297679251477e-05, "epoch": 0.430336, "percentage": 43.04, "elapsed_time": "6:10:43", "remaining_time": "8:10:42"}
|
1682 |
+
{"current_steps": 1682, "total_steps": 3906, "loss": 1.3379, "learning_rate": 2.4685499107390845e-05, "epoch": 0.430592, "percentage": 43.06, "elapsed_time": "6:10:57", "remaining_time": "8:10:29"}
|
1683 |
+
{"current_steps": 1683, "total_steps": 3906, "loss": 1.3943, "learning_rate": 2.4669697441447765e-05, "epoch": 0.430848, "percentage": 43.09, "elapsed_time": "6:11:10", "remaining_time": "8:10:15"}
|
1684 |
+
{"current_steps": 1684, "total_steps": 3906, "loss": 1.3809, "learning_rate": 2.4653892691856917e-05, "epoch": 0.431104, "percentage": 43.11, "elapsed_time": "6:11:23", "remaining_time": "8:10:02"}
|
1685 |
+
{"current_steps": 1685, "total_steps": 3906, "loss": 1.3692, "learning_rate": 2.463808486905501e-05, "epoch": 0.43136, "percentage": 43.14, "elapsed_time": "6:11:36", "remaining_time": "8:09:49"}
|
1686 |
+
{"current_steps": 1686, "total_steps": 3906, "loss": 1.4103, "learning_rate": 2.4622273983480786e-05, "epoch": 0.431616, "percentage": 43.16, "elapsed_time": "6:11:49", "remaining_time": "8:09:36"}
|
1687 |
+
{"current_steps": 1687, "total_steps": 3906, "loss": 1.4062, "learning_rate": 2.4606460045575e-05, "epoch": 0.431872, "percentage": 43.19, "elapsed_time": "6:12:03", "remaining_time": "8:09:22"}
|
1688 |
+
{"current_steps": 1688, "total_steps": 3906, "loss": 1.3561, "learning_rate": 2.4590643065780442e-05, "epoch": 0.432128, "percentage": 43.22, "elapsed_time": "6:12:16", "remaining_time": "8:09:09"}
|
1689 |
+
{"current_steps": 1689, "total_steps": 3906, "loss": 1.376, "learning_rate": 2.4574823054541888e-05, "epoch": 0.432384, "percentage": 43.24, "elapsed_time": "6:12:29", "remaining_time": "8:08:56"}
|
1690 |
+
{"current_steps": 1690, "total_steps": 3906, "loss": 1.3618, "learning_rate": 2.4559000022306138e-05, "epoch": 0.43264, "percentage": 43.27, "elapsed_time": "6:12:42", "remaining_time": "8:08:42"}
|
1691 |
+
{"current_steps": 1691, "total_steps": 3906, "loss": 1.3738, "learning_rate": 2.4543173979521963e-05, "epoch": 0.432896, "percentage": 43.29, "elapsed_time": "6:12:55", "remaining_time": "8:08:29"}
|
1692 |
+
{"current_steps": 1692, "total_steps": 3906, "loss": 1.3997, "learning_rate": 2.4527344936640144e-05, "epoch": 0.433152, "percentage": 43.32, "elapsed_time": "6:13:08", "remaining_time": "8:08:16"}
|
1693 |
+
{"current_steps": 1693, "total_steps": 3906, "loss": 1.3701, "learning_rate": 2.451151290411342e-05, "epoch": 0.433408, "percentage": 43.34, "elapsed_time": "6:13:22", "remaining_time": "8:08:02"}
|
1694 |
+
{"current_steps": 1694, "total_steps": 3906, "loss": 1.3876, "learning_rate": 2.4495677892396536e-05, "epoch": 0.433664, "percentage": 43.37, "elapsed_time": "6:13:35", "remaining_time": "8:07:49"}
|
1695 |
+
{"current_steps": 1695, "total_steps": 3906, "loss": 1.3846, "learning_rate": 2.447983991194618e-05, "epoch": 0.43392, "percentage": 43.39, "elapsed_time": "6:13:48", "remaining_time": "8:07:36"}
|
1696 |
+
{"current_steps": 1696, "total_steps": 3906, "loss": 1.3401, "learning_rate": 2.4463998973221008e-05, "epoch": 0.434176, "percentage": 43.42, "elapsed_time": "6:14:01", "remaining_time": "8:07:22"}
|
1697 |
+
{"current_steps": 1697, "total_steps": 3906, "loss": 1.3799, "learning_rate": 2.444815508668162e-05, "epoch": 0.434432, "percentage": 43.45, "elapsed_time": "6:14:14", "remaining_time": "8:07:09"}
|
1698 |
+
{"current_steps": 1698, "total_steps": 3906, "loss": 1.3811, "learning_rate": 2.4432308262790578e-05, "epoch": 0.434688, "percentage": 43.47, "elapsed_time": "6:14:27", "remaining_time": "8:06:56"}
|
1699 |
+
{"current_steps": 1699, "total_steps": 3906, "loss": 1.3798, "learning_rate": 2.441645851201237e-05, "epoch": 0.434944, "percentage": 43.5, "elapsed_time": "6:14:41", "remaining_time": "8:06:42"}
|
1700 |
+
{"current_steps": 1700, "total_steps": 3906, "loss": 1.4282, "learning_rate": 2.4400605844813435e-05, "epoch": 0.4352, "percentage": 43.52, "elapsed_time": "6:14:54", "remaining_time": "8:06:29"}
|
1701 |
+
{"current_steps": 1701, "total_steps": 3906, "loss": 1.358, "learning_rate": 2.4384750271662115e-05, "epoch": 0.435456, "percentage": 43.55, "elapsed_time": "6:15:07", "remaining_time": "8:06:16"}
|
1702 |
+
{"current_steps": 1702, "total_steps": 3906, "loss": 1.3855, "learning_rate": 2.4368891803028686e-05, "epoch": 0.435712, "percentage": 43.57, "elapsed_time": "6:15:20", "remaining_time": "8:06:03"}
|
1703 |
+
{"current_steps": 1703, "total_steps": 3906, "loss": 1.402, "learning_rate": 2.435303044938533e-05, "epoch": 0.435968, "percentage": 43.6, "elapsed_time": "6:15:33", "remaining_time": "8:05:49"}
|
1704 |
+
{"current_steps": 1704, "total_steps": 3906, "loss": 1.3923, "learning_rate": 2.433716622120615e-05, "epoch": 0.436224, "percentage": 43.63, "elapsed_time": "6:15:47", "remaining_time": "8:05:36"}
|
1705 |
+
{"current_steps": 1705, "total_steps": 3906, "loss": 1.3442, "learning_rate": 2.432129912896711e-05, "epoch": 0.43648, "percentage": 43.65, "elapsed_time": "6:16:00", "remaining_time": "8:05:23"}
|
1706 |
+
{"current_steps": 1706, "total_steps": 3906, "loss": 1.3597, "learning_rate": 2.4305429183146103e-05, "epoch": 0.436736, "percentage": 43.68, "elapsed_time": "6:16:13", "remaining_time": "8:05:09"}
|
1707 |
+
{"current_steps": 1707, "total_steps": 3906, "loss": 1.3902, "learning_rate": 2.4289556394222898e-05, "epoch": 0.436992, "percentage": 43.7, "elapsed_time": "6:16:26", "remaining_time": "8:04:56"}
|
1708 |
+
{"current_steps": 1708, "total_steps": 3906, "loss": 1.3581, "learning_rate": 2.4273680772679117e-05, "epoch": 0.437248, "percentage": 43.73, "elapsed_time": "6:16:39", "remaining_time": "8:04:43"}
|
1709 |
+
{"current_steps": 1709, "total_steps": 3906, "loss": 1.3445, "learning_rate": 2.4257802328998282e-05, "epoch": 0.437504, "percentage": 43.75, "elapsed_time": "6:16:52", "remaining_time": "8:04:30"}
|
1710 |
+
{"current_steps": 1710, "total_steps": 3906, "loss": 1.4048, "learning_rate": 2.4241921073665774e-05, "epoch": 0.43776, "percentage": 43.78, "elapsed_time": "6:17:06", "remaining_time": "8:04:16"}
|
1711 |
+
{"current_steps": 1711, "total_steps": 3906, "loss": 1.3807, "learning_rate": 2.422603701716881e-05, "epoch": 0.438016, "percentage": 43.8, "elapsed_time": "6:17:19", "remaining_time": "8:04:03"}
|
1712 |
+
{"current_steps": 1712, "total_steps": 3906, "loss": 1.3966, "learning_rate": 2.4210150169996488e-05, "epoch": 0.438272, "percentage": 43.83, "elapsed_time": "6:17:32", "remaining_time": "8:03:50"}
|
1713 |
+
{"current_steps": 1713, "total_steps": 3906, "loss": 1.4274, "learning_rate": 2.4194260542639718e-05, "epoch": 0.438528, "percentage": 43.86, "elapsed_time": "6:17:45", "remaining_time": "8:03:36"}
|
1714 |
+
{"current_steps": 1714, "total_steps": 3906, "loss": 1.4513, "learning_rate": 2.4178368145591263e-05, "epoch": 0.438784, "percentage": 43.88, "elapsed_time": "6:17:58", "remaining_time": "8:03:23"}
|
1715 |
+
{"current_steps": 1715, "total_steps": 3906, "loss": 1.3453, "learning_rate": 2.4162472989345712e-05, "epoch": 0.43904, "percentage": 43.91, "elapsed_time": "6:18:12", "remaining_time": "8:03:10"}
|
1716 |
+
{"current_steps": 1716, "total_steps": 3906, "loss": 1.389, "learning_rate": 2.4146575084399486e-05, "epoch": 0.439296, "percentage": 43.93, "elapsed_time": "6:18:25", "remaining_time": "8:02:56"}
|
1717 |
+
{"current_steps": 1717, "total_steps": 3906, "loss": 1.3979, "learning_rate": 2.4130674441250794e-05, "epoch": 0.439552, "percentage": 43.96, "elapsed_time": "6:18:38", "remaining_time": "8:02:43"}
|
1718 |
+
{"current_steps": 1718, "total_steps": 3906, "loss": 1.3684, "learning_rate": 2.4114771070399682e-05, "epoch": 0.439808, "percentage": 43.98, "elapsed_time": "6:18:51", "remaining_time": "8:02:30"}
|
1719 |
+
{"current_steps": 1719, "total_steps": 3906, "loss": 1.421, "learning_rate": 2.4098864982347987e-05, "epoch": 0.440064, "percentage": 44.01, "elapsed_time": "6:19:04", "remaining_time": "8:02:17"}
|
1720 |
+
{"current_steps": 1720, "total_steps": 3906, "loss": 1.4194, "learning_rate": 2.408295618759932e-05, "epoch": 0.44032, "percentage": 44.03, "elapsed_time": "6:19:17", "remaining_time": "8:02:03"}
|
1721 |
+
{"current_steps": 1721, "total_steps": 3906, "loss": 1.3582, "learning_rate": 2.4067044696659123e-05, "epoch": 0.440576, "percentage": 44.06, "elapsed_time": "6:19:31", "remaining_time": "8:01:50"}
|
1722 |
+
{"current_steps": 1722, "total_steps": 3906, "loss": 1.4082, "learning_rate": 2.4051130520034583e-05, "epoch": 0.440832, "percentage": 44.09, "elapsed_time": "6:19:44", "remaining_time": "8:01:37"}
|
1723 |
+
{"current_steps": 1723, "total_steps": 3906, "loss": 1.3576, "learning_rate": 2.4035213668234665e-05, "epoch": 0.441088, "percentage": 44.11, "elapsed_time": "6:19:57", "remaining_time": "8:01:23"}
|
1724 |
+
{"current_steps": 1724, "total_steps": 3906, "loss": 1.397, "learning_rate": 2.4019294151770112e-05, "epoch": 0.441344, "percentage": 44.14, "elapsed_time": "6:20:10", "remaining_time": "8:01:10"}
|
1725 |
+
{"current_steps": 1725, "total_steps": 3906, "loss": 1.3646, "learning_rate": 2.400337198115343e-05, "epoch": 0.4416, "percentage": 44.16, "elapsed_time": "6:20:23", "remaining_time": "8:00:57"}
|
1726 |
+
{"current_steps": 1726, "total_steps": 3906, "loss": 1.3027, "learning_rate": 2.3987447166898852e-05, "epoch": 0.441856, "percentage": 44.19, "elapsed_time": "6:20:37", "remaining_time": "8:00:43"}
|
1727 |
+
{"current_steps": 1727, "total_steps": 3906, "loss": 1.382, "learning_rate": 2.397151971952238e-05, "epoch": 0.442112, "percentage": 44.21, "elapsed_time": "6:20:50", "remaining_time": "8:00:30"}
|
1728 |
+
{"current_steps": 1728, "total_steps": 3906, "loss": 1.411, "learning_rate": 2.395558964954176e-05, "epoch": 0.442368, "percentage": 44.24, "elapsed_time": "6:21:03", "remaining_time": "8:00:17"}
|
1729 |
+
{"current_steps": 1729, "total_steps": 3906, "loss": 1.3677, "learning_rate": 2.3939656967476438e-05, "epoch": 0.442624, "percentage": 44.27, "elapsed_time": "6:21:16", "remaining_time": "8:00:04"}
|
1730 |
+
{"current_steps": 1730, "total_steps": 3906, "loss": 1.3604, "learning_rate": 2.3923721683847622e-05, "epoch": 0.44288, "percentage": 44.29, "elapsed_time": "6:21:29", "remaining_time": "7:59:50"}
|
1731 |
+
{"current_steps": 1731, "total_steps": 3906, "loss": 1.4061, "learning_rate": 2.3907783809178217e-05, "epoch": 0.443136, "percentage": 44.32, "elapsed_time": "6:21:42", "remaining_time": "7:59:37"}
|
1732 |
+
{"current_steps": 1732, "total_steps": 3906, "loss": 1.3478, "learning_rate": 2.3891843353992834e-05, "epoch": 0.443392, "percentage": 44.34, "elapsed_time": "6:21:55", "remaining_time": "7:59:23"}
|
1733 |
+
{"current_steps": 1733, "total_steps": 3906, "loss": 1.3767, "learning_rate": 2.3875900328817803e-05, "epoch": 0.443648, "percentage": 44.37, "elapsed_time": "6:22:09", "remaining_time": "7:59:10"}
|
1734 |
+
{"current_steps": 1734, "total_steps": 3906, "loss": 1.4088, "learning_rate": 2.3859954744181155e-05, "epoch": 0.443904, "percentage": 44.39, "elapsed_time": "6:22:22", "remaining_time": "7:58:57"}
|
1735 |
+
{"current_steps": 1735, "total_steps": 3906, "loss": 1.4105, "learning_rate": 2.384400661061259e-05, "epoch": 0.44416, "percentage": 44.42, "elapsed_time": "6:22:35", "remaining_time": "7:58:44"}
|
1736 |
+
{"current_steps": 1736, "total_steps": 3906, "loss": 1.3837, "learning_rate": 2.38280559386435e-05, "epoch": 0.444416, "percentage": 44.44, "elapsed_time": "6:22:48", "remaining_time": "7:58:30"}
|
1737 |
+
{"current_steps": 1737, "total_steps": 3906, "loss": 1.3887, "learning_rate": 2.3812102738806972e-05, "epoch": 0.444672, "percentage": 44.47, "elapsed_time": "6:23:01", "remaining_time": "7:58:17"}
|
1738 |
+
{"current_steps": 1738, "total_steps": 3906, "loss": 1.3768, "learning_rate": 2.3796147021637727e-05, "epoch": 0.444928, "percentage": 44.5, "elapsed_time": "6:23:15", "remaining_time": "7:58:04"}
|
1739 |
+
{"current_steps": 1739, "total_steps": 3906, "loss": 1.3837, "learning_rate": 2.3780188797672174e-05, "epoch": 0.445184, "percentage": 44.52, "elapsed_time": "6:23:28", "remaining_time": "7:57:51"}
|
1740 |
+
{"current_steps": 1740, "total_steps": 3906, "loss": 1.3211, "learning_rate": 2.376422807744838e-05, "epoch": 0.44544, "percentage": 44.55, "elapsed_time": "6:23:41", "remaining_time": "7:57:37"}
|
1741 |
+
{"current_steps": 1741, "total_steps": 3906, "loss": 1.3819, "learning_rate": 2.3748264871506046e-05, "epoch": 0.445696, "percentage": 44.57, "elapsed_time": "6:23:54", "remaining_time": "7:57:24"}
|
1742 |
+
{"current_steps": 1742, "total_steps": 3906, "loss": 1.3866, "learning_rate": 2.373229919038651e-05, "epoch": 0.445952, "percentage": 44.6, "elapsed_time": "6:24:07", "remaining_time": "7:57:11"}
|
1743 |
+
{"current_steps": 1743, "total_steps": 3906, "loss": 1.3869, "learning_rate": 2.371633104463277e-05, "epoch": 0.446208, "percentage": 44.62, "elapsed_time": "6:24:21", "remaining_time": "7:56:57"}
|
1744 |
+
{"current_steps": 1744, "total_steps": 3906, "loss": 1.3628, "learning_rate": 2.370036044478942e-05, "epoch": 0.446464, "percentage": 44.65, "elapsed_time": "6:24:34", "remaining_time": "7:56:44"}
|
1745 |
+
{"current_steps": 1745, "total_steps": 3906, "loss": 1.3705, "learning_rate": 2.36843874014027e-05, "epoch": 0.44672, "percentage": 44.67, "elapsed_time": "6:24:47", "remaining_time": "7:56:31"}
|
1746 |
+
{"current_steps": 1746, "total_steps": 3906, "loss": 1.4091, "learning_rate": 2.3668411925020455e-05, "epoch": 0.446976, "percentage": 44.7, "elapsed_time": "6:25:00", "remaining_time": "7:56:18"}
|
1747 |
+
{"current_steps": 1747, "total_steps": 3906, "loss": 1.3655, "learning_rate": 2.3652434026192133e-05, "epoch": 0.447232, "percentage": 44.73, "elapsed_time": "6:25:13", "remaining_time": "7:56:04"}
|
1748 |
+
{"current_steps": 1748, "total_steps": 3906, "loss": 1.4011, "learning_rate": 2.363645371546878e-05, "epoch": 0.447488, "percentage": 44.75, "elapsed_time": "6:25:26", "remaining_time": "7:55:51"}
|
1749 |
+
{"current_steps": 1749, "total_steps": 3906, "loss": 1.3843, "learning_rate": 2.3620471003403042e-05, "epoch": 0.447744, "percentage": 44.78, "elapsed_time": "6:25:40", "remaining_time": "7:55:38"}
|
1750 |
+
{"current_steps": 1750, "total_steps": 3906, "loss": 1.3105, "learning_rate": 2.360448590054915e-05, "epoch": 0.448, "percentage": 44.8, "elapsed_time": "6:25:53", "remaining_time": "7:55:24"}
|
1751 |
+
{"current_steps": 1751, "total_steps": 3906, "loss": 1.3721, "learning_rate": 2.358849841746291e-05, "epoch": 0.448256, "percentage": 44.83, "elapsed_time": "6:26:06", "remaining_time": "7:55:11"}
|
1752 |
+
{"current_steps": 1752, "total_steps": 3906, "loss": 1.4077, "learning_rate": 2.35725085647017e-05, "epoch": 0.448512, "percentage": 44.85, "elapsed_time": "6:26:19", "remaining_time": "7:54:58"}
|
1753 |
+
{"current_steps": 1753, "total_steps": 3906, "loss": 1.4041, "learning_rate": 2.3556516352824463e-05, "epoch": 0.448768, "percentage": 44.88, "elapsed_time": "6:26:32", "remaining_time": "7:54:44"}
|
1754 |
+
{"current_steps": 1754, "total_steps": 3906, "loss": 1.3843, "learning_rate": 2.3540521792391702e-05, "epoch": 0.449024, "percentage": 44.91, "elapsed_time": "6:26:45", "remaining_time": "7:54:31"}
|
1755 |
+
{"current_steps": 1755, "total_steps": 3906, "loss": 1.3633, "learning_rate": 2.352452489396547e-05, "epoch": 0.44928, "percentage": 44.93, "elapsed_time": "6:26:59", "remaining_time": "7:54:18"}
|
1756 |
+
{"current_steps": 1756, "total_steps": 3906, "loss": 1.3413, "learning_rate": 2.3508525668109372e-05, "epoch": 0.449536, "percentage": 44.96, "elapsed_time": "6:27:12", "remaining_time": "7:54:05"}
|
1757 |
+
{"current_steps": 1757, "total_steps": 3906, "loss": 1.3797, "learning_rate": 2.3492524125388528e-05, "epoch": 0.449792, "percentage": 44.98, "elapsed_time": "6:27:25", "remaining_time": "7:53:51"}
|
1758 |
+
{"current_steps": 1758, "total_steps": 3906, "loss": 1.395, "learning_rate": 2.3476520276369604e-05, "epoch": 0.450048, "percentage": 45.01, "elapsed_time": "6:27:38", "remaining_time": "7:53:38"}
|
1759 |
+
{"current_steps": 1759, "total_steps": 3906, "loss": 1.3687, "learning_rate": 2.3460514131620794e-05, "epoch": 0.450304, "percentage": 45.03, "elapsed_time": "6:27:51", "remaining_time": "7:53:25"}
|
1760 |
+
{"current_steps": 1760, "total_steps": 3906, "loss": 1.3633, "learning_rate": 2.3444505701711797e-05, "epoch": 0.45056, "percentage": 45.06, "elapsed_time": "6:28:05", "remaining_time": "7:53:11"}
|
1761 |
+
{"current_steps": 1761, "total_steps": 3906, "loss": 1.4106, "learning_rate": 2.3428494997213815e-05, "epoch": 0.450816, "percentage": 45.08, "elapsed_time": "6:28:18", "remaining_time": "7:52:58"}
|
1762 |
+
{"current_steps": 1762, "total_steps": 3906, "loss": 1.37, "learning_rate": 2.3412482028699578e-05, "epoch": 0.451072, "percentage": 45.11, "elapsed_time": "6:28:31", "remaining_time": "7:52:45"}
|
1763 |
+
{"current_steps": 1763, "total_steps": 3906, "loss": 1.393, "learning_rate": 2.339646680674329e-05, "epoch": 0.451328, "percentage": 45.14, "elapsed_time": "6:28:44", "remaining_time": "7:52:32"}
|
1764 |
+
{"current_steps": 1764, "total_steps": 3906, "loss": 1.3641, "learning_rate": 2.3380449341920636e-05, "epoch": 0.451584, "percentage": 45.16, "elapsed_time": "6:28:57", "remaining_time": "7:52:18"}
|
1765 |
+
{"current_steps": 1765, "total_steps": 3906, "loss": 1.408, "learning_rate": 2.33644296448088e-05, "epoch": 0.45184, "percentage": 45.19, "elapsed_time": "6:29:10", "remaining_time": "7:52:05"}
|
1766 |
+
{"current_steps": 1766, "total_steps": 3906, "loss": 1.4023, "learning_rate": 2.334840772598644e-05, "epoch": 0.452096, "percentage": 45.21, "elapsed_time": "6:29:24", "remaining_time": "7:51:52"}
|
1767 |
+
{"current_steps": 1767, "total_steps": 3906, "loss": 1.3824, "learning_rate": 2.3332383596033667e-05, "epoch": 0.452352, "percentage": 45.24, "elapsed_time": "6:29:37", "remaining_time": "7:51:38"}
|
1768 |
+
{"current_steps": 1768, "total_steps": 3906, "loss": 1.3809, "learning_rate": 2.331635726553207e-05, "epoch": 0.452608, "percentage": 45.26, "elapsed_time": "6:29:50", "remaining_time": "7:51:25"}
|
1769 |
+
{"current_steps": 1769, "total_steps": 3906, "loss": 1.378, "learning_rate": 2.3300328745064676e-05, "epoch": 0.452864, "percentage": 45.29, "elapsed_time": "6:30:03", "remaining_time": "7:51:12"}
|
1770 |
+
{"current_steps": 1770, "total_steps": 3906, "loss": 1.4255, "learning_rate": 2.328429804521595e-05, "epoch": 0.45312, "percentage": 45.31, "elapsed_time": "6:30:16", "remaining_time": "7:50:59"}
|
1771 |
+
{"current_steps": 1771, "total_steps": 3906, "loss": 1.3837, "learning_rate": 2.326826517657183e-05, "epoch": 0.453376, "percentage": 45.34, "elapsed_time": "6:30:30", "remaining_time": "7:50:45"}
|
1772 |
+
{"current_steps": 1772, "total_steps": 3906, "loss": 1.3953, "learning_rate": 2.325223014971966e-05, "epoch": 0.453632, "percentage": 45.37, "elapsed_time": "6:30:43", "remaining_time": "7:50:32"}
|
1773 |
+
{"current_steps": 1773, "total_steps": 3906, "loss": 1.3593, "learning_rate": 2.3236192975248204e-05, "epoch": 0.453888, "percentage": 45.39, "elapsed_time": "6:30:56", "remaining_time": "7:50:19"}
|
1774 |
+
{"current_steps": 1774, "total_steps": 3906, "loss": 1.3864, "learning_rate": 2.3220153663747675e-05, "epoch": 0.454144, "percentage": 45.42, "elapsed_time": "6:31:09", "remaining_time": "7:50:05"}
|
1775 |
+
{"current_steps": 1775, "total_steps": 3906, "loss": 1.3559, "learning_rate": 2.3204112225809653e-05, "epoch": 0.4544, "percentage": 45.44, "elapsed_time": "6:31:22", "remaining_time": "7:49:52"}
|
1776 |
+
{"current_steps": 1776, "total_steps": 3906, "loss": 1.3656, "learning_rate": 2.3188068672027163e-05, "epoch": 0.454656, "percentage": 45.47, "elapsed_time": "6:31:35", "remaining_time": "7:49:39"}
|
1777 |
+
{"current_steps": 1777, "total_steps": 3906, "loss": 1.3833, "learning_rate": 2.3172023012994608e-05, "epoch": 0.454912, "percentage": 45.49, "elapsed_time": "6:31:48", "remaining_time": "7:49:25"}
|
1778 |
+
{"current_steps": 1778, "total_steps": 3906, "loss": 1.3408, "learning_rate": 2.3155975259307778e-05, "epoch": 0.455168, "percentage": 45.52, "elapsed_time": "6:32:02", "remaining_time": "7:49:12"}
|
1779 |
+
{"current_steps": 1779, "total_steps": 3906, "loss": 1.3417, "learning_rate": 2.3139925421563863e-05, "epoch": 0.455424, "percentage": 45.55, "elapsed_time": "6:32:15", "remaining_time": "7:48:59"}
|
1780 |
+
{"current_steps": 1780, "total_steps": 3906, "loss": 1.3441, "learning_rate": 2.312387351036141e-05, "epoch": 0.45568, "percentage": 45.57, "elapsed_time": "6:32:28", "remaining_time": "7:48:45"}
|
1781 |
+
{"current_steps": 1781, "total_steps": 3906, "loss": 1.3752, "learning_rate": 2.310781953630034e-05, "epoch": 0.455936, "percentage": 45.6, "elapsed_time": "6:32:41", "remaining_time": "7:48:32"}
|
1782 |
+
{"current_steps": 1782, "total_steps": 3906, "loss": 1.3558, "learning_rate": 2.3091763509981946e-05, "epoch": 0.456192, "percentage": 45.62, "elapsed_time": "6:32:54", "remaining_time": "7:48:19"}
|
1783 |
+
{"current_steps": 1783, "total_steps": 3906, "loss": 1.3719, "learning_rate": 2.3075705442008874e-05, "epoch": 0.456448, "percentage": 45.65, "elapsed_time": "6:33:08", "remaining_time": "7:48:06"}
|
1784 |
+
{"current_steps": 1784, "total_steps": 3906, "loss": 1.3435, "learning_rate": 2.30596453429851e-05, "epoch": 0.456704, "percentage": 45.67, "elapsed_time": "6:33:21", "remaining_time": "7:47:53"}
|
1785 |
+
{"current_steps": 1785, "total_steps": 3906, "loss": 1.4077, "learning_rate": 2.304358322351597e-05, "epoch": 0.45696, "percentage": 45.7, "elapsed_time": "6:33:34", "remaining_time": "7:47:39"}
|
1786 |
+
{"current_steps": 1786, "total_steps": 3906, "loss": 1.3475, "learning_rate": 2.3027519094208148e-05, "epoch": 0.457216, "percentage": 45.72, "elapsed_time": "6:33:47", "remaining_time": "7:47:26"}
|
1787 |
+
{"current_steps": 1787, "total_steps": 3906, "loss": 1.4272, "learning_rate": 2.301145296566961e-05, "epoch": 0.457472, "percentage": 45.75, "elapsed_time": "6:34:00", "remaining_time": "7:47:13"}
|
1788 |
+
{"current_steps": 1788, "total_steps": 3906, "loss": 1.3644, "learning_rate": 2.2995384848509684e-05, "epoch": 0.457728, "percentage": 45.78, "elapsed_time": "6:34:14", "remaining_time": "7:46:59"}
|
1789 |
+
{"current_steps": 1789, "total_steps": 3906, "loss": 1.3725, "learning_rate": 2.2979314753339e-05, "epoch": 0.457984, "percentage": 45.8, "elapsed_time": "6:34:27", "remaining_time": "7:46:46"}
|
1790 |
+
{"current_steps": 1790, "total_steps": 3906, "loss": 1.3867, "learning_rate": 2.296324269076948e-05, "epoch": 0.45824, "percentage": 45.83, "elapsed_time": "6:34:40", "remaining_time": "7:46:33"}
|
1791 |
+
{"current_steps": 1791, "total_steps": 3906, "loss": 1.3811, "learning_rate": 2.294716867141436e-05, "epoch": 0.458496, "percentage": 45.85, "elapsed_time": "6:34:53", "remaining_time": "7:46:20"}
|
1792 |
+
{"current_steps": 1792, "total_steps": 3906, "loss": 1.4255, "learning_rate": 2.2931092705888167e-05, "epoch": 0.458752, "percentage": 45.88, "elapsed_time": "6:35:07", "remaining_time": "7:46:06"}
|
1793 |
+
{"current_steps": 1793, "total_steps": 3906, "loss": 1.34, "learning_rate": 2.2915014804806693e-05, "epoch": 0.459008, "percentage": 45.9, "elapsed_time": "6:35:20", "remaining_time": "7:45:53"}
|
1794 |
+
{"current_steps": 1794, "total_steps": 3906, "loss": 1.3865, "learning_rate": 2.289893497878705e-05, "epoch": 0.459264, "percentage": 45.93, "elapsed_time": "6:35:33", "remaining_time": "7:45:40"}
|
1795 |
+
{"current_steps": 1795, "total_steps": 3906, "loss": 1.3564, "learning_rate": 2.2882853238447576e-05, "epoch": 0.45952, "percentage": 45.95, "elapsed_time": "6:35:46", "remaining_time": "7:45:27"}
|
1796 |
+
{"current_steps": 1796, "total_steps": 3906, "loss": 1.3872, "learning_rate": 2.28667695944079e-05, "epoch": 0.459776, "percentage": 45.98, "elapsed_time": "6:35:59", "remaining_time": "7:45:13"}
|
1797 |
+
{"current_steps": 1797, "total_steps": 3906, "loss": 1.3826, "learning_rate": 2.285068405728891e-05, "epoch": 0.460032, "percentage": 46.01, "elapsed_time": "6:36:12", "remaining_time": "7:45:00"}
|
1798 |
+
{"current_steps": 1798, "total_steps": 3906, "loss": 1.3999, "learning_rate": 2.2834596637712724e-05, "epoch": 0.460288, "percentage": 46.03, "elapsed_time": "6:36:26", "remaining_time": "7:44:47"}
|
1799 |
+
{"current_steps": 1799, "total_steps": 3906, "loss": 1.3887, "learning_rate": 2.281850734630272e-05, "epoch": 0.460544, "percentage": 46.06, "elapsed_time": "6:36:39", "remaining_time": "7:44:33"}
|
1800 |
+
{"current_steps": 1800, "total_steps": 3906, "loss": 1.3663, "learning_rate": 2.28024161936835e-05, "epoch": 0.4608, "percentage": 46.08, "elapsed_time": "6:36:52", "remaining_time": "7:44:20"}
|
1801 |
+
{"current_steps": 1801, "total_steps": 3906, "loss": 1.3218, "learning_rate": 2.2786323190480907e-05, "epoch": 0.461056, "percentage": 46.11, "elapsed_time": "6:37:05", "remaining_time": "7:44:07"}
|
1802 |
+
{"current_steps": 1802, "total_steps": 3906, "loss": 1.3694, "learning_rate": 2.2770228347322005e-05, "epoch": 0.461312, "percentage": 46.13, "elapsed_time": "6:37:18", "remaining_time": "7:43:54"}
|
1803 |
+
{"current_steps": 1803, "total_steps": 3906, "loss": 1.3591, "learning_rate": 2.2754131674835057e-05, "epoch": 0.461568, "percentage": 46.16, "elapsed_time": "6:37:32", "remaining_time": "7:43:40"}
|
1804 |
+
{"current_steps": 1804, "total_steps": 3906, "loss": 1.4035, "learning_rate": 2.273803318364956e-05, "epoch": 0.461824, "percentage": 46.19, "elapsed_time": "6:37:45", "remaining_time": "7:43:27"}
|
1805 |
+
{"current_steps": 1805, "total_steps": 3906, "loss": 1.3387, "learning_rate": 2.272193288439618e-05, "epoch": 0.46208, "percentage": 46.21, "elapsed_time": "6:37:58", "remaining_time": "7:43:14"}
|
1806 |
+
{"current_steps": 1806, "total_steps": 3906, "loss": 1.3106, "learning_rate": 2.270583078770681e-05, "epoch": 0.462336, "percentage": 46.24, "elapsed_time": "6:38:11", "remaining_time": "7:43:01"}
|
1807 |
+
{"current_steps": 1807, "total_steps": 3906, "loss": 1.3806, "learning_rate": 2.2689726904214514e-05, "epoch": 0.462592, "percentage": 46.26, "elapsed_time": "6:38:24", "remaining_time": "7:42:47"}
|
1808 |
+
{"current_steps": 1808, "total_steps": 3906, "loss": 1.3552, "learning_rate": 2.2673621244553526e-05, "epoch": 0.462848, "percentage": 46.29, "elapsed_time": "6:38:38", "remaining_time": "7:42:34"}
|
1809 |
+
{"current_steps": 1809, "total_steps": 3906, "loss": 1.3692, "learning_rate": 2.2657513819359273e-05, "epoch": 0.463104, "percentage": 46.31, "elapsed_time": "6:38:51", "remaining_time": "7:42:21"}
|
1810 |
+
{"current_steps": 1810, "total_steps": 3906, "loss": 1.397, "learning_rate": 2.2641404639268333e-05, "epoch": 0.46336, "percentage": 46.34, "elapsed_time": "6:39:04", "remaining_time": "7:42:07"}
|
1811 |
+
{"current_steps": 1811, "total_steps": 3906, "loss": 1.3409, "learning_rate": 2.262529371491845e-05, "epoch": 0.463616, "percentage": 46.36, "elapsed_time": "6:39:17", "remaining_time": "7:41:54"}
|
1812 |
+
{"current_steps": 1812, "total_steps": 3906, "loss": 1.3821, "learning_rate": 2.2609181056948514e-05, "epoch": 0.463872, "percentage": 46.39, "elapsed_time": "6:39:30", "remaining_time": "7:41:41"}
|
1813 |
+
{"current_steps": 1813, "total_steps": 3906, "loss": 1.3788, "learning_rate": 2.2593066675998573e-05, "epoch": 0.464128, "percentage": 46.42, "elapsed_time": "6:39:43", "remaining_time": "7:41:27"}
|
1814 |
+
{"current_steps": 1814, "total_steps": 3906, "loss": 1.3963, "learning_rate": 2.2576950582709798e-05, "epoch": 0.464384, "percentage": 46.44, "elapsed_time": "6:39:57", "remaining_time": "7:41:14"}
|
1815 |
+
{"current_steps": 1815, "total_steps": 3906, "loss": 1.4016, "learning_rate": 2.256083278772449e-05, "epoch": 0.46464, "percentage": 46.47, "elapsed_time": "6:40:10", "remaining_time": "7:41:01"}
|
1816 |
+
{"current_steps": 1816, "total_steps": 3906, "loss": 1.3973, "learning_rate": 2.2544713301686095e-05, "epoch": 0.464896, "percentage": 46.49, "elapsed_time": "6:40:23", "remaining_time": "7:40:48"}
|
1817 |
+
{"current_steps": 1817, "total_steps": 3906, "loss": 1.3817, "learning_rate": 2.252859213523915e-05, "epoch": 0.465152, "percentage": 46.52, "elapsed_time": "6:40:36", "remaining_time": "7:40:34"}
|
1818 |
+
{"current_steps": 1818, "total_steps": 3906, "loss": 1.3936, "learning_rate": 2.2512469299029305e-05, "epoch": 0.465408, "percentage": 46.54, "elapsed_time": "6:40:49", "remaining_time": "7:40:21"}
|
1819 |
+
{"current_steps": 1819, "total_steps": 3906, "loss": 1.3946, "learning_rate": 2.249634480370334e-05, "epoch": 0.465664, "percentage": 46.57, "elapsed_time": "6:41:02", "remaining_time": "7:40:08"}
|
1820 |
+
{"current_steps": 1820, "total_steps": 3906, "loss": 1.3553, "learning_rate": 2.24802186599091e-05, "epoch": 0.46592, "percentage": 46.59, "elapsed_time": "6:41:16", "remaining_time": "7:39:54"}
|
1821 |
+
{"current_steps": 1821, "total_steps": 3906, "loss": 1.3999, "learning_rate": 2.246409087829552e-05, "epoch": 0.466176, "percentage": 46.62, "elapsed_time": "6:41:29", "remaining_time": "7:39:41"}
|
1822 |
+
{"current_steps": 1822, "total_steps": 3906, "loss": 1.3506, "learning_rate": 2.244796146951264e-05, "epoch": 0.466432, "percentage": 46.65, "elapsed_time": "6:41:42", "remaining_time": "7:39:28"}
|
1823 |
+
{"current_steps": 1823, "total_steps": 3906, "loss": 1.3772, "learning_rate": 2.243183044421156e-05, "epoch": 0.466688, "percentage": 46.67, "elapsed_time": "6:41:55", "remaining_time": "7:39:15"}
|
1824 |
+
{"current_steps": 1824, "total_steps": 3906, "loss": 1.394, "learning_rate": 2.2415697813044437e-05, "epoch": 0.466944, "percentage": 46.7, "elapsed_time": "6:42:08", "remaining_time": "7:39:01"}
|
1825 |
+
{"current_steps": 1825, "total_steps": 3906, "loss": 1.3555, "learning_rate": 2.2399563586664506e-05, "epoch": 0.4672, "percentage": 46.72, "elapsed_time": "6:42:22", "remaining_time": "7:38:48"}
|
1826 |
+
{"current_steps": 1826, "total_steps": 3906, "loss": 1.3721, "learning_rate": 2.2383427775726048e-05, "epoch": 0.467456, "percentage": 46.75, "elapsed_time": "6:42:35", "remaining_time": "7:38:35"}
|
1827 |
+
{"current_steps": 1827, "total_steps": 3906, "loss": 1.3459, "learning_rate": 2.2367290390884388e-05, "epoch": 0.467712, "percentage": 46.77, "elapsed_time": "6:42:48", "remaining_time": "7:38:22"}
|
1828 |
+
{"current_steps": 1828, "total_steps": 3906, "loss": 1.3884, "learning_rate": 2.23511514427959e-05, "epoch": 0.467968, "percentage": 46.8, "elapsed_time": "6:43:01", "remaining_time": "7:38:08"}
|
1829 |
+
{"current_steps": 1829, "total_steps": 3906, "loss": 1.3986, "learning_rate": 2.2335010942117975e-05, "epoch": 0.468224, "percentage": 46.83, "elapsed_time": "6:43:14", "remaining_time": "7:37:55"}
|
1830 |
+
{"current_steps": 1830, "total_steps": 3906, "loss": 1.382, "learning_rate": 2.2318868899509044e-05, "epoch": 0.46848, "percentage": 46.85, "elapsed_time": "6:43:28", "remaining_time": "7:37:42"}
|
1831 |
+
{"current_steps": 1831, "total_steps": 3906, "loss": 1.4032, "learning_rate": 2.2302725325628546e-05, "epoch": 0.468736, "percentage": 46.88, "elapsed_time": "6:43:41", "remaining_time": "7:37:29"}
|
1832 |
+
{"current_steps": 1832, "total_steps": 3906, "loss": 1.3506, "learning_rate": 2.2286580231136944e-05, "epoch": 0.468992, "percentage": 46.9, "elapsed_time": "6:43:54", "remaining_time": "7:37:15"}
|
1833 |
+
{"current_steps": 1833, "total_steps": 3906, "loss": 1.3476, "learning_rate": 2.227043362669568e-05, "epoch": 0.469248, "percentage": 46.93, "elapsed_time": "6:44:07", "remaining_time": "7:37:02"}
|
1834 |
+
{"current_steps": 1834, "total_steps": 3906, "loss": 1.363, "learning_rate": 2.2254285522967222e-05, "epoch": 0.469504, "percentage": 46.95, "elapsed_time": "6:44:20", "remaining_time": "7:36:49"}
|
1835 |
+
{"current_steps": 1835, "total_steps": 3906, "loss": 1.3761, "learning_rate": 2.2238135930615013e-05, "epoch": 0.46976, "percentage": 46.98, "elapsed_time": "6:44:34", "remaining_time": "7:36:35"}
|
1836 |
+
{"current_steps": 1836, "total_steps": 3906, "loss": 1.3248, "learning_rate": 2.222198486030348e-05, "epoch": 0.470016, "percentage": 47.0, "elapsed_time": "6:44:47", "remaining_time": "7:36:22"}
|
1837 |
+
{"current_steps": 1837, "total_steps": 3906, "loss": 1.3613, "learning_rate": 2.2205832322698032e-05, "epoch": 0.470272, "percentage": 47.03, "elapsed_time": "6:45:00", "remaining_time": "7:36:09"}
|
1838 |
+
{"current_steps": 1838, "total_steps": 3906, "loss": 1.3583, "learning_rate": 2.2189678328465037e-05, "epoch": 0.470528, "percentage": 47.06, "elapsed_time": "6:45:13", "remaining_time": "7:35:56"}
|
1839 |
+
{"current_steps": 1839, "total_steps": 3906, "loss": 1.3427, "learning_rate": 2.2173522888271828e-05, "epoch": 0.470784, "percentage": 47.08, "elapsed_time": "6:45:26", "remaining_time": "7:35:42"}
|
1840 |
+
{"current_steps": 1840, "total_steps": 3906, "loss": 1.4158, "learning_rate": 2.2157366012786697e-05, "epoch": 0.47104, "percentage": 47.11, "elapsed_time": "6:45:39", "remaining_time": "7:35:29"}
|
1841 |
+
{"current_steps": 1841, "total_steps": 3906, "loss": 1.3519, "learning_rate": 2.2141207712678892e-05, "epoch": 0.471296, "percentage": 47.13, "elapsed_time": "6:45:53", "remaining_time": "7:35:16"}
|
1842 |
+
{"current_steps": 1842, "total_steps": 3906, "loss": 1.417, "learning_rate": 2.2125047998618585e-05, "epoch": 0.471552, "percentage": 47.16, "elapsed_time": "6:46:06", "remaining_time": "7:35:03"}
|
1843 |
+
{"current_steps": 1843, "total_steps": 3906, "loss": 1.4381, "learning_rate": 2.2108886881276876e-05, "epoch": 0.471808, "percentage": 47.18, "elapsed_time": "6:46:19", "remaining_time": "7:34:49"}
|
1844 |
+
{"current_steps": 1844, "total_steps": 3906, "loss": 1.3945, "learning_rate": 2.2092724371325822e-05, "epoch": 0.472064, "percentage": 47.21, "elapsed_time": "6:46:32", "remaining_time": "7:34:36"}
|
1845 |
+
{"current_steps": 1845, "total_steps": 3906, "loss": 1.3791, "learning_rate": 2.207656047943837e-05, "epoch": 0.47232, "percentage": 47.24, "elapsed_time": "6:46:46", "remaining_time": "7:34:23"}
|
1846 |
+
{"current_steps": 1846, "total_steps": 3906, "loss": 1.4425, "learning_rate": 2.20603952162884e-05, "epoch": 0.472576, "percentage": 47.26, "elapsed_time": "6:46:59", "remaining_time": "7:34:09"}
|
1847 |
+
{"current_steps": 1847, "total_steps": 3906, "loss": 1.3741, "learning_rate": 2.2044228592550692e-05, "epoch": 0.472832, "percentage": 47.29, "elapsed_time": "6:47:12", "remaining_time": "7:33:56"}
|
1848 |
+
{"current_steps": 1848, "total_steps": 3906, "loss": 1.3422, "learning_rate": 2.2028060618900904e-05, "epoch": 0.473088, "percentage": 47.31, "elapsed_time": "6:47:25", "remaining_time": "7:33:43"}
|
1849 |
+
{"current_steps": 1849, "total_steps": 3906, "loss": 1.3301, "learning_rate": 2.2011891306015613e-05, "epoch": 0.473344, "percentage": 47.34, "elapsed_time": "6:47:38", "remaining_time": "7:33:30"}
|
1850 |
+
{"current_steps": 1850, "total_steps": 3906, "loss": 1.4056, "learning_rate": 2.1995720664572274e-05, "epoch": 0.4736, "percentage": 47.36, "elapsed_time": "6:47:51", "remaining_time": "7:33:16"}
|
1851 |
+
{"current_steps": 1851, "total_steps": 3906, "loss": 1.337, "learning_rate": 2.1979548705249207e-05, "epoch": 0.473856, "percentage": 47.39, "elapsed_time": "6:48:05", "remaining_time": "7:33:03"}
|
1852 |
+
{"current_steps": 1852, "total_steps": 3906, "loss": 1.4087, "learning_rate": 2.1963375438725616e-05, "epoch": 0.474112, "percentage": 47.41, "elapsed_time": "6:48:18", "remaining_time": "7:32:50"}
|
1853 |
+
{"current_steps": 1853, "total_steps": 3906, "loss": 1.3276, "learning_rate": 2.1947200875681562e-05, "epoch": 0.474368, "percentage": 47.44, "elapsed_time": "6:48:31", "remaining_time": "7:32:37"}
|
1854 |
+
{"current_steps": 1854, "total_steps": 3906, "loss": 1.3708, "learning_rate": 2.193102502679796e-05, "epoch": 0.474624, "percentage": 47.47, "elapsed_time": "6:48:44", "remaining_time": "7:32:23"}
|
1855 |
+
{"current_steps": 1855, "total_steps": 3906, "loss": 1.3561, "learning_rate": 2.1914847902756576e-05, "epoch": 0.47488, "percentage": 47.49, "elapsed_time": "6:48:57", "remaining_time": "7:32:10"}
|
1856 |
+
{"current_steps": 1856, "total_steps": 3906, "loss": 1.4004, "learning_rate": 2.1898669514240027e-05, "epoch": 0.475136, "percentage": 47.52, "elapsed_time": "6:49:11", "remaining_time": "7:31:57"}
|
1857 |
+
{"current_steps": 1857, "total_steps": 3906, "loss": 1.348, "learning_rate": 2.188248987193175e-05, "epoch": 0.475392, "percentage": 47.54, "elapsed_time": "6:49:24", "remaining_time": "7:31:44"}
|
1858 |
+
{"current_steps": 1858, "total_steps": 3906, "loss": 1.4026, "learning_rate": 2.1866308986516024e-05, "epoch": 0.475648, "percentage": 47.57, "elapsed_time": "6:49:37", "remaining_time": "7:31:30"}
|
1859 |
+
{"current_steps": 1859, "total_steps": 3906, "loss": 1.3524, "learning_rate": 2.185012686867794e-05, "epoch": 0.475904, "percentage": 47.59, "elapsed_time": "6:49:50", "remaining_time": "7:31:17"}
|
1860 |
+
{"current_steps": 1860, "total_steps": 3906, "loss": 1.3222, "learning_rate": 2.183394352910339e-05, "epoch": 0.47616, "percentage": 47.62, "elapsed_time": "6:50:03", "remaining_time": "7:31:04"}
|
1861 |
+
{"current_steps": 1861, "total_steps": 3906, "loss": 1.3663, "learning_rate": 2.181775897847911e-05, "epoch": 0.476416, "percentage": 47.64, "elapsed_time": "6:50:17", "remaining_time": "7:30:50"}
|
1862 |
+
{"current_steps": 1862, "total_steps": 3906, "loss": 1.3772, "learning_rate": 2.18015732274926e-05, "epoch": 0.476672, "percentage": 47.67, "elapsed_time": "6:50:30", "remaining_time": "7:30:37"}
|
1863 |
+
{"current_steps": 1863, "total_steps": 3906, "loss": 1.3503, "learning_rate": 2.1785386286832168e-05, "epoch": 0.476928, "percentage": 47.7, "elapsed_time": "6:50:43", "remaining_time": "7:30:24"}
|
1864 |
+
{"current_steps": 1864, "total_steps": 3906, "loss": 1.3254, "learning_rate": 2.176919816718691e-05, "epoch": 0.477184, "percentage": 47.72, "elapsed_time": "6:50:56", "remaining_time": "7:30:11"}
|
1865 |
+
{"current_steps": 1865, "total_steps": 3906, "loss": 1.374, "learning_rate": 2.175300887924669e-05, "epoch": 0.47744, "percentage": 47.75, "elapsed_time": "6:51:09", "remaining_time": "7:29:57"}
|
1866 |
+
{"current_steps": 1866, "total_steps": 3906, "loss": 1.3544, "learning_rate": 2.1736818433702154e-05, "epoch": 0.477696, "percentage": 47.77, "elapsed_time": "6:51:22", "remaining_time": "7:29:44"}
|
1867 |
+
{"current_steps": 1867, "total_steps": 3906, "loss": 1.3532, "learning_rate": 2.17206268412447e-05, "epoch": 0.477952, "percentage": 47.8, "elapsed_time": "6:51:36", "remaining_time": "7:29:31"}
|
1868 |
+
{"current_steps": 1868, "total_steps": 3906, "loss": 1.3726, "learning_rate": 2.170443411256651e-05, "epoch": 0.478208, "percentage": 47.82, "elapsed_time": "6:51:49", "remaining_time": "7:29:17"}
|
1869 |
+
{"current_steps": 1869, "total_steps": 3906, "loss": 1.341, "learning_rate": 2.168824025836047e-05, "epoch": 0.478464, "percentage": 47.85, "elapsed_time": "6:52:02", "remaining_time": "7:29:04"}
|
1870 |
+
{"current_steps": 1870, "total_steps": 3906, "loss": 1.367, "learning_rate": 2.1672045289320266e-05, "epoch": 0.47872, "percentage": 47.88, "elapsed_time": "6:52:15", "remaining_time": "7:28:51"}
|
1871 |
+
{"current_steps": 1871, "total_steps": 3906, "loss": 1.3279, "learning_rate": 2.1655849216140275e-05, "epoch": 0.478976, "percentage": 47.9, "elapsed_time": "6:52:28", "remaining_time": "7:28:38"}
|
1872 |
+
{"current_steps": 1872, "total_steps": 3906, "loss": 1.3552, "learning_rate": 2.1639652049515612e-05, "epoch": 0.479232, "percentage": 47.93, "elapsed_time": "6:52:42", "remaining_time": "7:28:24"}
|
1873 |
+
{"current_steps": 1873, "total_steps": 3906, "loss": 1.3455, "learning_rate": 2.1623453800142137e-05, "epoch": 0.479488, "percentage": 47.95, "elapsed_time": "6:52:55", "remaining_time": "7:28:11"}
|
1874 |
+
{"current_steps": 1874, "total_steps": 3906, "loss": 1.3495, "learning_rate": 2.1607254478716408e-05, "epoch": 0.479744, "percentage": 47.98, "elapsed_time": "6:53:08", "remaining_time": "7:27:58"}
|
1875 |
+
{"current_steps": 1875, "total_steps": 3906, "loss": 1.3665, "learning_rate": 2.1591054095935685e-05, "epoch": 0.48, "percentage": 48.0, "elapsed_time": "6:53:21", "remaining_time": "7:27:45"}
|
1876 |
+
{"current_steps": 1876, "total_steps": 3906, "loss": 1.3692, "learning_rate": 2.1574852662497934e-05, "epoch": 0.480256, "percentage": 48.03, "elapsed_time": "6:53:34", "remaining_time": "7:27:31"}
|
1877 |
+
{"current_steps": 1877, "total_steps": 3906, "loss": 1.3769, "learning_rate": 2.1558650189101827e-05, "epoch": 0.480512, "percentage": 48.05, "elapsed_time": "6:53:47", "remaining_time": "7:27:18"}
|
1878 |
+
{"current_steps": 1878, "total_steps": 3906, "loss": 1.3559, "learning_rate": 2.1542446686446703e-05, "epoch": 0.480768, "percentage": 48.08, "elapsed_time": "6:54:00", "remaining_time": "7:27:04"}
|
1879 |
+
{"current_steps": 1879, "total_steps": 3906, "loss": 1.3625, "learning_rate": 2.1526242165232593e-05, "epoch": 0.481024, "percentage": 48.11, "elapsed_time": "6:54:14", "remaining_time": "7:26:51"}
|
1880 |
+
{"current_steps": 1880, "total_steps": 3906, "loss": 1.3586, "learning_rate": 2.1510036636160203e-05, "epoch": 0.48128, "percentage": 48.13, "elapsed_time": "6:54:27", "remaining_time": "7:26:38"}
|
1881 |
+
{"current_steps": 1881, "total_steps": 3906, "loss": 1.3681, "learning_rate": 2.14938301099309e-05, "epoch": 0.481536, "percentage": 48.16, "elapsed_time": "6:54:40", "remaining_time": "7:26:25"}
|
1882 |
+
{"current_steps": 1882, "total_steps": 3906, "loss": 1.3809, "learning_rate": 2.1477622597246698e-05, "epoch": 0.481792, "percentage": 48.18, "elapsed_time": "6:54:53", "remaining_time": "7:26:11"}
|
1883 |
+
{"current_steps": 1883, "total_steps": 3906, "loss": 1.3692, "learning_rate": 2.1461414108810275e-05, "epoch": 0.482048, "percentage": 48.21, "elapsed_time": "6:55:06", "remaining_time": "7:25:58"}
|
1884 |
+
{"current_steps": 1884, "total_steps": 3906, "loss": 1.3847, "learning_rate": 2.1445204655324966e-05, "epoch": 0.482304, "percentage": 48.23, "elapsed_time": "6:55:19", "remaining_time": "7:25:45"}
|
1885 |
+
{"current_steps": 1885, "total_steps": 3906, "loss": 1.3132, "learning_rate": 2.1428994247494717e-05, "epoch": 0.48256, "percentage": 48.26, "elapsed_time": "6:55:33", "remaining_time": "7:25:32"}
|
1886 |
+
{"current_steps": 1886, "total_steps": 3906, "loss": 1.4053, "learning_rate": 2.1412782896024123e-05, "epoch": 0.482816, "percentage": 48.28, "elapsed_time": "6:55:46", "remaining_time": "7:25:18"}
|
1887 |
+
{"current_steps": 1887, "total_steps": 3906, "loss": 1.412, "learning_rate": 2.139657061161839e-05, "epoch": 0.483072, "percentage": 48.31, "elapsed_time": "6:55:59", "remaining_time": "7:25:05"}
|
1888 |
+
{"current_steps": 1888, "total_steps": 3906, "loss": 1.3454, "learning_rate": 2.1380357404983337e-05, "epoch": 0.483328, "percentage": 48.34, "elapsed_time": "6:56:12", "remaining_time": "7:24:52"}
|
1889 |
+
{"current_steps": 1889, "total_steps": 3906, "loss": 1.3567, "learning_rate": 2.1364143286825416e-05, "epoch": 0.483584, "percentage": 48.36, "elapsed_time": "6:56:25", "remaining_time": "7:24:38"}
|
1890 |
+
{"current_steps": 1890, "total_steps": 3906, "loss": 1.3464, "learning_rate": 2.134792826785166e-05, "epoch": 0.48384, "percentage": 48.39, "elapsed_time": "6:56:39", "remaining_time": "7:24:25"}
|
1891 |
+
{"current_steps": 1891, "total_steps": 3906, "loss": 1.3741, "learning_rate": 2.13317123587697e-05, "epoch": 0.484096, "percentage": 48.41, "elapsed_time": "6:56:52", "remaining_time": "7:24:12"}
|
1892 |
+
{"current_steps": 1892, "total_steps": 3906, "loss": 1.3658, "learning_rate": 2.1315495570287763e-05, "epoch": 0.484352, "percentage": 48.44, "elapsed_time": "6:57:05", "remaining_time": "7:23:59"}
|
1893 |
+
{"current_steps": 1893, "total_steps": 3906, "loss": 1.3823, "learning_rate": 2.1299277913114644e-05, "epoch": 0.484608, "percentage": 48.46, "elapsed_time": "6:57:18", "remaining_time": "7:23:45"}
|
1894 |
+
{"current_steps": 1894, "total_steps": 3906, "loss": 1.346, "learning_rate": 2.1283059397959715e-05, "epoch": 0.484864, "percentage": 48.49, "elapsed_time": "6:57:31", "remaining_time": "7:23:32"}
|
1895 |
+
{"current_steps": 1895, "total_steps": 3906, "loss": 1.3401, "learning_rate": 2.1266840035532924e-05, "epoch": 0.48512, "percentage": 48.52, "elapsed_time": "6:57:45", "remaining_time": "7:23:19"}
|
1896 |
+
{"current_steps": 1896, "total_steps": 3906, "loss": 1.3979, "learning_rate": 2.1250619836544777e-05, "epoch": 0.485376, "percentage": 48.54, "elapsed_time": "6:57:58", "remaining_time": "7:23:06"}
|
1897 |
+
{"current_steps": 1897, "total_steps": 3906, "loss": 1.3448, "learning_rate": 2.1234398811706316e-05, "epoch": 0.485632, "percentage": 48.57, "elapsed_time": "6:58:11", "remaining_time": "7:22:52"}
|
1898 |
+
{"current_steps": 1898, "total_steps": 3906, "loss": 1.3527, "learning_rate": 2.121817697172915e-05, "epoch": 0.485888, "percentage": 48.59, "elapsed_time": "6:58:24", "remaining_time": "7:22:39"}
|
1899 |
+
{"current_steps": 1899, "total_steps": 3906, "loss": 1.3447, "learning_rate": 2.120195432732542e-05, "epoch": 0.486144, "percentage": 48.62, "elapsed_time": "6:58:37", "remaining_time": "7:22:26"}
|
1900 |
+
{"current_steps": 1900, "total_steps": 3906, "loss": 1.3671, "learning_rate": 2.1185730889207776e-05, "epoch": 0.4864, "percentage": 48.64, "elapsed_time": "6:58:50", "remaining_time": "7:22:13"}
|
1901 |
+
{"current_steps": 1901, "total_steps": 3906, "loss": 1.3707, "learning_rate": 2.116950666808942e-05, "epoch": 0.486656, "percentage": 48.67, "elapsed_time": "6:59:04", "remaining_time": "7:21:59"}
|
1902 |
+
{"current_steps": 1902, "total_steps": 3906, "loss": 1.3536, "learning_rate": 2.115328167468407e-05, "epoch": 0.486912, "percentage": 48.69, "elapsed_time": "6:59:17", "remaining_time": "7:21:46"}
|
1903 |
+
{"current_steps": 1903, "total_steps": 3906, "loss": 1.3556, "learning_rate": 2.1137055919705943e-05, "epoch": 0.487168, "percentage": 48.72, "elapsed_time": "6:59:30", "remaining_time": "7:21:33"}
|
1904 |
+
{"current_steps": 1904, "total_steps": 3906, "loss": 1.3176, "learning_rate": 2.1120829413869753e-05, "epoch": 0.487424, "percentage": 48.75, "elapsed_time": "6:59:43", "remaining_time": "7:21:19"}
|
1905 |
+
{"current_steps": 1905, "total_steps": 3906, "loss": 1.3928, "learning_rate": 2.110460216789073e-05, "epoch": 0.48768, "percentage": 48.77, "elapsed_time": "6:59:56", "remaining_time": "7:21:06"}
|
1906 |
+
{"current_steps": 1906, "total_steps": 3906, "loss": 1.3438, "learning_rate": 2.1088374192484563e-05, "epoch": 0.487936, "percentage": 48.8, "elapsed_time": "7:00:10", "remaining_time": "7:20:53"}
|
1907 |
+
{"current_steps": 1907, "total_steps": 3906, "loss": 1.3727, "learning_rate": 2.107214549836746e-05, "epoch": 0.488192, "percentage": 48.82, "elapsed_time": "7:00:23", "remaining_time": "7:20:40"}
|
1908 |
+
{"current_steps": 1908, "total_steps": 3906, "loss": 1.3998, "learning_rate": 2.105591609625608e-05, "epoch": 0.488448, "percentage": 48.85, "elapsed_time": "7:00:36", "remaining_time": "7:20:26"}
|
1909 |
+
{"current_steps": 1909, "total_steps": 3906, "loss": 1.3545, "learning_rate": 2.1039685996867553e-05, "epoch": 0.488704, "percentage": 48.87, "elapsed_time": "7:00:49", "remaining_time": "7:20:13"}
|
1910 |
+
{"current_steps": 1910, "total_steps": 3906, "loss": 1.3421, "learning_rate": 2.1023455210919464e-05, "epoch": 0.48896, "percentage": 48.9, "elapsed_time": "7:01:02", "remaining_time": "7:20:00"}
|
1911 |
+
{"current_steps": 1911, "total_steps": 3906, "loss": 1.39, "learning_rate": 2.1007223749129868e-05, "epoch": 0.489216, "percentage": 48.92, "elapsed_time": "7:01:16", "remaining_time": "7:19:47"}
|
1912 |
+
{"current_steps": 1912, "total_steps": 3906, "loss": 1.3524, "learning_rate": 2.0990991622217245e-05, "epoch": 0.489472, "percentage": 48.95, "elapsed_time": "7:01:29", "remaining_time": "7:19:33"}
|
1913 |
+
{"current_steps": 1913, "total_steps": 3906, "loss": 1.3887, "learning_rate": 2.097475884090053e-05, "epoch": 0.489728, "percentage": 48.98, "elapsed_time": "7:01:42", "remaining_time": "7:19:20"}
|
1914 |
+
{"current_steps": 1914, "total_steps": 3906, "loss": 1.3522, "learning_rate": 2.0958525415899097e-05, "epoch": 0.489984, "percentage": 49.0, "elapsed_time": "7:01:55", "remaining_time": "7:19:07"}
|
1915 |
+
{"current_steps": 1915, "total_steps": 3906, "loss": 1.3368, "learning_rate": 2.094229135793272e-05, "epoch": 0.49024, "percentage": 49.03, "elapsed_time": "7:02:08", "remaining_time": "7:18:53"}
|
1916 |
+
{"current_steps": 1916, "total_steps": 3906, "loss": 1.3398, "learning_rate": 2.092605667772161e-05, "epoch": 0.490496, "percentage": 49.05, "elapsed_time": "7:02:21", "remaining_time": "7:18:40"}
|
1917 |
+
{"current_steps": 1917, "total_steps": 3906, "loss": 1.4091, "learning_rate": 2.090982138598638e-05, "epoch": 0.490752, "percentage": 49.08, "elapsed_time": "7:02:35", "remaining_time": "7:18:27"}
|
1918 |
+
{"current_steps": 1918, "total_steps": 3906, "loss": 1.3868, "learning_rate": 2.089358549344805e-05, "epoch": 0.491008, "percentage": 49.1, "elapsed_time": "7:02:48", "remaining_time": "7:18:14"}
|
1919 |
+
{"current_steps": 1919, "total_steps": 3906, "loss": 1.3416, "learning_rate": 2.0877349010828044e-05, "epoch": 0.491264, "percentage": 49.13, "elapsed_time": "7:03:01", "remaining_time": "7:18:00"}
|
1920 |
+
{"current_steps": 1920, "total_steps": 3906, "loss": 1.4201, "learning_rate": 2.0861111948848164e-05, "epoch": 0.49152, "percentage": 49.16, "elapsed_time": "7:03:14", "remaining_time": "7:17:47"}
|
1921 |
+
{"current_steps": 1921, "total_steps": 3906, "loss": 1.3307, "learning_rate": 2.08448743182306e-05, "epoch": 0.491776, "percentage": 49.18, "elapsed_time": "7:03:27", "remaining_time": "7:17:34"}
|
1922 |
+
{"current_steps": 1922, "total_steps": 3906, "loss": 1.3439, "learning_rate": 2.0828636129697913e-05, "epoch": 0.492032, "percentage": 49.21, "elapsed_time": "7:03:40", "remaining_time": "7:17:21"}
|
1923 |
+
{"current_steps": 1923, "total_steps": 3906, "loss": 1.4366, "learning_rate": 2.081239739397304e-05, "epoch": 0.492288, "percentage": 49.23, "elapsed_time": "7:03:54", "remaining_time": "7:17:07"}
|
1924 |
+
{"current_steps": 1924, "total_steps": 3906, "loss": 1.3363, "learning_rate": 2.0796158121779276e-05, "epoch": 0.492544, "percentage": 49.26, "elapsed_time": "7:04:07", "remaining_time": "7:16:54"}
|
1925 |
+
{"current_steps": 1925, "total_steps": 3906, "loss": 1.3487, "learning_rate": 2.0779918323840272e-05, "epoch": 0.4928, "percentage": 49.28, "elapsed_time": "7:04:20", "remaining_time": "7:16:41"}
|
1926 |
+
{"current_steps": 1926, "total_steps": 3906, "loss": 1.3578, "learning_rate": 2.0763678010880022e-05, "epoch": 0.493056, "percentage": 49.31, "elapsed_time": "7:04:33", "remaining_time": "7:16:27"}
|
1927 |
+
{"current_steps": 1927, "total_steps": 3906, "loss": 1.3834, "learning_rate": 2.0747437193622853e-05, "epoch": 0.493312, "percentage": 49.33, "elapsed_time": "7:04:46", "remaining_time": "7:16:14"}
|
1928 |
+
{"current_steps": 1928, "total_steps": 3906, "loss": 1.392, "learning_rate": 2.0731195882793444e-05, "epoch": 0.493568, "percentage": 49.36, "elapsed_time": "7:05:00", "remaining_time": "7:16:01"}
|
1929 |
+
{"current_steps": 1929, "total_steps": 3906, "loss": 1.3432, "learning_rate": 2.071495408911679e-05, "epoch": 0.493824, "percentage": 49.39, "elapsed_time": "7:05:13", "remaining_time": "7:15:48"}
|
1930 |
+
{"current_steps": 1930, "total_steps": 3906, "loss": 1.4333, "learning_rate": 2.0698711823318206e-05, "epoch": 0.49408, "percentage": 49.41, "elapsed_time": "7:05:26", "remaining_time": "7:15:34"}
|
1931 |
+
{"current_steps": 1931, "total_steps": 3906, "loss": 1.3208, "learning_rate": 2.068246909612331e-05, "epoch": 0.494336, "percentage": 49.44, "elapsed_time": "7:05:39", "remaining_time": "7:15:21"}
|
1932 |
+
{"current_steps": 1932, "total_steps": 3906, "loss": 1.3466, "learning_rate": 2.0666225918258044e-05, "epoch": 0.494592, "percentage": 49.46, "elapsed_time": "7:05:52", "remaining_time": "7:15:08"}
|
1933 |
+
{"current_steps": 1933, "total_steps": 3906, "loss": 1.3353, "learning_rate": 2.0649982300448622e-05, "epoch": 0.494848, "percentage": 49.49, "elapsed_time": "7:06:05", "remaining_time": "7:14:54"}
|
1934 |
+
{"current_steps": 1934, "total_steps": 3906, "loss": 1.4063, "learning_rate": 2.0633738253421568e-05, "epoch": 0.495104, "percentage": 49.51, "elapsed_time": "7:06:19", "remaining_time": "7:14:41"}
|
1935 |
+
{"current_steps": 1935, "total_steps": 3906, "loss": 1.3536, "learning_rate": 2.0617493787903693e-05, "epoch": 0.49536, "percentage": 49.54, "elapsed_time": "7:06:32", "remaining_time": "7:14:28"}
|
1936 |
+
{"current_steps": 1936, "total_steps": 3906, "loss": 1.3588, "learning_rate": 2.060124891462206e-05, "epoch": 0.495616, "percentage": 49.56, "elapsed_time": "7:06:45", "remaining_time": "7:14:15"}
|
1937 |
+
{"current_steps": 1937, "total_steps": 3906, "loss": 1.3787, "learning_rate": 2.058500364430403e-05, "epoch": 0.495872, "percentage": 49.59, "elapsed_time": "7:06:58", "remaining_time": "7:14:01"}
|
1938 |
+
{"current_steps": 1938, "total_steps": 3906, "loss": 1.3801, "learning_rate": 2.056875798767721e-05, "epoch": 0.496128, "percentage": 49.62, "elapsed_time": "7:07:11", "remaining_time": "7:13:48"}
|
1939 |
+
{"current_steps": 1939, "total_steps": 3906, "loss": 1.3386, "learning_rate": 2.055251195546945e-05, "epoch": 0.496384, "percentage": 49.64, "elapsed_time": "7:07:24", "remaining_time": "7:13:35"}
|
1940 |
+
{"current_steps": 1940, "total_steps": 3906, "loss": 1.3933, "learning_rate": 2.0536265558408882e-05, "epoch": 0.49664, "percentage": 49.67, "elapsed_time": "7:07:38", "remaining_time": "7:13:21"}
|
1941 |
+
{"current_steps": 1941, "total_steps": 3906, "loss": 1.4091, "learning_rate": 2.052001880722385e-05, "epoch": 0.496896, "percentage": 49.69, "elapsed_time": "7:07:51", "remaining_time": "7:13:08"}
|
1942 |
+
{"current_steps": 1942, "total_steps": 3906, "loss": 1.3607, "learning_rate": 2.0503771712642943e-05, "epoch": 0.497152, "percentage": 49.72, "elapsed_time": "7:08:04", "remaining_time": "7:12:55"}
|
1943 |
+
{"current_steps": 1943, "total_steps": 3906, "loss": 1.3971, "learning_rate": 2.048752428539498e-05, "epoch": 0.497408, "percentage": 49.74, "elapsed_time": "7:08:17", "remaining_time": "7:12:41"}
|
1944 |
+
{"current_steps": 1944, "total_steps": 3906, "loss": 1.3329, "learning_rate": 2.0471276536208985e-05, "epoch": 0.497664, "percentage": 49.77, "elapsed_time": "7:08:30", "remaining_time": "7:12:28"}
|
1945 |
+
{"current_steps": 1945, "total_steps": 3906, "loss": 1.3677, "learning_rate": 2.045502847581422e-05, "epoch": 0.49792, "percentage": 49.8, "elapsed_time": "7:08:43", "remaining_time": "7:12:15"}
|
1946 |
+
{"current_steps": 1946, "total_steps": 3906, "loss": 1.3767, "learning_rate": 2.043878011494012e-05, "epoch": 0.498176, "percentage": 49.82, "elapsed_time": "7:08:56", "remaining_time": "7:12:02"}
|
1947 |
+
{"current_steps": 1947, "total_steps": 3906, "loss": 1.371, "learning_rate": 2.0422531464316348e-05, "epoch": 0.498432, "percentage": 49.85, "elapsed_time": "7:09:10", "remaining_time": "7:11:48"}
|
1948 |
+
{"current_steps": 1948, "total_steps": 3906, "loss": 1.3701, "learning_rate": 2.040628253467274e-05, "epoch": 0.498688, "percentage": 49.87, "elapsed_time": "7:09:23", "remaining_time": "7:11:35"}
|
1949 |
+
{"current_steps": 1949, "total_steps": 3906, "loss": 1.3573, "learning_rate": 2.0390033336739324e-05, "epoch": 0.498944, "percentage": 49.9, "elapsed_time": "7:09:36", "remaining_time": "7:11:22"}
|
1950 |
+
{"current_steps": 1950, "total_steps": 3906, "loss": 1.3735, "learning_rate": 2.0373783881246304e-05, "epoch": 0.4992, "percentage": 49.92, "elapsed_time": "7:09:49", "remaining_time": "7:11:08"}
|
1951 |
+
{"current_steps": 1951, "total_steps": 3906, "loss": 1.3625, "learning_rate": 2.035753417892405e-05, "epoch": 0.499456, "percentage": 49.95, "elapsed_time": "7:10:02", "remaining_time": "7:10:55"}
|
1952 |
+
{"current_steps": 1952, "total_steps": 3906, "loss": 1.3675, "learning_rate": 2.03412842405031e-05, "epoch": 0.499712, "percentage": 49.97, "elapsed_time": "7:10:15", "remaining_time": "7:10:42"}
|
1953 |
+
{"current_steps": 1953, "total_steps": 3906, "loss": 1.3578, "learning_rate": 2.0325034076714154e-05, "epoch": 0.499968, "percentage": 50.0, "elapsed_time": "7:10:29", "remaining_time": "7:10:29"}
|
1954 |
+
{"current_steps": 1954, "total_steps": 3906, "loss": 1.3417, "learning_rate": 2.030878369828804e-05, "epoch": 0.500224, "percentage": 50.03, "elapsed_time": "7:10:42", "remaining_time": "7:10:15"}
|
1955 |
+
{"current_steps": 1955, "total_steps": 3906, "loss": 1.3483, "learning_rate": 2.0292533115955747e-05, "epoch": 0.50048, "percentage": 50.05, "elapsed_time": "7:10:55", "remaining_time": "7:10:02"}
|
1956 |
+
{"current_steps": 1956, "total_steps": 3906, "loss": 1.3746, "learning_rate": 2.0276282340448388e-05, "epoch": 0.500736, "percentage": 50.08, "elapsed_time": "7:11:08", "remaining_time": "7:09:49"}
|
1957 |
+
{"current_steps": 1957, "total_steps": 3906, "loss": 1.3732, "learning_rate": 2.0260031382497223e-05, "epoch": 0.500992, "percentage": 50.1, "elapsed_time": "7:11:21", "remaining_time": "7:09:36"}
|
1958 |
+
{"current_steps": 1958, "total_steps": 3906, "loss": 1.3855, "learning_rate": 2.0243780252833595e-05, "epoch": 0.501248, "percentage": 50.13, "elapsed_time": "7:11:35", "remaining_time": "7:09:22"}
|
1959 |
+
{"current_steps": 1959, "total_steps": 3906, "loss": 1.3594, "learning_rate": 2.0227528962189007e-05, "epoch": 0.501504, "percentage": 50.15, "elapsed_time": "7:11:48", "remaining_time": "7:09:09"}
|
1960 |
+
{"current_steps": 1960, "total_steps": 3906, "loss": 1.3553, "learning_rate": 2.0211277521295028e-05, "epoch": 0.50176, "percentage": 50.18, "elapsed_time": "7:12:01", "remaining_time": "7:08:56"}
|
1961 |
+
{"current_steps": 1961, "total_steps": 3906, "loss": 1.3613, "learning_rate": 2.0195025940883345e-05, "epoch": 0.502016, "percentage": 50.2, "elapsed_time": "7:12:14", "remaining_time": "7:08:43"}
|
1962 |
+
{"current_steps": 1962, "total_steps": 3906, "loss": 1.4203, "learning_rate": 2.0178774231685737e-05, "epoch": 0.502272, "percentage": 50.23, "elapsed_time": "7:12:27", "remaining_time": "7:08:29"}
|
1963 |
+
{"current_steps": 1963, "total_steps": 3906, "loss": 1.3929, "learning_rate": 2.0162522404434064e-05, "epoch": 0.502528, "percentage": 50.26, "elapsed_time": "7:12:40", "remaining_time": "7:08:16"}
|
1964 |
+
{"current_steps": 1964, "total_steps": 3906, "loss": 1.3769, "learning_rate": 2.0146270469860267e-05, "epoch": 0.502784, "percentage": 50.28, "elapsed_time": "7:12:54", "remaining_time": "7:08:03"}
|
1965 |
+
{"current_steps": 1965, "total_steps": 3906, "loss": 1.3841, "learning_rate": 2.0130018438696356e-05, "epoch": 0.50304, "percentage": 50.31, "elapsed_time": "7:13:07", "remaining_time": "7:07:49"}
|
1966 |
+
{"current_steps": 1966, "total_steps": 3906, "loss": 1.3184, "learning_rate": 2.011376632167441e-05, "epoch": 0.503296, "percentage": 50.33, "elapsed_time": "7:13:20", "remaining_time": "7:07:36"}
|
1967 |
+
{"current_steps": 1967, "total_steps": 3906, "loss": 1.3799, "learning_rate": 2.009751412952655e-05, "epoch": 0.503552, "percentage": 50.36, "elapsed_time": "7:13:33", "remaining_time": "7:07:23"}
|
1968 |
+
{"current_steps": 1968, "total_steps": 3906, "loss": 1.3499, "learning_rate": 2.008126187298496e-05, "epoch": 0.503808, "percentage": 50.38, "elapsed_time": "7:13:46", "remaining_time": "7:07:10"}
|
1969 |
+
{"current_steps": 1969, "total_steps": 3906, "loss": 1.2956, "learning_rate": 2.006500956278187e-05, "epoch": 0.504064, "percentage": 50.41, "elapsed_time": "7:14:00", "remaining_time": "7:06:56"}
|
1970 |
+
{"current_steps": 1970, "total_steps": 3906, "loss": 1.3707, "learning_rate": 2.004875720964953e-05, "epoch": 0.50432, "percentage": 50.44, "elapsed_time": "7:14:13", "remaining_time": "7:06:43"}
|
1971 |
+
{"current_steps": 1971, "total_steps": 3906, "loss": 1.3477, "learning_rate": 2.0032504824320232e-05, "epoch": 0.504576, "percentage": 50.46, "elapsed_time": "7:14:26", "remaining_time": "7:06:30"}
|
1972 |
+
{"current_steps": 1972, "total_steps": 3906, "loss": 1.4148, "learning_rate": 2.0016252417526286e-05, "epoch": 0.504832, "percentage": 50.49, "elapsed_time": "7:14:39", "remaining_time": "7:06:16"}
|
1973 |
+
{"current_steps": 1973, "total_steps": 3906, "loss": 1.3116, "learning_rate": 2e-05, "epoch": 0.505088, "percentage": 50.51, "elapsed_time": "7:14:52", "remaining_time": "7:06:03"}
|
1974 |
+
{"current_steps": 1974, "total_steps": 3906, "loss": 1.3867, "learning_rate": 1.998374758247372e-05, "epoch": 0.505344, "percentage": 50.54, "elapsed_time": "7:15:05", "remaining_time": "7:05:50"}
|
1975 |
+
{"current_steps": 1975, "total_steps": 3906, "loss": 1.3371, "learning_rate": 1.9967495175679767e-05, "epoch": 0.5056, "percentage": 50.56, "elapsed_time": "7:15:19", "remaining_time": "7:05:37"}
|
1976 |
+
{"current_steps": 1976, "total_steps": 3906, "loss": 1.3702, "learning_rate": 1.9951242790350473e-05, "epoch": 0.505856, "percentage": 50.59, "elapsed_time": "7:15:32", "remaining_time": "7:05:23"}
|
1977 |
+
{"current_steps": 1977, "total_steps": 3906, "loss": 1.4002, "learning_rate": 1.993499043721814e-05, "epoch": 0.506112, "percentage": 50.61, "elapsed_time": "7:15:45", "remaining_time": "7:05:10"}
|
1978 |
+
{"current_steps": 1978, "total_steps": 3906, "loss": 1.4003, "learning_rate": 1.9918738127015044e-05, "epoch": 0.506368, "percentage": 50.64, "elapsed_time": "7:15:58", "remaining_time": "7:04:57"}
|
1979 |
+
{"current_steps": 1979, "total_steps": 3906, "loss": 1.3934, "learning_rate": 1.9902485870473456e-05, "epoch": 0.506624, "percentage": 50.67, "elapsed_time": "7:16:11", "remaining_time": "7:04:44"}
|
1980 |
+
{"current_steps": 1980, "total_steps": 3906, "loss": 1.323, "learning_rate": 1.9886233678325593e-05, "epoch": 0.50688, "percentage": 50.69, "elapsed_time": "7:16:24", "remaining_time": "7:04:30"}
|
1981 |
+
{"current_steps": 1981, "total_steps": 3906, "loss": 1.3769, "learning_rate": 1.9869981561303648e-05, "epoch": 0.507136, "percentage": 50.72, "elapsed_time": "7:16:38", "remaining_time": "7:04:17"}
|
1982 |
+
{"current_steps": 1982, "total_steps": 3906, "loss": 1.3545, "learning_rate": 1.9853729530139733e-05, "epoch": 0.507392, "percentage": 50.74, "elapsed_time": "7:16:51", "remaining_time": "7:04:04"}
|
1983 |
+
{"current_steps": 1983, "total_steps": 3906, "loss": 1.3723, "learning_rate": 1.9837477595565943e-05, "epoch": 0.507648, "percentage": 50.77, "elapsed_time": "7:17:04", "remaining_time": "7:03:51"}
|
1984 |
+
{"current_steps": 1984, "total_steps": 3906, "loss": 1.3552, "learning_rate": 1.9821225768314273e-05, "epoch": 0.507904, "percentage": 50.79, "elapsed_time": "7:17:17", "remaining_time": "7:03:37"}
|
1985 |
+
{"current_steps": 1985, "total_steps": 3906, "loss": 1.4037, "learning_rate": 1.9804974059116662e-05, "epoch": 0.50816, "percentage": 50.82, "elapsed_time": "7:17:31", "remaining_time": "7:03:24"}
|
1986 |
+
{"current_steps": 1986, "total_steps": 3906, "loss": 1.3649, "learning_rate": 1.9788722478704976e-05, "epoch": 0.508416, "percentage": 50.84, "elapsed_time": "7:17:44", "remaining_time": "7:03:11"}
|
1987 |
+
{"current_steps": 1987, "total_steps": 3906, "loss": 1.3329, "learning_rate": 1.9772471037811e-05, "epoch": 0.508672, "percentage": 50.87, "elapsed_time": "7:17:57", "remaining_time": "7:02:58"}
|
1988 |
+
{"current_steps": 1988, "total_steps": 3906, "loss": 1.3828, "learning_rate": 1.9756219747166402e-05, "epoch": 0.508928, "percentage": 50.9, "elapsed_time": "7:18:10", "remaining_time": "7:02:44"}
|
1989 |
+
{"current_steps": 1989, "total_steps": 3906, "loss": 1.4168, "learning_rate": 1.973996861750279e-05, "epoch": 0.509184, "percentage": 50.92, "elapsed_time": "7:18:23", "remaining_time": "7:02:31"}
|
1990 |
+
{"current_steps": 1990, "total_steps": 3906, "loss": 1.3796, "learning_rate": 1.9723717659551615e-05, "epoch": 0.50944, "percentage": 50.95, "elapsed_time": "7:18:36", "remaining_time": "7:02:18"}
|
1991 |
+
{"current_steps": 1991, "total_steps": 3906, "loss": 1.3685, "learning_rate": 1.970746688404426e-05, "epoch": 0.509696, "percentage": 50.97, "elapsed_time": "7:18:50", "remaining_time": "7:02:05"}
|
1992 |
+
{"current_steps": 1992, "total_steps": 3906, "loss": 1.3765, "learning_rate": 1.9691216301711966e-05, "epoch": 0.509952, "percentage": 51.0, "elapsed_time": "7:19:03", "remaining_time": "7:01:51"}
|
1993 |
+
{"current_steps": 1993, "total_steps": 3906, "loss": 1.3655, "learning_rate": 1.9674965923285853e-05, "epoch": 0.510208, "percentage": 51.02, "elapsed_time": "7:19:16", "remaining_time": "7:01:38"}
|
1994 |
+
{"current_steps": 1994, "total_steps": 3906, "loss": 1.3714, "learning_rate": 1.96587157594969e-05, "epoch": 0.510464, "percentage": 51.05, "elapsed_time": "7:19:29", "remaining_time": "7:01:25"}
|
1995 |
+
{"current_steps": 1995, "total_steps": 3906, "loss": 1.3364, "learning_rate": 1.964246582107596e-05, "epoch": 0.51072, "percentage": 51.08, "elapsed_time": "7:19:42", "remaining_time": "7:01:11"}
|
1996 |
+
{"current_steps": 1996, "total_steps": 3906, "loss": 1.3001, "learning_rate": 1.9626216118753706e-05, "epoch": 0.510976, "percentage": 51.1, "elapsed_time": "7:19:55", "remaining_time": "7:00:58"}
|
1997 |
+
{"current_steps": 1997, "total_steps": 3906, "loss": 1.3261, "learning_rate": 1.960996666326068e-05, "epoch": 0.511232, "percentage": 51.13, "elapsed_time": "7:20:09", "remaining_time": "7:00:45"}
|
1998 |
+
{"current_steps": 1998, "total_steps": 3906, "loss": 1.4068, "learning_rate": 1.959371746532726e-05, "epoch": 0.511488, "percentage": 51.15, "elapsed_time": "7:20:22", "remaining_time": "7:00:32"}
|
1999 |
+
{"current_steps": 1999, "total_steps": 3906, "loss": 1.3977, "learning_rate": 1.9577468535683656e-05, "epoch": 0.511744, "percentage": 51.18, "elapsed_time": "7:20:35", "remaining_time": "7:00:18"}
|
2000 |
+
{"current_steps": 2000, "total_steps": 3906, "loss": 1.3409, "learning_rate": 1.9561219885059882e-05, "epoch": 0.512, "percentage": 51.2, "elapsed_time": "7:20:48", "remaining_time": "7:00:05"}
|
2001 |
+
{"current_steps": 2001, "total_steps": 3906, "loss": 1.3669, "learning_rate": 1.9544971524185788e-05, "epoch": 0.512256, "percentage": 51.23, "elapsed_time": "7:21:21", "remaining_time": "7:00:10"}
|
2002 |
+
{"current_steps": 2002, "total_steps": 3906, "loss": 1.3806, "learning_rate": 1.9528723463791018e-05, "epoch": 0.512512, "percentage": 51.25, "elapsed_time": "7:21:34", "remaining_time": "6:59:57"}
|
2003 |
+
{"current_steps": 2003, "total_steps": 3906, "loss": 1.3475, "learning_rate": 1.9512475714605026e-05, "epoch": 0.512768, "percentage": 51.28, "elapsed_time": "7:21:47", "remaining_time": "6:59:44"}
|
2004 |
+
{"current_steps": 2004, "total_steps": 3906, "loss": 1.3496, "learning_rate": 1.949622828735706e-05, "epoch": 0.513024, "percentage": 51.31, "elapsed_time": "7:22:00", "remaining_time": "6:59:30"}
|
2005 |
+
{"current_steps": 2005, "total_steps": 3906, "loss": 1.3804, "learning_rate": 1.9479981192776155e-05, "epoch": 0.51328, "percentage": 51.33, "elapsed_time": "7:22:13", "remaining_time": "6:59:17"}
|
2006 |
+
{"current_steps": 2006, "total_steps": 3906, "loss": 1.3265, "learning_rate": 1.946373444159112e-05, "epoch": 0.513536, "percentage": 51.36, "elapsed_time": "7:22:27", "remaining_time": "6:59:04"}
|
2007 |
+
{"current_steps": 2007, "total_steps": 3906, "loss": 1.3409, "learning_rate": 1.9447488044530562e-05, "epoch": 0.513792, "percentage": 51.38, "elapsed_time": "7:22:40", "remaining_time": "6:58:51"}
|
2008 |
+
{"current_steps": 2008, "total_steps": 3906, "loss": 1.356, "learning_rate": 1.9431242012322804e-05, "epoch": 0.514048, "percentage": 51.41, "elapsed_time": "7:22:53", "remaining_time": "6:58:37"}
|
2009 |
+
{"current_steps": 2009, "total_steps": 3906, "loss": 1.416, "learning_rate": 1.9414996355695977e-05, "epoch": 0.514304, "percentage": 51.43, "elapsed_time": "7:23:06", "remaining_time": "6:58:24"}
|
2010 |
+
{"current_steps": 2010, "total_steps": 3906, "loss": 1.3442, "learning_rate": 1.9398751085377947e-05, "epoch": 0.51456, "percentage": 51.46, "elapsed_time": "7:23:19", "remaining_time": "6:58:11"}
|
2011 |
+
{"current_steps": 2011, "total_steps": 3906, "loss": 1.3587, "learning_rate": 1.9382506212096314e-05, "epoch": 0.514816, "percentage": 51.48, "elapsed_time": "7:23:33", "remaining_time": "6:57:58"}
|
2012 |
+
{"current_steps": 2012, "total_steps": 3906, "loss": 1.3568, "learning_rate": 1.9366261746578436e-05, "epoch": 0.515072, "percentage": 51.51, "elapsed_time": "7:23:46", "remaining_time": "6:57:44"}
|