{ "best_metric": 0.4402177035808563, "best_model_checkpoint": "NHS-BiomedNLP-BiomedBERT-hypop-512\\checkpoint-794", "epoch": 3.0, "eval_steps": 500, "global_step": 1191, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "grad_norm": 10.960957527160645, "learning_rate": 2.9987405541561713e-05, "loss": 0.7342, "step": 1 }, { "epoch": 0.01, "grad_norm": 9.521003723144531, "learning_rate": 2.997481108312343e-05, "loss": 0.7329, "step": 2 }, { "epoch": 0.01, "grad_norm": 8.874771118164062, "learning_rate": 2.9962216624685138e-05, "loss": 0.7373, "step": 3 }, { "epoch": 0.01, "grad_norm": 7.235786437988281, "learning_rate": 2.994962216624685e-05, "loss": 0.6377, "step": 4 }, { "epoch": 0.01, "grad_norm": 13.307806968688965, "learning_rate": 2.9937027707808566e-05, "loss": 0.7346, "step": 5 }, { "epoch": 0.02, "grad_norm": 5.770761013031006, "learning_rate": 2.992443324937028e-05, "loss": 0.6492, "step": 6 }, { "epoch": 0.02, "grad_norm": 6.9356560707092285, "learning_rate": 2.9911838790931988e-05, "loss": 0.6832, "step": 7 }, { "epoch": 0.02, "grad_norm": 13.655845642089844, "learning_rate": 2.9899244332493704e-05, "loss": 0.8057, "step": 8 }, { "epoch": 0.02, "grad_norm": 8.541060447692871, "learning_rate": 2.9886649874055416e-05, "loss": 0.6656, "step": 9 }, { "epoch": 0.03, "grad_norm": 6.785902500152588, "learning_rate": 2.987405541561713e-05, "loss": 0.6787, "step": 10 }, { "epoch": 0.03, "grad_norm": 6.900331974029541, "learning_rate": 2.9861460957178844e-05, "loss": 0.6868, "step": 11 }, { "epoch": 0.03, "grad_norm": 9.852057456970215, "learning_rate": 2.9848866498740553e-05, "loss": 0.6991, "step": 12 }, { "epoch": 0.03, "grad_norm": 6.37741756439209, "learning_rate": 2.9836272040302266e-05, "loss": 0.6302, "step": 13 }, { "epoch": 0.04, "grad_norm": 5.589299201965332, "learning_rate": 2.982367758186398e-05, "loss": 0.6055, "step": 14 }, { "epoch": 0.04, "grad_norm": 10.325979232788086, "learning_rate": 2.9811083123425694e-05, "loss": 0.7367, "step": 15 }, { "epoch": 0.04, "grad_norm": 10.072173118591309, "learning_rate": 2.9798488664987406e-05, "loss": 0.6229, "step": 16 }, { "epoch": 0.04, "grad_norm": 10.84774398803711, "learning_rate": 2.978589420654912e-05, "loss": 0.601, "step": 17 }, { "epoch": 0.05, "grad_norm": 12.358999252319336, "learning_rate": 2.977329974811083e-05, "loss": 0.6277, "step": 18 }, { "epoch": 0.05, "grad_norm": 13.19227409362793, "learning_rate": 2.9760705289672544e-05, "loss": 0.5326, "step": 19 }, { "epoch": 0.05, "grad_norm": 6.927993297576904, "learning_rate": 2.974811083123426e-05, "loss": 0.5725, "step": 20 }, { "epoch": 0.05, "grad_norm": 11.326826095581055, "learning_rate": 2.973551637279597e-05, "loss": 0.6021, "step": 21 }, { "epoch": 0.06, "grad_norm": 7.477938652038574, "learning_rate": 2.972292191435768e-05, "loss": 0.4767, "step": 22 }, { "epoch": 0.06, "grad_norm": 17.05657958984375, "learning_rate": 2.9710327455919397e-05, "loss": 0.4546, "step": 23 }, { "epoch": 0.06, "grad_norm": 10.406664848327637, "learning_rate": 2.969773299748111e-05, "loss": 0.4608, "step": 24 }, { "epoch": 0.06, "grad_norm": 7.126372337341309, "learning_rate": 2.968513853904282e-05, "loss": 0.5253, "step": 25 }, { "epoch": 0.07, "grad_norm": 22.958934783935547, "learning_rate": 2.9672544080604534e-05, "loss": 0.7278, "step": 26 }, { "epoch": 0.07, "grad_norm": 15.145363807678223, "learning_rate": 2.9659949622166247e-05, "loss": 0.6793, "step": 27 }, { "epoch": 0.07, "grad_norm": 6.81843900680542, "learning_rate": 2.964735516372796e-05, "loss": 0.5066, "step": 28 }, { "epoch": 0.07, "grad_norm": 5.444020748138428, "learning_rate": 2.9634760705289675e-05, "loss": 0.2889, "step": 29 }, { "epoch": 0.08, "grad_norm": 20.001272201538086, "learning_rate": 2.9622166246851387e-05, "loss": 0.4007, "step": 30 }, { "epoch": 0.08, "grad_norm": 9.261313438415527, "learning_rate": 2.9609571788413096e-05, "loss": 0.4314, "step": 31 }, { "epoch": 0.08, "grad_norm": 9.045961380004883, "learning_rate": 2.9596977329974812e-05, "loss": 0.7514, "step": 32 }, { "epoch": 0.08, "grad_norm": 8.198954582214355, "learning_rate": 2.9584382871536524e-05, "loss": 0.498, "step": 33 }, { "epoch": 0.09, "grad_norm": 13.375472068786621, "learning_rate": 2.9571788413098237e-05, "loss": 0.3819, "step": 34 }, { "epoch": 0.09, "grad_norm": 9.266693115234375, "learning_rate": 2.955919395465995e-05, "loss": 0.4062, "step": 35 }, { "epoch": 0.09, "grad_norm": 10.688496589660645, "learning_rate": 2.9546599496221662e-05, "loss": 0.37, "step": 36 }, { "epoch": 0.09, "grad_norm": 4.4655046463012695, "learning_rate": 2.9534005037783378e-05, "loss": 0.1862, "step": 37 }, { "epoch": 0.1, "grad_norm": 20.452878952026367, "learning_rate": 2.952141057934509e-05, "loss": 0.6387, "step": 38 }, { "epoch": 0.1, "grad_norm": 6.115667819976807, "learning_rate": 2.9508816120906802e-05, "loss": 0.404, "step": 39 }, { "epoch": 0.1, "grad_norm": 18.10906982421875, "learning_rate": 2.9496221662468515e-05, "loss": 0.5456, "step": 40 }, { "epoch": 0.1, "grad_norm": 11.372830390930176, "learning_rate": 2.9483627204030227e-05, "loss": 0.2977, "step": 41 }, { "epoch": 0.11, "grad_norm": 8.976436614990234, "learning_rate": 2.947103274559194e-05, "loss": 0.4582, "step": 42 }, { "epoch": 0.11, "grad_norm": 16.73163414001465, "learning_rate": 2.9458438287153656e-05, "loss": 0.3448, "step": 43 }, { "epoch": 0.11, "grad_norm": 8.729334831237793, "learning_rate": 2.9445843828715368e-05, "loss": 0.1669, "step": 44 }, { "epoch": 0.11, "grad_norm": 4.726550579071045, "learning_rate": 2.9433249370277077e-05, "loss": 0.1398, "step": 45 }, { "epoch": 0.12, "grad_norm": 4.698641300201416, "learning_rate": 2.9420654911838793e-05, "loss": 0.1377, "step": 46 }, { "epoch": 0.12, "grad_norm": 15.612041473388672, "learning_rate": 2.9408060453400505e-05, "loss": 0.5077, "step": 47 }, { "epoch": 0.12, "grad_norm": 18.663637161254883, "learning_rate": 2.9395465994962218e-05, "loss": 0.5674, "step": 48 }, { "epoch": 0.12, "grad_norm": 24.12005043029785, "learning_rate": 2.938287153652393e-05, "loss": 0.546, "step": 49 }, { "epoch": 0.13, "grad_norm": 16.070539474487305, "learning_rate": 2.9370277078085643e-05, "loss": 0.4623, "step": 50 }, { "epoch": 0.13, "grad_norm": 26.91022491455078, "learning_rate": 2.9357682619647355e-05, "loss": 0.3909, "step": 51 }, { "epoch": 0.13, "grad_norm": 13.222580909729004, "learning_rate": 2.934508816120907e-05, "loss": 0.5905, "step": 52 }, { "epoch": 0.13, "grad_norm": 13.613443374633789, "learning_rate": 2.9332493702770783e-05, "loss": 0.3731, "step": 53 }, { "epoch": 0.14, "grad_norm": 20.913328170776367, "learning_rate": 2.9319899244332492e-05, "loss": 0.789, "step": 54 }, { "epoch": 0.14, "grad_norm": 10.94701099395752, "learning_rate": 2.9307304785894208e-05, "loss": 0.1207, "step": 55 }, { "epoch": 0.14, "grad_norm": 10.41066837310791, "learning_rate": 2.929471032745592e-05, "loss": 0.4236, "step": 56 }, { "epoch": 0.14, "grad_norm": 19.012123107910156, "learning_rate": 2.9282115869017633e-05, "loss": 0.474, "step": 57 }, { "epoch": 0.15, "grad_norm": 6.834192276000977, "learning_rate": 2.926952141057935e-05, "loss": 0.3311, "step": 58 }, { "epoch": 0.15, "grad_norm": 9.958150863647461, "learning_rate": 2.9256926952141058e-05, "loss": 0.3635, "step": 59 }, { "epoch": 0.15, "grad_norm": 7.105954647064209, "learning_rate": 2.924433249370277e-05, "loss": 0.3664, "step": 60 }, { "epoch": 0.15, "grad_norm": 10.265762329101562, "learning_rate": 2.9231738035264486e-05, "loss": 0.4841, "step": 61 }, { "epoch": 0.16, "grad_norm": 7.81852912902832, "learning_rate": 2.92191435768262e-05, "loss": 0.2711, "step": 62 }, { "epoch": 0.16, "grad_norm": 8.933136940002441, "learning_rate": 2.9206549118387908e-05, "loss": 0.4243, "step": 63 }, { "epoch": 0.16, "grad_norm": 12.95289421081543, "learning_rate": 2.9193954659949623e-05, "loss": 0.3405, "step": 64 }, { "epoch": 0.16, "grad_norm": 6.353665828704834, "learning_rate": 2.9181360201511336e-05, "loss": 0.3388, "step": 65 }, { "epoch": 0.17, "grad_norm": 6.9620795249938965, "learning_rate": 2.9168765743073048e-05, "loss": 0.2494, "step": 66 }, { "epoch": 0.17, "grad_norm": 21.418628692626953, "learning_rate": 2.9156171284634764e-05, "loss": 0.7494, "step": 67 }, { "epoch": 0.17, "grad_norm": 11.0220308303833, "learning_rate": 2.9143576826196473e-05, "loss": 0.3143, "step": 68 }, { "epoch": 0.17, "grad_norm": 17.424291610717773, "learning_rate": 2.9130982367758185e-05, "loss": 0.9001, "step": 69 }, { "epoch": 0.18, "grad_norm": 20.80506134033203, "learning_rate": 2.91183879093199e-05, "loss": 0.6554, "step": 70 }, { "epoch": 0.18, "grad_norm": 10.28822135925293, "learning_rate": 2.9105793450881614e-05, "loss": 0.6043, "step": 71 }, { "epoch": 0.18, "grad_norm": 15.434713363647461, "learning_rate": 2.9093198992443326e-05, "loss": 0.4225, "step": 72 }, { "epoch": 0.18, "grad_norm": 12.180867195129395, "learning_rate": 2.908060453400504e-05, "loss": 0.3573, "step": 73 }, { "epoch": 0.19, "grad_norm": 11.896459579467773, "learning_rate": 2.906801007556675e-05, "loss": 0.5342, "step": 74 }, { "epoch": 0.19, "grad_norm": 7.4386444091796875, "learning_rate": 2.9055415617128463e-05, "loss": 0.4236, "step": 75 }, { "epoch": 0.19, "grad_norm": 11.402010917663574, "learning_rate": 2.904282115869018e-05, "loss": 0.3661, "step": 76 }, { "epoch": 0.19, "grad_norm": 9.488443374633789, "learning_rate": 2.903022670025189e-05, "loss": 0.7061, "step": 77 }, { "epoch": 0.2, "grad_norm": 15.988265991210938, "learning_rate": 2.90176322418136e-05, "loss": 0.6259, "step": 78 }, { "epoch": 0.2, "grad_norm": 11.976941108703613, "learning_rate": 2.9005037783375317e-05, "loss": 0.5461, "step": 79 }, { "epoch": 0.2, "grad_norm": 10.190409660339355, "learning_rate": 2.899244332493703e-05, "loss": 0.5816, "step": 80 }, { "epoch": 0.2, "grad_norm": 6.492282867431641, "learning_rate": 2.897984886649874e-05, "loss": 0.2346, "step": 81 }, { "epoch": 0.21, "grad_norm": 13.058635711669922, "learning_rate": 2.8967254408060454e-05, "loss": 0.3657, "step": 82 }, { "epoch": 0.21, "grad_norm": 14.294368743896484, "learning_rate": 2.8954659949622166e-05, "loss": 0.5426, "step": 83 }, { "epoch": 0.21, "grad_norm": 11.543447494506836, "learning_rate": 2.894206549118388e-05, "loss": 0.5771, "step": 84 }, { "epoch": 0.21, "grad_norm": 7.160464286804199, "learning_rate": 2.8929471032745595e-05, "loss": 0.3088, "step": 85 }, { "epoch": 0.22, "grad_norm": 16.53839683532715, "learning_rate": 2.8916876574307307e-05, "loss": 0.7966, "step": 86 }, { "epoch": 0.22, "grad_norm": 10.61913013458252, "learning_rate": 2.8904282115869016e-05, "loss": 0.507, "step": 87 }, { "epoch": 0.22, "grad_norm": 5.3180084228515625, "learning_rate": 2.8891687657430732e-05, "loss": 0.4556, "step": 88 }, { "epoch": 0.22, "grad_norm": 7.9845805168151855, "learning_rate": 2.8879093198992444e-05, "loss": 0.462, "step": 89 }, { "epoch": 0.23, "grad_norm": 19.176565170288086, "learning_rate": 2.8866498740554157e-05, "loss": 0.6049, "step": 90 }, { "epoch": 0.23, "grad_norm": 6.490673065185547, "learning_rate": 2.885390428211587e-05, "loss": 0.4461, "step": 91 }, { "epoch": 0.23, "grad_norm": 10.937054634094238, "learning_rate": 2.884130982367758e-05, "loss": 0.4222, "step": 92 }, { "epoch": 0.23, "grad_norm": 14.133673667907715, "learning_rate": 2.8828715365239294e-05, "loss": 0.6477, "step": 93 }, { "epoch": 0.24, "grad_norm": 7.424984931945801, "learning_rate": 2.881612090680101e-05, "loss": 0.3932, "step": 94 }, { "epoch": 0.24, "grad_norm": 6.378269195556641, "learning_rate": 2.8803526448362722e-05, "loss": 0.3421, "step": 95 }, { "epoch": 0.24, "grad_norm": 14.893821716308594, "learning_rate": 2.879093198992443e-05, "loss": 0.4213, "step": 96 }, { "epoch": 0.24, "grad_norm": 7.521154880523682, "learning_rate": 2.8778337531486147e-05, "loss": 0.3938, "step": 97 }, { "epoch": 0.25, "grad_norm": 7.131565093994141, "learning_rate": 2.876574307304786e-05, "loss": 0.4146, "step": 98 }, { "epoch": 0.25, "grad_norm": 17.402559280395508, "learning_rate": 2.8753148614609572e-05, "loss": 0.6407, "step": 99 }, { "epoch": 0.25, "grad_norm": 17.919559478759766, "learning_rate": 2.8740554156171288e-05, "loss": 0.5773, "step": 100 }, { "epoch": 0.25, "grad_norm": 18.85222625732422, "learning_rate": 2.8727959697732997e-05, "loss": 0.8411, "step": 101 }, { "epoch": 0.26, "grad_norm": 17.24286651611328, "learning_rate": 2.871536523929471e-05, "loss": 0.6594, "step": 102 }, { "epoch": 0.26, "grad_norm": 8.199654579162598, "learning_rate": 2.8702770780856425e-05, "loss": 0.2904, "step": 103 }, { "epoch": 0.26, "grad_norm": 11.349382400512695, "learning_rate": 2.8690176322418137e-05, "loss": 0.4303, "step": 104 }, { "epoch": 0.26, "grad_norm": 5.197649002075195, "learning_rate": 2.8677581863979846e-05, "loss": 0.3887, "step": 105 }, { "epoch": 0.27, "grad_norm": 6.405046463012695, "learning_rate": 2.8664987405541562e-05, "loss": 0.2129, "step": 106 }, { "epoch": 0.27, "grad_norm": 11.765724182128906, "learning_rate": 2.8652392947103275e-05, "loss": 0.6096, "step": 107 }, { "epoch": 0.27, "grad_norm": 15.173542022705078, "learning_rate": 2.8639798488664987e-05, "loss": 0.8199, "step": 108 }, { "epoch": 0.27, "grad_norm": 15.437769889831543, "learning_rate": 2.8627204030226703e-05, "loss": 0.6021, "step": 109 }, { "epoch": 0.28, "grad_norm": 8.98073673248291, "learning_rate": 2.8614609571788412e-05, "loss": 0.4355, "step": 110 }, { "epoch": 0.28, "grad_norm": 8.301546096801758, "learning_rate": 2.8602015113350128e-05, "loss": 0.6601, "step": 111 }, { "epoch": 0.28, "grad_norm": 14.387860298156738, "learning_rate": 2.858942065491184e-05, "loss": 0.482, "step": 112 }, { "epoch": 0.28, "grad_norm": 14.607189178466797, "learning_rate": 2.8576826196473553e-05, "loss": 0.5696, "step": 113 }, { "epoch": 0.29, "grad_norm": 8.80168628692627, "learning_rate": 2.8564231738035265e-05, "loss": 0.3216, "step": 114 }, { "epoch": 0.29, "grad_norm": 10.599414825439453, "learning_rate": 2.8551637279596978e-05, "loss": 0.6489, "step": 115 }, { "epoch": 0.29, "grad_norm": 11.09121322631836, "learning_rate": 2.853904282115869e-05, "loss": 0.2516, "step": 116 }, { "epoch": 0.29, "grad_norm": 7.300434112548828, "learning_rate": 2.8526448362720406e-05, "loss": 0.4725, "step": 117 }, { "epoch": 0.3, "grad_norm": 16.0921688079834, "learning_rate": 2.8513853904282118e-05, "loss": 0.6409, "step": 118 }, { "epoch": 0.3, "grad_norm": 6.426705360412598, "learning_rate": 2.8501259445843827e-05, "loss": 0.5262, "step": 119 }, { "epoch": 0.3, "grad_norm": 7.335501670837402, "learning_rate": 2.8488664987405543e-05, "loss": 0.4493, "step": 120 }, { "epoch": 0.3, "grad_norm": 7.7583794593811035, "learning_rate": 2.8476070528967256e-05, "loss": 0.4348, "step": 121 }, { "epoch": 0.31, "grad_norm": 6.631812572479248, "learning_rate": 2.8463476070528968e-05, "loss": 0.2344, "step": 122 }, { "epoch": 0.31, "grad_norm": 19.749597549438477, "learning_rate": 2.8450881612090684e-05, "loss": 0.6976, "step": 123 }, { "epoch": 0.31, "grad_norm": 8.363977432250977, "learning_rate": 2.8438287153652393e-05, "loss": 0.4216, "step": 124 }, { "epoch": 0.31, "grad_norm": 10.750311851501465, "learning_rate": 2.8425692695214105e-05, "loss": 0.5134, "step": 125 }, { "epoch": 0.32, "grad_norm": 8.392335891723633, "learning_rate": 2.841309823677582e-05, "loss": 0.485, "step": 126 }, { "epoch": 0.32, "grad_norm": 9.613751411437988, "learning_rate": 2.8400503778337533e-05, "loss": 0.5329, "step": 127 }, { "epoch": 0.32, "grad_norm": 5.709622859954834, "learning_rate": 2.8387909319899243e-05, "loss": 0.4024, "step": 128 }, { "epoch": 0.32, "grad_norm": 11.972945213317871, "learning_rate": 2.837531486146096e-05, "loss": 0.3772, "step": 129 }, { "epoch": 0.33, "grad_norm": 9.365653991699219, "learning_rate": 2.836272040302267e-05, "loss": 0.9204, "step": 130 }, { "epoch": 0.33, "grad_norm": 8.527847290039062, "learning_rate": 2.8350125944584383e-05, "loss": 0.3677, "step": 131 }, { "epoch": 0.33, "grad_norm": 10.193449020385742, "learning_rate": 2.83375314861461e-05, "loss": 0.4139, "step": 132 }, { "epoch": 0.34, "grad_norm": 9.360228538513184, "learning_rate": 2.8324937027707808e-05, "loss": 0.1871, "step": 133 }, { "epoch": 0.34, "grad_norm": 11.814716339111328, "learning_rate": 2.831234256926952e-05, "loss": 0.547, "step": 134 }, { "epoch": 0.34, "grad_norm": 6.990214824676514, "learning_rate": 2.8299748110831236e-05, "loss": 0.229, "step": 135 }, { "epoch": 0.34, "grad_norm": 9.345853805541992, "learning_rate": 2.828715365239295e-05, "loss": 0.3516, "step": 136 }, { "epoch": 0.35, "grad_norm": 8.358905792236328, "learning_rate": 2.827455919395466e-05, "loss": 0.4395, "step": 137 }, { "epoch": 0.35, "grad_norm": 4.4381327629089355, "learning_rate": 2.8261964735516374e-05, "loss": 0.4123, "step": 138 }, { "epoch": 0.35, "grad_norm": 5.183178424835205, "learning_rate": 2.8249370277078086e-05, "loss": 0.2701, "step": 139 }, { "epoch": 0.35, "grad_norm": 4.981905460357666, "learning_rate": 2.82367758186398e-05, "loss": 0.3501, "step": 140 }, { "epoch": 0.36, "grad_norm": 4.92559814453125, "learning_rate": 2.8224181360201514e-05, "loss": 0.3215, "step": 141 }, { "epoch": 0.36, "grad_norm": 10.248849868774414, "learning_rate": 2.8211586901763223e-05, "loss": 0.4752, "step": 142 }, { "epoch": 0.36, "grad_norm": 24.566904067993164, "learning_rate": 2.8198992443324936e-05, "loss": 0.4138, "step": 143 }, { "epoch": 0.36, "grad_norm": 14.266925811767578, "learning_rate": 2.818639798488665e-05, "loss": 0.2534, "step": 144 }, { "epoch": 0.37, "grad_norm": 12.045469284057617, "learning_rate": 2.8173803526448364e-05, "loss": 0.3652, "step": 145 }, { "epoch": 0.37, "grad_norm": 10.166648864746094, "learning_rate": 2.8161209068010076e-05, "loss": 0.4335, "step": 146 }, { "epoch": 0.37, "grad_norm": 6.525755882263184, "learning_rate": 2.814861460957179e-05, "loss": 0.3855, "step": 147 }, { "epoch": 0.37, "grad_norm": 7.986118316650391, "learning_rate": 2.81360201511335e-05, "loss": 0.3791, "step": 148 }, { "epoch": 0.38, "grad_norm": 10.436816215515137, "learning_rate": 2.8123425692695214e-05, "loss": 0.5131, "step": 149 }, { "epoch": 0.38, "grad_norm": 10.12028694152832, "learning_rate": 2.811083123425693e-05, "loss": 0.4801, "step": 150 }, { "epoch": 0.38, "grad_norm": 11.41160774230957, "learning_rate": 2.8098236775818642e-05, "loss": 0.4166, "step": 151 }, { "epoch": 0.38, "grad_norm": 11.20549201965332, "learning_rate": 2.808564231738035e-05, "loss": 0.5492, "step": 152 }, { "epoch": 0.39, "grad_norm": 8.953025817871094, "learning_rate": 2.8073047858942067e-05, "loss": 0.5637, "step": 153 }, { "epoch": 0.39, "grad_norm": 10.427567481994629, "learning_rate": 2.806045340050378e-05, "loss": 0.3726, "step": 154 }, { "epoch": 0.39, "grad_norm": 10.698348999023438, "learning_rate": 2.804785894206549e-05, "loss": 0.581, "step": 155 }, { "epoch": 0.39, "grad_norm": 11.016364097595215, "learning_rate": 2.8035264483627204e-05, "loss": 0.4332, "step": 156 }, { "epoch": 0.4, "grad_norm": 9.223053932189941, "learning_rate": 2.8022670025188917e-05, "loss": 0.325, "step": 157 }, { "epoch": 0.4, "grad_norm": 7.178832054138184, "learning_rate": 2.801007556675063e-05, "loss": 0.4933, "step": 158 }, { "epoch": 0.4, "grad_norm": 7.114950180053711, "learning_rate": 2.7997481108312345e-05, "loss": 0.3691, "step": 159 }, { "epoch": 0.4, "grad_norm": 11.535995483398438, "learning_rate": 2.7984886649874057e-05, "loss": 0.3785, "step": 160 }, { "epoch": 0.41, "grad_norm": 11.003711700439453, "learning_rate": 2.7972292191435766e-05, "loss": 0.458, "step": 161 }, { "epoch": 0.41, "grad_norm": 7.845949649810791, "learning_rate": 2.7959697732997482e-05, "loss": 0.3575, "step": 162 }, { "epoch": 0.41, "grad_norm": 7.582950592041016, "learning_rate": 2.7947103274559194e-05, "loss": 0.5969, "step": 163 }, { "epoch": 0.41, "grad_norm": 10.127944946289062, "learning_rate": 2.7934508816120907e-05, "loss": 0.4724, "step": 164 }, { "epoch": 0.42, "grad_norm": 7.671854496002197, "learning_rate": 2.7921914357682623e-05, "loss": 0.5269, "step": 165 }, { "epoch": 0.42, "grad_norm": 8.781110763549805, "learning_rate": 2.7909319899244332e-05, "loss": 0.7148, "step": 166 }, { "epoch": 0.42, "grad_norm": 8.324240684509277, "learning_rate": 2.7896725440806044e-05, "loss": 0.428, "step": 167 }, { "epoch": 0.42, "grad_norm": 9.713485717773438, "learning_rate": 2.788413098236776e-05, "loss": 0.5671, "step": 168 }, { "epoch": 0.43, "grad_norm": 6.988973617553711, "learning_rate": 2.7871536523929472e-05, "loss": 0.506, "step": 169 }, { "epoch": 0.43, "grad_norm": 19.869598388671875, "learning_rate": 2.785894206549118e-05, "loss": 0.5968, "step": 170 }, { "epoch": 0.43, "grad_norm": 10.117294311523438, "learning_rate": 2.7846347607052897e-05, "loss": 0.413, "step": 171 }, { "epoch": 0.43, "grad_norm": 7.587430477142334, "learning_rate": 2.783375314861461e-05, "loss": 0.4436, "step": 172 }, { "epoch": 0.44, "grad_norm": 3.9757165908813477, "learning_rate": 2.7821158690176322e-05, "loss": 0.2495, "step": 173 }, { "epoch": 0.44, "grad_norm": 6.600593566894531, "learning_rate": 2.7808564231738038e-05, "loss": 0.3974, "step": 174 }, { "epoch": 0.44, "grad_norm": 7.599936485290527, "learning_rate": 2.7795969773299747e-05, "loss": 0.4323, "step": 175 }, { "epoch": 0.44, "grad_norm": 8.428034782409668, "learning_rate": 2.778337531486146e-05, "loss": 0.48, "step": 176 }, { "epoch": 0.45, "grad_norm": 10.259221076965332, "learning_rate": 2.7770780856423175e-05, "loss": 0.4639, "step": 177 }, { "epoch": 0.45, "grad_norm": 8.270182609558105, "learning_rate": 2.7758186397984888e-05, "loss": 0.3487, "step": 178 }, { "epoch": 0.45, "grad_norm": 5.60026741027832, "learning_rate": 2.77455919395466e-05, "loss": 0.4958, "step": 179 }, { "epoch": 0.45, "grad_norm": 9.737464904785156, "learning_rate": 2.7732997481108313e-05, "loss": 0.4293, "step": 180 }, { "epoch": 0.46, "grad_norm": 10.6303129196167, "learning_rate": 2.7720403022670025e-05, "loss": 0.4395, "step": 181 }, { "epoch": 0.46, "grad_norm": 8.617947578430176, "learning_rate": 2.770780856423174e-05, "loss": 0.5972, "step": 182 }, { "epoch": 0.46, "grad_norm": 9.682026863098145, "learning_rate": 2.7695214105793453e-05, "loss": 0.2741, "step": 183 }, { "epoch": 0.46, "grad_norm": 5.532878398895264, "learning_rate": 2.7682619647355162e-05, "loss": 0.4699, "step": 184 }, { "epoch": 0.47, "grad_norm": 11.166542053222656, "learning_rate": 2.7670025188916878e-05, "loss": 0.5117, "step": 185 }, { "epoch": 0.47, "grad_norm": 6.78548526763916, "learning_rate": 2.765743073047859e-05, "loss": 0.3471, "step": 186 }, { "epoch": 0.47, "grad_norm": 12.514228820800781, "learning_rate": 2.7644836272040303e-05, "loss": 0.6523, "step": 187 }, { "epoch": 0.47, "grad_norm": 7.483336925506592, "learning_rate": 2.763224181360202e-05, "loss": 0.5714, "step": 188 }, { "epoch": 0.48, "grad_norm": 8.872393608093262, "learning_rate": 2.7619647355163728e-05, "loss": 0.4711, "step": 189 }, { "epoch": 0.48, "grad_norm": 5.779665946960449, "learning_rate": 2.760705289672544e-05, "loss": 0.3747, "step": 190 }, { "epoch": 0.48, "grad_norm": 6.670152187347412, "learning_rate": 2.7594458438287156e-05, "loss": 0.3524, "step": 191 }, { "epoch": 0.48, "grad_norm": 8.483373641967773, "learning_rate": 2.758186397984887e-05, "loss": 0.2449, "step": 192 }, { "epoch": 0.49, "grad_norm": 7.961008548736572, "learning_rate": 2.756926952141058e-05, "loss": 0.5184, "step": 193 }, { "epoch": 0.49, "grad_norm": 8.475528717041016, "learning_rate": 2.7556675062972293e-05, "loss": 0.4057, "step": 194 }, { "epoch": 0.49, "grad_norm": 5.382789611816406, "learning_rate": 2.7544080604534006e-05, "loss": 0.3483, "step": 195 }, { "epoch": 0.49, "grad_norm": 11.06598949432373, "learning_rate": 2.7531486146095718e-05, "loss": 0.4566, "step": 196 }, { "epoch": 0.5, "grad_norm": 7.172100067138672, "learning_rate": 2.7518891687657434e-05, "loss": 0.5069, "step": 197 }, { "epoch": 0.5, "grad_norm": 8.69959545135498, "learning_rate": 2.7506297229219143e-05, "loss": 0.4024, "step": 198 }, { "epoch": 0.5, "grad_norm": 3.8215749263763428, "learning_rate": 2.7493702770780855e-05, "loss": 0.1903, "step": 199 }, { "epoch": 0.5, "grad_norm": 6.7800726890563965, "learning_rate": 2.748110831234257e-05, "loss": 0.4091, "step": 200 }, { "epoch": 0.51, "grad_norm": 11.993806838989258, "learning_rate": 2.7468513853904284e-05, "loss": 0.4989, "step": 201 }, { "epoch": 0.51, "grad_norm": 14.418086051940918, "learning_rate": 2.7455919395465996e-05, "loss": 0.5218, "step": 202 }, { "epoch": 0.51, "grad_norm": 9.064515113830566, "learning_rate": 2.744332493702771e-05, "loss": 0.357, "step": 203 }, { "epoch": 0.51, "grad_norm": 4.245691299438477, "learning_rate": 2.743073047858942e-05, "loss": 0.3012, "step": 204 }, { "epoch": 0.52, "grad_norm": 10.50638198852539, "learning_rate": 2.7418136020151133e-05, "loss": 0.3103, "step": 205 }, { "epoch": 0.52, "grad_norm": 7.120365619659424, "learning_rate": 2.740554156171285e-05, "loss": 0.3018, "step": 206 }, { "epoch": 0.52, "grad_norm": 14.648916244506836, "learning_rate": 2.7392947103274562e-05, "loss": 0.8458, "step": 207 }, { "epoch": 0.52, "grad_norm": 11.240401268005371, "learning_rate": 2.738035264483627e-05, "loss": 0.2182, "step": 208 }, { "epoch": 0.53, "grad_norm": 15.59676456451416, "learning_rate": 2.7367758186397987e-05, "loss": 0.2614, "step": 209 }, { "epoch": 0.53, "grad_norm": 2.611619472503662, "learning_rate": 2.73551637279597e-05, "loss": 0.0725, "step": 210 }, { "epoch": 0.53, "grad_norm": 15.017423629760742, "learning_rate": 2.734256926952141e-05, "loss": 0.6751, "step": 211 }, { "epoch": 0.53, "grad_norm": 8.739672660827637, "learning_rate": 2.7329974811083124e-05, "loss": 0.1393, "step": 212 }, { "epoch": 0.54, "grad_norm": 13.18393611907959, "learning_rate": 2.7317380352644836e-05, "loss": 0.4262, "step": 213 }, { "epoch": 0.54, "grad_norm": 15.25338363647461, "learning_rate": 2.730478589420655e-05, "loss": 0.4582, "step": 214 }, { "epoch": 0.54, "grad_norm": 4.455483913421631, "learning_rate": 2.7292191435768265e-05, "loss": 0.2272, "step": 215 }, { "epoch": 0.54, "grad_norm": 5.5019145011901855, "learning_rate": 2.7279596977329977e-05, "loss": 0.2219, "step": 216 }, { "epoch": 0.55, "grad_norm": 6.36189603805542, "learning_rate": 2.7267002518891686e-05, "loss": 0.3612, "step": 217 }, { "epoch": 0.55, "grad_norm": 11.216201782226562, "learning_rate": 2.7254408060453402e-05, "loss": 0.4951, "step": 218 }, { "epoch": 0.55, "grad_norm": 7.516535758972168, "learning_rate": 2.7241813602015114e-05, "loss": 0.3473, "step": 219 }, { "epoch": 0.55, "grad_norm": 12.50316047668457, "learning_rate": 2.7229219143576827e-05, "loss": 0.5692, "step": 220 }, { "epoch": 0.56, "grad_norm": 7.9396562576293945, "learning_rate": 2.721662468513854e-05, "loss": 0.3818, "step": 221 }, { "epoch": 0.56, "grad_norm": 10.905122756958008, "learning_rate": 2.720403022670025e-05, "loss": 0.1508, "step": 222 }, { "epoch": 0.56, "grad_norm": 11.742682456970215, "learning_rate": 2.7191435768261964e-05, "loss": 0.5067, "step": 223 }, { "epoch": 0.56, "grad_norm": 9.895332336425781, "learning_rate": 2.717884130982368e-05, "loss": 0.4384, "step": 224 }, { "epoch": 0.57, "grad_norm": 13.22032642364502, "learning_rate": 2.7166246851385392e-05, "loss": 0.2769, "step": 225 }, { "epoch": 0.57, "grad_norm": 8.62972354888916, "learning_rate": 2.71536523929471e-05, "loss": 0.141, "step": 226 }, { "epoch": 0.57, "grad_norm": 13.368087768554688, "learning_rate": 2.7141057934508817e-05, "loss": 0.485, "step": 227 }, { "epoch": 0.57, "grad_norm": 5.622757911682129, "learning_rate": 2.712846347607053e-05, "loss": 0.3319, "step": 228 }, { "epoch": 0.58, "grad_norm": 6.637231349945068, "learning_rate": 2.7115869017632242e-05, "loss": 0.5557, "step": 229 }, { "epoch": 0.58, "grad_norm": 9.410402297973633, "learning_rate": 2.7103274559193958e-05, "loss": 0.4845, "step": 230 }, { "epoch": 0.58, "grad_norm": 18.66705894470215, "learning_rate": 2.7090680100755667e-05, "loss": 0.9765, "step": 231 }, { "epoch": 0.58, "grad_norm": 4.065969467163086, "learning_rate": 2.707808564231738e-05, "loss": 0.2138, "step": 232 }, { "epoch": 0.59, "grad_norm": 12.393918991088867, "learning_rate": 2.7065491183879095e-05, "loss": 0.6254, "step": 233 }, { "epoch": 0.59, "grad_norm": 13.665956497192383, "learning_rate": 2.7052896725440807e-05, "loss": 0.6912, "step": 234 }, { "epoch": 0.59, "grad_norm": 6.5648345947265625, "learning_rate": 2.7040302267002517e-05, "loss": 0.4344, "step": 235 }, { "epoch": 0.59, "grad_norm": 4.4842529296875, "learning_rate": 2.7027707808564232e-05, "loss": 0.3209, "step": 236 }, { "epoch": 0.6, "grad_norm": 3.8340840339660645, "learning_rate": 2.7015113350125945e-05, "loss": 0.4045, "step": 237 }, { "epoch": 0.6, "grad_norm": 10.3326997756958, "learning_rate": 2.7002518891687657e-05, "loss": 0.4998, "step": 238 }, { "epoch": 0.6, "grad_norm": 8.519319534301758, "learning_rate": 2.6989924433249373e-05, "loss": 0.3126, "step": 239 }, { "epoch": 0.6, "grad_norm": 5.563427448272705, "learning_rate": 2.6977329974811082e-05, "loss": 0.2947, "step": 240 }, { "epoch": 0.61, "grad_norm": 4.757438659667969, "learning_rate": 2.6964735516372794e-05, "loss": 0.3535, "step": 241 }, { "epoch": 0.61, "grad_norm": 7.1805419921875, "learning_rate": 2.695214105793451e-05, "loss": 0.4377, "step": 242 }, { "epoch": 0.61, "grad_norm": 8.782963752746582, "learning_rate": 2.6939546599496223e-05, "loss": 0.5135, "step": 243 }, { "epoch": 0.61, "grad_norm": 6.521146774291992, "learning_rate": 2.6926952141057935e-05, "loss": 0.4571, "step": 244 }, { "epoch": 0.62, "grad_norm": 9.651432991027832, "learning_rate": 2.6914357682619648e-05, "loss": 0.3851, "step": 245 }, { "epoch": 0.62, "grad_norm": 13.755444526672363, "learning_rate": 2.690176322418136e-05, "loss": 0.6536, "step": 246 }, { "epoch": 0.62, "grad_norm": 6.192646503448486, "learning_rate": 2.6889168765743072e-05, "loss": 0.3358, "step": 247 }, { "epoch": 0.62, "grad_norm": 10.30482292175293, "learning_rate": 2.6876574307304788e-05, "loss": 0.4194, "step": 248 }, { "epoch": 0.63, "grad_norm": 6.788898944854736, "learning_rate": 2.6863979848866497e-05, "loss": 0.5712, "step": 249 }, { "epoch": 0.63, "grad_norm": 10.21720027923584, "learning_rate": 2.685138539042821e-05, "loss": 0.3626, "step": 250 }, { "epoch": 0.63, "grad_norm": 7.160099029541016, "learning_rate": 2.6838790931989926e-05, "loss": 0.26, "step": 251 }, { "epoch": 0.63, "grad_norm": 10.605951309204102, "learning_rate": 2.6826196473551638e-05, "loss": 0.7586, "step": 252 }, { "epoch": 0.64, "grad_norm": 10.67288589477539, "learning_rate": 2.681360201511335e-05, "loss": 0.5889, "step": 253 }, { "epoch": 0.64, "grad_norm": 5.152278423309326, "learning_rate": 2.6801007556675063e-05, "loss": 0.2835, "step": 254 }, { "epoch": 0.64, "grad_norm": 6.4969563484191895, "learning_rate": 2.6788413098236775e-05, "loss": 0.3987, "step": 255 }, { "epoch": 0.64, "grad_norm": 8.223746299743652, "learning_rate": 2.677581863979849e-05, "loss": 0.3398, "step": 256 }, { "epoch": 0.65, "grad_norm": 13.44676399230957, "learning_rate": 2.6763224181360204e-05, "loss": 0.5064, "step": 257 }, { "epoch": 0.65, "grad_norm": 8.834419250488281, "learning_rate": 2.6750629722921916e-05, "loss": 0.4247, "step": 258 }, { "epoch": 0.65, "grad_norm": 7.546679973602295, "learning_rate": 2.673803526448363e-05, "loss": 0.3359, "step": 259 }, { "epoch": 0.65, "grad_norm": 11.630725860595703, "learning_rate": 2.672544080604534e-05, "loss": 0.4661, "step": 260 }, { "epoch": 0.66, "grad_norm": 6.135303497314453, "learning_rate": 2.6712846347607053e-05, "loss": 0.2913, "step": 261 }, { "epoch": 0.66, "grad_norm": 7.024556636810303, "learning_rate": 2.670025188916877e-05, "loss": 0.2892, "step": 262 }, { "epoch": 0.66, "grad_norm": 11.134332656860352, "learning_rate": 2.6687657430730478e-05, "loss": 0.2884, "step": 263 }, { "epoch": 0.66, "grad_norm": 19.467782974243164, "learning_rate": 2.667506297229219e-05, "loss": 0.5973, "step": 264 }, { "epoch": 0.67, "grad_norm": 6.133577346801758, "learning_rate": 2.6662468513853906e-05, "loss": 0.4258, "step": 265 }, { "epoch": 0.67, "grad_norm": 12.55841064453125, "learning_rate": 2.664987405541562e-05, "loss": 0.4355, "step": 266 }, { "epoch": 0.67, "grad_norm": 17.070459365844727, "learning_rate": 2.663727959697733e-05, "loss": 0.7106, "step": 267 }, { "epoch": 0.68, "grad_norm": 13.686192512512207, "learning_rate": 2.6624685138539044e-05, "loss": 0.526, "step": 268 }, { "epoch": 0.68, "grad_norm": 11.354989051818848, "learning_rate": 2.6612090680100756e-05, "loss": 0.3703, "step": 269 }, { "epoch": 0.68, "grad_norm": 9.988239288330078, "learning_rate": 2.659949622166247e-05, "loss": 0.4727, "step": 270 }, { "epoch": 0.68, "grad_norm": 8.656394958496094, "learning_rate": 2.6586901763224184e-05, "loss": 0.1363, "step": 271 }, { "epoch": 0.69, "grad_norm": 7.235680103302002, "learning_rate": 2.6574307304785897e-05, "loss": 0.2889, "step": 272 }, { "epoch": 0.69, "grad_norm": 18.26453399658203, "learning_rate": 2.6561712846347606e-05, "loss": 0.6898, "step": 273 }, { "epoch": 0.69, "grad_norm": 8.426389694213867, "learning_rate": 2.654911838790932e-05, "loss": 0.3522, "step": 274 }, { "epoch": 0.69, "grad_norm": 9.811342239379883, "learning_rate": 2.6536523929471034e-05, "loss": 0.5926, "step": 275 }, { "epoch": 0.7, "grad_norm": 6.944952011108398, "learning_rate": 2.6523929471032746e-05, "loss": 0.281, "step": 276 }, { "epoch": 0.7, "grad_norm": 7.4408979415893555, "learning_rate": 2.651133501259446e-05, "loss": 0.5496, "step": 277 }, { "epoch": 0.7, "grad_norm": 8.388789176940918, "learning_rate": 2.649874055415617e-05, "loss": 0.2775, "step": 278 }, { "epoch": 0.7, "grad_norm": 4.758674144744873, "learning_rate": 2.6486146095717884e-05, "loss": 0.4558, "step": 279 }, { "epoch": 0.71, "grad_norm": 6.666047096252441, "learning_rate": 2.64735516372796e-05, "loss": 0.4307, "step": 280 }, { "epoch": 0.71, "grad_norm": 13.093096733093262, "learning_rate": 2.6460957178841312e-05, "loss": 0.5806, "step": 281 }, { "epoch": 0.71, "grad_norm": 12.456526756286621, "learning_rate": 2.644836272040302e-05, "loss": 0.5669, "step": 282 }, { "epoch": 0.71, "grad_norm": 6.2549004554748535, "learning_rate": 2.6435768261964737e-05, "loss": 0.3338, "step": 283 }, { "epoch": 0.72, "grad_norm": 6.807823657989502, "learning_rate": 2.642317380352645e-05, "loss": 0.5899, "step": 284 }, { "epoch": 0.72, "grad_norm": 5.582499027252197, "learning_rate": 2.641057934508816e-05, "loss": 0.4312, "step": 285 }, { "epoch": 0.72, "grad_norm": 7.518726348876953, "learning_rate": 2.6397984886649878e-05, "loss": 0.3892, "step": 286 }, { "epoch": 0.72, "grad_norm": 5.416556358337402, "learning_rate": 2.6385390428211587e-05, "loss": 0.2505, "step": 287 }, { "epoch": 0.73, "grad_norm": 6.679108619689941, "learning_rate": 2.63727959697733e-05, "loss": 0.4348, "step": 288 }, { "epoch": 0.73, "grad_norm": 4.240274429321289, "learning_rate": 2.6360201511335015e-05, "loss": 0.5039, "step": 289 }, { "epoch": 0.73, "grad_norm": 9.389612197875977, "learning_rate": 2.6347607052896727e-05, "loss": 0.5208, "step": 290 }, { "epoch": 0.73, "grad_norm": 7.322227478027344, "learning_rate": 2.6335012594458436e-05, "loss": 0.5255, "step": 291 }, { "epoch": 0.74, "grad_norm": 6.759903430938721, "learning_rate": 2.6322418136020152e-05, "loss": 0.3596, "step": 292 }, { "epoch": 0.74, "grad_norm": 4.8109822273254395, "learning_rate": 2.6309823677581865e-05, "loss": 0.2835, "step": 293 }, { "epoch": 0.74, "grad_norm": 4.74676513671875, "learning_rate": 2.6297229219143577e-05, "loss": 0.321, "step": 294 }, { "epoch": 0.74, "grad_norm": 8.768762588500977, "learning_rate": 2.6284634760705293e-05, "loss": 0.3736, "step": 295 }, { "epoch": 0.75, "grad_norm": 6.698115825653076, "learning_rate": 2.6272040302267002e-05, "loss": 0.2979, "step": 296 }, { "epoch": 0.75, "grad_norm": 9.396990776062012, "learning_rate": 2.6259445843828714e-05, "loss": 0.6949, "step": 297 }, { "epoch": 0.75, "grad_norm": 6.292881011962891, "learning_rate": 2.624685138539043e-05, "loss": 0.5847, "step": 298 }, { "epoch": 0.75, "grad_norm": 5.798967361450195, "learning_rate": 2.6234256926952142e-05, "loss": 0.413, "step": 299 }, { "epoch": 0.76, "grad_norm": 8.518070220947266, "learning_rate": 2.6221662468513855e-05, "loss": 0.4463, "step": 300 }, { "epoch": 0.76, "grad_norm": 7.179047584533691, "learning_rate": 2.6209068010075567e-05, "loss": 0.4101, "step": 301 }, { "epoch": 0.76, "grad_norm": 8.893898963928223, "learning_rate": 2.619647355163728e-05, "loss": 0.3712, "step": 302 }, { "epoch": 0.76, "grad_norm": 7.180310249328613, "learning_rate": 2.6183879093198992e-05, "loss": 0.4253, "step": 303 }, { "epoch": 0.77, "grad_norm": 11.03390884399414, "learning_rate": 2.6171284634760708e-05, "loss": 0.3372, "step": 304 }, { "epoch": 0.77, "grad_norm": 6.37282133102417, "learning_rate": 2.6158690176322417e-05, "loss": 0.2729, "step": 305 }, { "epoch": 0.77, "grad_norm": 8.934061050415039, "learning_rate": 2.614609571788413e-05, "loss": 0.2433, "step": 306 }, { "epoch": 0.77, "grad_norm": 7.522178649902344, "learning_rate": 2.6133501259445845e-05, "loss": 0.4284, "step": 307 }, { "epoch": 0.78, "grad_norm": 8.500354766845703, "learning_rate": 2.6120906801007558e-05, "loss": 0.3639, "step": 308 }, { "epoch": 0.78, "grad_norm": 11.417220115661621, "learning_rate": 2.610831234256927e-05, "loss": 0.7663, "step": 309 }, { "epoch": 0.78, "grad_norm": 6.737428665161133, "learning_rate": 2.6095717884130983e-05, "loss": 0.2624, "step": 310 }, { "epoch": 0.78, "grad_norm": 6.082789897918701, "learning_rate": 2.6083123425692695e-05, "loss": 0.4592, "step": 311 }, { "epoch": 0.79, "grad_norm": 13.96414566040039, "learning_rate": 2.6070528967254407e-05, "loss": 0.7136, "step": 312 }, { "epoch": 0.79, "grad_norm": 14.575349807739258, "learning_rate": 2.6057934508816123e-05, "loss": 0.5151, "step": 313 }, { "epoch": 0.79, "grad_norm": 9.101264953613281, "learning_rate": 2.6045340050377836e-05, "loss": 0.3629, "step": 314 }, { "epoch": 0.79, "grad_norm": 14.307868957519531, "learning_rate": 2.6032745591939545e-05, "loss": 0.2515, "step": 315 }, { "epoch": 0.8, "grad_norm": 4.331357955932617, "learning_rate": 2.602015113350126e-05, "loss": 0.1049, "step": 316 }, { "epoch": 0.8, "grad_norm": 10.65372371673584, "learning_rate": 2.6007556675062973e-05, "loss": 0.7491, "step": 317 }, { "epoch": 0.8, "grad_norm": 7.100830554962158, "learning_rate": 2.5994962216624685e-05, "loss": 0.2969, "step": 318 }, { "epoch": 0.8, "grad_norm": 8.876154899597168, "learning_rate": 2.5982367758186398e-05, "loss": 0.2077, "step": 319 }, { "epoch": 0.81, "grad_norm": 9.560657501220703, "learning_rate": 2.596977329974811e-05, "loss": 0.6008, "step": 320 }, { "epoch": 0.81, "grad_norm": 6.376749515533447, "learning_rate": 2.5957178841309823e-05, "loss": 0.2282, "step": 321 }, { "epoch": 0.81, "grad_norm": 7.339711666107178, "learning_rate": 2.594458438287154e-05, "loss": 0.3252, "step": 322 }, { "epoch": 0.81, "grad_norm": 4.212242126464844, "learning_rate": 2.593198992443325e-05, "loss": 0.1242, "step": 323 }, { "epoch": 0.82, "grad_norm": 12.86931037902832, "learning_rate": 2.5919395465994963e-05, "loss": 0.5819, "step": 324 }, { "epoch": 0.82, "grad_norm": 9.049539566040039, "learning_rate": 2.5906801007556676e-05, "loss": 0.5282, "step": 325 }, { "epoch": 0.82, "grad_norm": 7.5245490074157715, "learning_rate": 2.5894206549118388e-05, "loss": 0.2818, "step": 326 }, { "epoch": 0.82, "grad_norm": 5.835328578948975, "learning_rate": 2.5881612090680104e-05, "loss": 0.2127, "step": 327 }, { "epoch": 0.83, "grad_norm": 6.4401469230651855, "learning_rate": 2.5869017632241813e-05, "loss": 0.327, "step": 328 }, { "epoch": 0.83, "grad_norm": 8.760528564453125, "learning_rate": 2.5856423173803526e-05, "loss": 0.533, "step": 329 }, { "epoch": 0.83, "grad_norm": 8.962298393249512, "learning_rate": 2.584382871536524e-05, "loss": 0.6162, "step": 330 }, { "epoch": 0.83, "grad_norm": 4.031172275543213, "learning_rate": 2.5831234256926954e-05, "loss": 0.2266, "step": 331 }, { "epoch": 0.84, "grad_norm": 12.244653701782227, "learning_rate": 2.5818639798488666e-05, "loss": 0.5697, "step": 332 }, { "epoch": 0.84, "grad_norm": 4.518488883972168, "learning_rate": 2.580604534005038e-05, "loss": 0.1391, "step": 333 }, { "epoch": 0.84, "grad_norm": 4.96370792388916, "learning_rate": 2.579345088161209e-05, "loss": 0.2026, "step": 334 }, { "epoch": 0.84, "grad_norm": 8.518425941467285, "learning_rate": 2.5780856423173803e-05, "loss": 0.392, "step": 335 }, { "epoch": 0.85, "grad_norm": 8.354524612426758, "learning_rate": 2.576826196473552e-05, "loss": 0.3324, "step": 336 }, { "epoch": 0.85, "grad_norm": 5.117928981781006, "learning_rate": 2.5755667506297232e-05, "loss": 0.3849, "step": 337 }, { "epoch": 0.85, "grad_norm": 5.336648464202881, "learning_rate": 2.574307304785894e-05, "loss": 0.241, "step": 338 }, { "epoch": 0.85, "grad_norm": 3.6382648944854736, "learning_rate": 2.5730478589420657e-05, "loss": 0.3806, "step": 339 }, { "epoch": 0.86, "grad_norm": 9.698990821838379, "learning_rate": 2.571788413098237e-05, "loss": 0.4922, "step": 340 }, { "epoch": 0.86, "grad_norm": 7.453464984893799, "learning_rate": 2.570528967254408e-05, "loss": 0.323, "step": 341 }, { "epoch": 0.86, "grad_norm": 5.546812534332275, "learning_rate": 2.5692695214105794e-05, "loss": 0.17, "step": 342 }, { "epoch": 0.86, "grad_norm": 4.6894941329956055, "learning_rate": 2.5680100755667506e-05, "loss": 0.3012, "step": 343 }, { "epoch": 0.87, "grad_norm": 12.336396217346191, "learning_rate": 2.566750629722922e-05, "loss": 0.7455, "step": 344 }, { "epoch": 0.87, "grad_norm": 10.94264030456543, "learning_rate": 2.5654911838790935e-05, "loss": 0.2863, "step": 345 }, { "epoch": 0.87, "grad_norm": 12.642402648925781, "learning_rate": 2.5642317380352647e-05, "loss": 0.2087, "step": 346 }, { "epoch": 0.87, "grad_norm": 1.6862198114395142, "learning_rate": 2.5629722921914356e-05, "loss": 0.0703, "step": 347 }, { "epoch": 0.88, "grad_norm": 6.8891191482543945, "learning_rate": 2.5617128463476072e-05, "loss": 0.3455, "step": 348 }, { "epoch": 0.88, "grad_norm": 7.763698577880859, "learning_rate": 2.5604534005037784e-05, "loss": 0.3147, "step": 349 }, { "epoch": 0.88, "grad_norm": 12.52840518951416, "learning_rate": 2.5591939546599497e-05, "loss": 0.6684, "step": 350 }, { "epoch": 0.88, "grad_norm": 11.723652839660645, "learning_rate": 2.5579345088161213e-05, "loss": 0.3871, "step": 351 }, { "epoch": 0.89, "grad_norm": 9.9241361618042, "learning_rate": 2.556675062972292e-05, "loss": 0.3243, "step": 352 }, { "epoch": 0.89, "grad_norm": 8.628238677978516, "learning_rate": 2.5554156171284634e-05, "loss": 0.82, "step": 353 }, { "epoch": 0.89, "grad_norm": 4.368703365325928, "learning_rate": 2.554156171284635e-05, "loss": 0.3554, "step": 354 }, { "epoch": 0.89, "grad_norm": 4.239711761474609, "learning_rate": 2.5528967254408062e-05, "loss": 0.1529, "step": 355 }, { "epoch": 0.9, "grad_norm": 9.269442558288574, "learning_rate": 2.551637279596977e-05, "loss": 0.4518, "step": 356 }, { "epoch": 0.9, "grad_norm": 5.514954566955566, "learning_rate": 2.5503778337531487e-05, "loss": 0.1421, "step": 357 }, { "epoch": 0.9, "grad_norm": 7.544525146484375, "learning_rate": 2.54911838790932e-05, "loss": 0.6922, "step": 358 }, { "epoch": 0.9, "grad_norm": 4.167691707611084, "learning_rate": 2.5478589420654912e-05, "loss": 0.0966, "step": 359 }, { "epoch": 0.91, "grad_norm": 7.282542705535889, "learning_rate": 2.5465994962216628e-05, "loss": 0.494, "step": 360 }, { "epoch": 0.91, "grad_norm": 12.464173316955566, "learning_rate": 2.5453400503778337e-05, "loss": 0.4292, "step": 361 }, { "epoch": 0.91, "grad_norm": 10.870161056518555, "learning_rate": 2.544080604534005e-05, "loss": 0.625, "step": 362 }, { "epoch": 0.91, "grad_norm": 8.240769386291504, "learning_rate": 2.5428211586901765e-05, "loss": 0.7351, "step": 363 }, { "epoch": 0.92, "grad_norm": 9.801168441772461, "learning_rate": 2.5415617128463477e-05, "loss": 0.2285, "step": 364 }, { "epoch": 0.92, "grad_norm": 11.260651588439941, "learning_rate": 2.540302267002519e-05, "loss": 0.4224, "step": 365 }, { "epoch": 0.92, "grad_norm": 14.770816802978516, "learning_rate": 2.5390428211586902e-05, "loss": 0.6083, "step": 366 }, { "epoch": 0.92, "grad_norm": 12.272558212280273, "learning_rate": 2.5377833753148615e-05, "loss": 0.4445, "step": 367 }, { "epoch": 0.93, "grad_norm": 6.899696350097656, "learning_rate": 2.5365239294710327e-05, "loss": 0.3629, "step": 368 }, { "epoch": 0.93, "grad_norm": 8.54549789428711, "learning_rate": 2.5352644836272043e-05, "loss": 0.4865, "step": 369 }, { "epoch": 0.93, "grad_norm": 8.818639755249023, "learning_rate": 2.5340050377833752e-05, "loss": 0.4459, "step": 370 }, { "epoch": 0.93, "grad_norm": 10.441351890563965, "learning_rate": 2.5327455919395464e-05, "loss": 0.3823, "step": 371 }, { "epoch": 0.94, "grad_norm": 10.242267608642578, "learning_rate": 2.531486146095718e-05, "loss": 0.4975, "step": 372 }, { "epoch": 0.94, "grad_norm": 9.25827407836914, "learning_rate": 2.5302267002518893e-05, "loss": 0.3961, "step": 373 }, { "epoch": 0.94, "grad_norm": 5.644846439361572, "learning_rate": 2.5289672544080605e-05, "loss": 0.1543, "step": 374 }, { "epoch": 0.94, "grad_norm": 8.441045761108398, "learning_rate": 2.5277078085642318e-05, "loss": 0.3673, "step": 375 }, { "epoch": 0.95, "grad_norm": 9.566481590270996, "learning_rate": 2.526448362720403e-05, "loss": 0.3034, "step": 376 }, { "epoch": 0.95, "grad_norm": 11.111581802368164, "learning_rate": 2.5251889168765742e-05, "loss": 0.4208, "step": 377 }, { "epoch": 0.95, "grad_norm": 10.74457836151123, "learning_rate": 2.5239294710327458e-05, "loss": 0.5335, "step": 378 }, { "epoch": 0.95, "grad_norm": 7.5216450691223145, "learning_rate": 2.522670025188917e-05, "loss": 0.4418, "step": 379 }, { "epoch": 0.96, "grad_norm": 10.419405937194824, "learning_rate": 2.521410579345088e-05, "loss": 0.4507, "step": 380 }, { "epoch": 0.96, "grad_norm": 6.768444061279297, "learning_rate": 2.5201511335012596e-05, "loss": 0.6152, "step": 381 }, { "epoch": 0.96, "grad_norm": 9.090213775634766, "learning_rate": 2.5188916876574308e-05, "loss": 0.6423, "step": 382 }, { "epoch": 0.96, "grad_norm": 7.854204177856445, "learning_rate": 2.517632241813602e-05, "loss": 0.3511, "step": 383 }, { "epoch": 0.97, "grad_norm": 8.918037414550781, "learning_rate": 2.5163727959697733e-05, "loss": 0.4304, "step": 384 }, { "epoch": 0.97, "grad_norm": 7.300523281097412, "learning_rate": 2.5151133501259445e-05, "loss": 0.2343, "step": 385 }, { "epoch": 0.97, "grad_norm": 5.29052734375, "learning_rate": 2.5138539042821158e-05, "loss": 0.2357, "step": 386 }, { "epoch": 0.97, "grad_norm": 13.124969482421875, "learning_rate": 2.5125944584382874e-05, "loss": 0.4188, "step": 387 }, { "epoch": 0.98, "grad_norm": 13.79240894317627, "learning_rate": 2.5113350125944586e-05, "loss": 0.4014, "step": 388 }, { "epoch": 0.98, "grad_norm": 5.569485664367676, "learning_rate": 2.5100755667506295e-05, "loss": 0.2489, "step": 389 }, { "epoch": 0.98, "grad_norm": 9.553081512451172, "learning_rate": 2.508816120906801e-05, "loss": 0.3377, "step": 390 }, { "epoch": 0.98, "grad_norm": 13.31069564819336, "learning_rate": 2.5075566750629723e-05, "loss": 0.3421, "step": 391 }, { "epoch": 0.99, "grad_norm": 8.653861999511719, "learning_rate": 2.5062972292191436e-05, "loss": 0.3805, "step": 392 }, { "epoch": 0.99, "grad_norm": 8.786714553833008, "learning_rate": 2.505037783375315e-05, "loss": 0.2799, "step": 393 }, { "epoch": 0.99, "grad_norm": 6.493752479553223, "learning_rate": 2.503778337531486e-05, "loss": 0.2796, "step": 394 }, { "epoch": 0.99, "grad_norm": 6.619555473327637, "learning_rate": 2.5025188916876573e-05, "loss": 0.2407, "step": 395 }, { "epoch": 1.0, "grad_norm": 6.245583534240723, "learning_rate": 2.501259445843829e-05, "loss": 0.2345, "step": 396 }, { "epoch": 1.0, "grad_norm": 97.71422576904297, "learning_rate": 2.5e-05, "loss": 1.7921, "step": 397 }, { "epoch": 1.0, "eval_accuracy": 0.8195583596214511, "eval_f1": 0.818793132119398, "eval_loss": 0.44511958956718445, "eval_precision": 0.8183509587189882, "eval_recall": 0.8213043799940811, "eval_runtime": 4442.9836, "eval_samples_per_second": 0.357, "eval_steps_per_second": 0.023, "step": 397 }, { "epoch": 1.0, "grad_norm": 7.2050909996032715, "learning_rate": 2.4987405541561714e-05, "loss": 0.2411, "step": 398 }, { "epoch": 1.01, "grad_norm": 5.562217712402344, "learning_rate": 2.4974811083123426e-05, "loss": 0.2874, "step": 399 }, { "epoch": 1.01, "grad_norm": 10.240361213684082, "learning_rate": 2.496221662468514e-05, "loss": 0.2287, "step": 400 }, { "epoch": 1.01, "grad_norm": 11.001642227172852, "learning_rate": 2.4949622166246854e-05, "loss": 0.7281, "step": 401 }, { "epoch": 1.01, "grad_norm": 14.056950569152832, "learning_rate": 2.4937027707808567e-05, "loss": 0.3952, "step": 402 }, { "epoch": 1.02, "grad_norm": 7.004691123962402, "learning_rate": 2.4924433249370276e-05, "loss": 0.3898, "step": 403 }, { "epoch": 1.02, "grad_norm": 5.306135654449463, "learning_rate": 2.491183879093199e-05, "loss": 0.3549, "step": 404 }, { "epoch": 1.02, "grad_norm": 7.861663818359375, "learning_rate": 2.4899244332493704e-05, "loss": 0.2773, "step": 405 }, { "epoch": 1.02, "grad_norm": 3.698382616043091, "learning_rate": 2.4886649874055416e-05, "loss": 0.2846, "step": 406 }, { "epoch": 1.03, "grad_norm": 14.958970069885254, "learning_rate": 2.4874055415617132e-05, "loss": 0.2251, "step": 407 }, { "epoch": 1.03, "grad_norm": 8.70470905303955, "learning_rate": 2.486146095717884e-05, "loss": 0.3024, "step": 408 }, { "epoch": 1.03, "grad_norm": 7.765173435211182, "learning_rate": 2.4848866498740554e-05, "loss": 0.346, "step": 409 }, { "epoch": 1.03, "grad_norm": 19.742538452148438, "learning_rate": 2.483627204030227e-05, "loss": 0.459, "step": 410 }, { "epoch": 1.04, "grad_norm": 8.575105667114258, "learning_rate": 2.4823677581863982e-05, "loss": 0.4486, "step": 411 }, { "epoch": 1.04, "grad_norm": 21.82974624633789, "learning_rate": 2.481108312342569e-05, "loss": 0.4693, "step": 412 }, { "epoch": 1.04, "grad_norm": 6.933416366577148, "learning_rate": 2.4798488664987407e-05, "loss": 0.2311, "step": 413 }, { "epoch": 1.04, "grad_norm": 4.775768756866455, "learning_rate": 2.478589420654912e-05, "loss": 0.2769, "step": 414 }, { "epoch": 1.05, "grad_norm": 1.4635595083236694, "learning_rate": 2.4773299748110832e-05, "loss": 0.0557, "step": 415 }, { "epoch": 1.05, "grad_norm": 17.7808895111084, "learning_rate": 2.4760705289672548e-05, "loss": 0.6811, "step": 416 }, { "epoch": 1.05, "grad_norm": 16.204805374145508, "learning_rate": 2.4748110831234257e-05, "loss": 0.4921, "step": 417 }, { "epoch": 1.05, "grad_norm": 3.415022611618042, "learning_rate": 2.473551637279597e-05, "loss": 0.2516, "step": 418 }, { "epoch": 1.06, "grad_norm": 12.413846969604492, "learning_rate": 2.4722921914357685e-05, "loss": 0.2815, "step": 419 }, { "epoch": 1.06, "grad_norm": 7.598104476928711, "learning_rate": 2.4710327455919397e-05, "loss": 0.1306, "step": 420 }, { "epoch": 1.06, "grad_norm": 5.7529120445251465, "learning_rate": 2.469773299748111e-05, "loss": 0.1619, "step": 421 }, { "epoch": 1.06, "grad_norm": 4.8923516273498535, "learning_rate": 2.4685138539042822e-05, "loss": 0.1385, "step": 422 }, { "epoch": 1.07, "grad_norm": 3.9952762126922607, "learning_rate": 2.4672544080604535e-05, "loss": 0.0965, "step": 423 }, { "epoch": 1.07, "grad_norm": 14.615901947021484, "learning_rate": 2.4659949622166247e-05, "loss": 0.3344, "step": 424 }, { "epoch": 1.07, "grad_norm": 11.933664321899414, "learning_rate": 2.4647355163727963e-05, "loss": 0.4366, "step": 425 }, { "epoch": 1.07, "grad_norm": 6.072282314300537, "learning_rate": 2.4634760705289672e-05, "loss": 0.1211, "step": 426 }, { "epoch": 1.08, "grad_norm": 12.289721488952637, "learning_rate": 2.4622166246851384e-05, "loss": 0.4488, "step": 427 }, { "epoch": 1.08, "grad_norm": 24.086645126342773, "learning_rate": 2.46095717884131e-05, "loss": 0.977, "step": 428 }, { "epoch": 1.08, "grad_norm": 13.474416732788086, "learning_rate": 2.4596977329974812e-05, "loss": 0.6102, "step": 429 }, { "epoch": 1.08, "grad_norm": 5.976496696472168, "learning_rate": 2.4584382871536525e-05, "loss": 0.2712, "step": 430 }, { "epoch": 1.09, "grad_norm": 21.28632926940918, "learning_rate": 2.4571788413098237e-05, "loss": 0.4736, "step": 431 }, { "epoch": 1.09, "grad_norm": 10.714360237121582, "learning_rate": 2.455919395465995e-05, "loss": 0.7356, "step": 432 }, { "epoch": 1.09, "grad_norm": 15.236488342285156, "learning_rate": 2.4546599496221662e-05, "loss": 0.7644, "step": 433 }, { "epoch": 1.09, "grad_norm": 11.787192344665527, "learning_rate": 2.4534005037783378e-05, "loss": 0.7589, "step": 434 }, { "epoch": 1.1, "grad_norm": 12.629518508911133, "learning_rate": 2.4521410579345087e-05, "loss": 0.5611, "step": 435 }, { "epoch": 1.1, "grad_norm": 10.646512985229492, "learning_rate": 2.45088161209068e-05, "loss": 0.5095, "step": 436 }, { "epoch": 1.1, "grad_norm": 10.494978904724121, "learning_rate": 2.4496221662468515e-05, "loss": 0.2447, "step": 437 }, { "epoch": 1.1, "grad_norm": 4.493246078491211, "learning_rate": 2.4483627204030228e-05, "loss": 0.0987, "step": 438 }, { "epoch": 1.11, "grad_norm": 15.06545639038086, "learning_rate": 2.447103274559194e-05, "loss": 0.4021, "step": 439 }, { "epoch": 1.11, "grad_norm": 1.9468973875045776, "learning_rate": 2.4458438287153653e-05, "loss": 0.0678, "step": 440 }, { "epoch": 1.11, "grad_norm": 6.745187759399414, "learning_rate": 2.4445843828715365e-05, "loss": 0.2559, "step": 441 }, { "epoch": 1.11, "grad_norm": 9.412677764892578, "learning_rate": 2.4433249370277077e-05, "loss": 0.5198, "step": 442 }, { "epoch": 1.12, "grad_norm": 7.285678863525391, "learning_rate": 2.4420654911838793e-05, "loss": 0.5288, "step": 443 }, { "epoch": 1.12, "grad_norm": 11.449817657470703, "learning_rate": 2.4408060453400506e-05, "loss": 0.5802, "step": 444 }, { "epoch": 1.12, "grad_norm": 12.689784049987793, "learning_rate": 2.4395465994962215e-05, "loss": 0.2298, "step": 445 }, { "epoch": 1.12, "grad_norm": 25.520742416381836, "learning_rate": 2.438287153652393e-05, "loss": 0.6149, "step": 446 }, { "epoch": 1.13, "grad_norm": 7.158485412597656, "learning_rate": 2.4370277078085643e-05, "loss": 0.3499, "step": 447 }, { "epoch": 1.13, "grad_norm": 5.239826679229736, "learning_rate": 2.4357682619647355e-05, "loss": 0.1604, "step": 448 }, { "epoch": 1.13, "grad_norm": 9.69916820526123, "learning_rate": 2.4345088161209068e-05, "loss": 0.5048, "step": 449 }, { "epoch": 1.13, "grad_norm": 7.610560417175293, "learning_rate": 2.433249370277078e-05, "loss": 0.345, "step": 450 }, { "epoch": 1.14, "grad_norm": 9.45479965209961, "learning_rate": 2.4319899244332493e-05, "loss": 0.4998, "step": 451 }, { "epoch": 1.14, "grad_norm": 6.088526725769043, "learning_rate": 2.430730478589421e-05, "loss": 0.3311, "step": 452 }, { "epoch": 1.14, "grad_norm": 8.135771751403809, "learning_rate": 2.429471032745592e-05, "loss": 0.5105, "step": 453 }, { "epoch": 1.14, "grad_norm": 4.41991662979126, "learning_rate": 2.428211586901763e-05, "loss": 0.3346, "step": 454 }, { "epoch": 1.15, "grad_norm": 6.451519966125488, "learning_rate": 2.4269521410579346e-05, "loss": 0.2548, "step": 455 }, { "epoch": 1.15, "grad_norm": 7.247366428375244, "learning_rate": 2.4256926952141058e-05, "loss": 0.5191, "step": 456 }, { "epoch": 1.15, "grad_norm": 11.858521461486816, "learning_rate": 2.424433249370277e-05, "loss": 0.4406, "step": 457 }, { "epoch": 1.15, "grad_norm": 9.090578079223633, "learning_rate": 2.4231738035264486e-05, "loss": 0.6375, "step": 458 }, { "epoch": 1.16, "grad_norm": 9.295748710632324, "learning_rate": 2.4219143576826196e-05, "loss": 0.5871, "step": 459 }, { "epoch": 1.16, "grad_norm": 5.650725364685059, "learning_rate": 2.4206549118387908e-05, "loss": 0.1907, "step": 460 }, { "epoch": 1.16, "grad_norm": 14.45252513885498, "learning_rate": 2.4193954659949624e-05, "loss": 0.2786, "step": 461 }, { "epoch": 1.16, "grad_norm": 6.475660800933838, "learning_rate": 2.4181360201511336e-05, "loss": 0.3618, "step": 462 }, { "epoch": 1.17, "grad_norm": 8.952823638916016, "learning_rate": 2.4168765743073045e-05, "loss": 0.3385, "step": 463 }, { "epoch": 1.17, "grad_norm": 4.224743843078613, "learning_rate": 2.415617128463476e-05, "loss": 0.3246, "step": 464 }, { "epoch": 1.17, "grad_norm": 7.692142963409424, "learning_rate": 2.4143576826196473e-05, "loss": 0.4094, "step": 465 }, { "epoch": 1.17, "grad_norm": 5.044216156005859, "learning_rate": 2.4130982367758186e-05, "loss": 0.2324, "step": 466 }, { "epoch": 1.18, "grad_norm": 5.303596496582031, "learning_rate": 2.4118387909319902e-05, "loss": 0.2725, "step": 467 }, { "epoch": 1.18, "grad_norm": 8.907066345214844, "learning_rate": 2.410579345088161e-05, "loss": 0.371, "step": 468 }, { "epoch": 1.18, "grad_norm": 5.100632667541504, "learning_rate": 2.4093198992443327e-05, "loss": 0.3409, "step": 469 }, { "epoch": 1.18, "grad_norm": 8.859848022460938, "learning_rate": 2.408060453400504e-05, "loss": 0.3745, "step": 470 }, { "epoch": 1.19, "grad_norm": 7.951125144958496, "learning_rate": 2.406801007556675e-05, "loss": 0.5921, "step": 471 }, { "epoch": 1.19, "grad_norm": 11.735611915588379, "learning_rate": 2.4055415617128467e-05, "loss": 0.359, "step": 472 }, { "epoch": 1.19, "grad_norm": 7.200911998748779, "learning_rate": 2.4042821158690176e-05, "loss": 0.2809, "step": 473 }, { "epoch": 1.19, "grad_norm": 11.294443130493164, "learning_rate": 2.403022670025189e-05, "loss": 0.5223, "step": 474 }, { "epoch": 1.2, "grad_norm": 7.186520099639893, "learning_rate": 2.4017632241813605e-05, "loss": 0.2466, "step": 475 }, { "epoch": 1.2, "grad_norm": 13.5769624710083, "learning_rate": 2.4005037783375317e-05, "loss": 0.5432, "step": 476 }, { "epoch": 1.2, "grad_norm": 6.9906768798828125, "learning_rate": 2.3992443324937026e-05, "loss": 0.1902, "step": 477 }, { "epoch": 1.2, "grad_norm": 5.3905110359191895, "learning_rate": 2.3979848866498742e-05, "loss": 0.2713, "step": 478 }, { "epoch": 1.21, "grad_norm": 4.828819274902344, "learning_rate": 2.3967254408060454e-05, "loss": 0.3166, "step": 479 }, { "epoch": 1.21, "grad_norm": 9.989480018615723, "learning_rate": 2.3954659949622167e-05, "loss": 0.4494, "step": 480 }, { "epoch": 1.21, "grad_norm": 9.79456615447998, "learning_rate": 2.3942065491183883e-05, "loss": 0.4268, "step": 481 }, { "epoch": 1.21, "grad_norm": 7.9519267082214355, "learning_rate": 2.392947103274559e-05, "loss": 0.3281, "step": 482 }, { "epoch": 1.22, "grad_norm": 6.644056797027588, "learning_rate": 2.3916876574307304e-05, "loss": 0.1665, "step": 483 }, { "epoch": 1.22, "grad_norm": 13.469619750976562, "learning_rate": 2.390428211586902e-05, "loss": 0.3806, "step": 484 }, { "epoch": 1.22, "grad_norm": 8.886473655700684, "learning_rate": 2.3891687657430732e-05, "loss": 0.2951, "step": 485 }, { "epoch": 1.22, "grad_norm": 16.603439331054688, "learning_rate": 2.3879093198992445e-05, "loss": 0.3215, "step": 486 }, { "epoch": 1.23, "grad_norm": 5.1792192459106445, "learning_rate": 2.3866498740554157e-05, "loss": 0.2263, "step": 487 }, { "epoch": 1.23, "grad_norm": 22.057132720947266, "learning_rate": 2.385390428211587e-05, "loss": 0.5097, "step": 488 }, { "epoch": 1.23, "grad_norm": 9.38779067993164, "learning_rate": 2.3841309823677582e-05, "loss": 0.4057, "step": 489 }, { "epoch": 1.23, "grad_norm": 4.386843681335449, "learning_rate": 2.3828715365239298e-05, "loss": 0.0996, "step": 490 }, { "epoch": 1.24, "grad_norm": 21.32297706604004, "learning_rate": 2.3816120906801007e-05, "loss": 1.2138, "step": 491 }, { "epoch": 1.24, "grad_norm": 5.6553874015808105, "learning_rate": 2.380352644836272e-05, "loss": 0.1967, "step": 492 }, { "epoch": 1.24, "grad_norm": 6.671334266662598, "learning_rate": 2.3790931989924435e-05, "loss": 0.1301, "step": 493 }, { "epoch": 1.24, "grad_norm": 10.686269760131836, "learning_rate": 2.3778337531486147e-05, "loss": 0.2715, "step": 494 }, { "epoch": 1.25, "grad_norm": 6.189944744110107, "learning_rate": 2.376574307304786e-05, "loss": 0.4195, "step": 495 }, { "epoch": 1.25, "grad_norm": 11.906808853149414, "learning_rate": 2.3753148614609572e-05, "loss": 0.3496, "step": 496 }, { "epoch": 1.25, "grad_norm": 6.705618381500244, "learning_rate": 2.3740554156171285e-05, "loss": 0.2344, "step": 497 }, { "epoch": 1.25, "grad_norm": 8.328521728515625, "learning_rate": 2.3727959697732997e-05, "loss": 0.1997, "step": 498 }, { "epoch": 1.26, "grad_norm": 11.336451530456543, "learning_rate": 2.3715365239294713e-05, "loss": 0.5954, "step": 499 }, { "epoch": 1.26, "grad_norm": 11.99789810180664, "learning_rate": 2.3702770780856425e-05, "loss": 0.355, "step": 500 }, { "epoch": 1.26, "grad_norm": 17.559165954589844, "learning_rate": 2.3690176322418134e-05, "loss": 0.3469, "step": 501 }, { "epoch": 1.26, "grad_norm": 7.059667110443115, "learning_rate": 2.367758186397985e-05, "loss": 0.1771, "step": 502 }, { "epoch": 1.27, "grad_norm": 5.824542045593262, "learning_rate": 2.3664987405541563e-05, "loss": 0.1534, "step": 503 }, { "epoch": 1.27, "grad_norm": 8.428627967834473, "learning_rate": 2.3652392947103275e-05, "loss": 0.2015, "step": 504 }, { "epoch": 1.27, "grad_norm": 14.697854042053223, "learning_rate": 2.3639798488664988e-05, "loss": 0.6048, "step": 505 }, { "epoch": 1.27, "grad_norm": 10.655426025390625, "learning_rate": 2.36272040302267e-05, "loss": 0.6308, "step": 506 }, { "epoch": 1.28, "grad_norm": 6.204010009765625, "learning_rate": 2.3614609571788412e-05, "loss": 0.145, "step": 507 }, { "epoch": 1.28, "grad_norm": 5.01179838180542, "learning_rate": 2.3602015113350128e-05, "loss": 0.1016, "step": 508 }, { "epoch": 1.28, "grad_norm": 5.69111442565918, "learning_rate": 2.358942065491184e-05, "loss": 0.1569, "step": 509 }, { "epoch": 1.28, "grad_norm": 8.344858169555664, "learning_rate": 2.357682619647355e-05, "loss": 0.1635, "step": 510 }, { "epoch": 1.29, "grad_norm": 2.3715789318084717, "learning_rate": 2.3564231738035266e-05, "loss": 0.06, "step": 511 }, { "epoch": 1.29, "grad_norm": 4.8666605949401855, "learning_rate": 2.3551637279596978e-05, "loss": 0.1141, "step": 512 }, { "epoch": 1.29, "grad_norm": 9.464341163635254, "learning_rate": 2.353904282115869e-05, "loss": 0.2374, "step": 513 }, { "epoch": 1.29, "grad_norm": 5.035337924957275, "learning_rate": 2.3526448362720406e-05, "loss": 0.3652, "step": 514 }, { "epoch": 1.3, "grad_norm": 2.6235170364379883, "learning_rate": 2.3513853904282115e-05, "loss": 0.0372, "step": 515 }, { "epoch": 1.3, "grad_norm": 14.479222297668457, "learning_rate": 2.3501259445843828e-05, "loss": 0.6558, "step": 516 }, { "epoch": 1.3, "grad_norm": 9.507341384887695, "learning_rate": 2.3488664987405544e-05, "loss": 0.293, "step": 517 }, { "epoch": 1.3, "grad_norm": 12.590784072875977, "learning_rate": 2.3476070528967256e-05, "loss": 0.1925, "step": 518 }, { "epoch": 1.31, "grad_norm": 8.395529747009277, "learning_rate": 2.3463476070528965e-05, "loss": 0.3387, "step": 519 }, { "epoch": 1.31, "grad_norm": 9.460827827453613, "learning_rate": 2.345088161209068e-05, "loss": 0.2957, "step": 520 }, { "epoch": 1.31, "grad_norm": 12.13961410522461, "learning_rate": 2.3438287153652393e-05, "loss": 0.3519, "step": 521 }, { "epoch": 1.31, "grad_norm": 3.21748685836792, "learning_rate": 2.3425692695214106e-05, "loss": 0.0368, "step": 522 }, { "epoch": 1.32, "grad_norm": 13.214022636413574, "learning_rate": 2.341309823677582e-05, "loss": 0.4089, "step": 523 }, { "epoch": 1.32, "grad_norm": 15.640990257263184, "learning_rate": 2.340050377833753e-05, "loss": 0.3837, "step": 524 }, { "epoch": 1.32, "grad_norm": 7.136631011962891, "learning_rate": 2.3387909319899243e-05, "loss": 0.3011, "step": 525 }, { "epoch": 1.32, "grad_norm": 8.388585090637207, "learning_rate": 2.337531486146096e-05, "loss": 0.2053, "step": 526 }, { "epoch": 1.33, "grad_norm": 12.422948837280273, "learning_rate": 2.336272040302267e-05, "loss": 0.7142, "step": 527 }, { "epoch": 1.33, "grad_norm": 6.136419773101807, "learning_rate": 2.3350125944584384e-05, "loss": 0.5098, "step": 528 }, { "epoch": 1.33, "grad_norm": 14.847418785095215, "learning_rate": 2.3337531486146096e-05, "loss": 0.2521, "step": 529 }, { "epoch": 1.34, "grad_norm": 9.92980670928955, "learning_rate": 2.332493702770781e-05, "loss": 0.3919, "step": 530 }, { "epoch": 1.34, "grad_norm": 25.052722930908203, "learning_rate": 2.331234256926952e-05, "loss": 0.5285, "step": 531 }, { "epoch": 1.34, "grad_norm": 11.806641578674316, "learning_rate": 2.3299748110831237e-05, "loss": 0.514, "step": 532 }, { "epoch": 1.34, "grad_norm": 21.424930572509766, "learning_rate": 2.3287153652392946e-05, "loss": 0.2707, "step": 533 }, { "epoch": 1.35, "grad_norm": 19.599058151245117, "learning_rate": 2.3274559193954658e-05, "loss": 0.3077, "step": 534 }, { "epoch": 1.35, "grad_norm": 20.604076385498047, "learning_rate": 2.3261964735516374e-05, "loss": 1.0837, "step": 535 }, { "epoch": 1.35, "grad_norm": 2.573046922683716, "learning_rate": 2.3249370277078086e-05, "loss": 0.0493, "step": 536 }, { "epoch": 1.35, "grad_norm": 6.39470100402832, "learning_rate": 2.32367758186398e-05, "loss": 0.1642, "step": 537 }, { "epoch": 1.36, "grad_norm": 8.209395408630371, "learning_rate": 2.322418136020151e-05, "loss": 0.1703, "step": 538 }, { "epoch": 1.36, "grad_norm": 12.432522773742676, "learning_rate": 2.3211586901763224e-05, "loss": 0.3736, "step": 539 }, { "epoch": 1.36, "grad_norm": 7.828361511230469, "learning_rate": 2.3198992443324936e-05, "loss": 0.1279, "step": 540 }, { "epoch": 1.36, "grad_norm": 14.346171379089355, "learning_rate": 2.3186397984886652e-05, "loss": 0.2658, "step": 541 }, { "epoch": 1.37, "grad_norm": 15.226658821105957, "learning_rate": 2.317380352644836e-05, "loss": 0.5368, "step": 542 }, { "epoch": 1.37, "grad_norm": 12.5059175491333, "learning_rate": 2.3161209068010077e-05, "loss": 0.3146, "step": 543 }, { "epoch": 1.37, "grad_norm": 8.119623184204102, "learning_rate": 2.314861460957179e-05, "loss": 0.1653, "step": 544 }, { "epoch": 1.37, "grad_norm": 10.031901359558105, "learning_rate": 2.3136020151133502e-05, "loss": 0.3746, "step": 545 }, { "epoch": 1.38, "grad_norm": 17.581632614135742, "learning_rate": 2.3123425692695218e-05, "loss": 0.6627, "step": 546 }, { "epoch": 1.38, "grad_norm": 10.162378311157227, "learning_rate": 2.3110831234256927e-05, "loss": 0.5804, "step": 547 }, { "epoch": 1.38, "grad_norm": 3.8320300579071045, "learning_rate": 2.309823677581864e-05, "loss": 0.0739, "step": 548 }, { "epoch": 1.38, "grad_norm": 5.822529315948486, "learning_rate": 2.3085642317380355e-05, "loss": 0.075, "step": 549 }, { "epoch": 1.39, "grad_norm": 4.666683673858643, "learning_rate": 2.3073047858942067e-05, "loss": 0.2555, "step": 550 }, { "epoch": 1.39, "grad_norm": 6.855231285095215, "learning_rate": 2.306045340050378e-05, "loss": 0.316, "step": 551 }, { "epoch": 1.39, "grad_norm": 14.629341125488281, "learning_rate": 2.3047858942065492e-05, "loss": 0.6001, "step": 552 }, { "epoch": 1.39, "grad_norm": 6.551723480224609, "learning_rate": 2.3035264483627205e-05, "loss": 0.2065, "step": 553 }, { "epoch": 1.4, "grad_norm": 10.324251174926758, "learning_rate": 2.3022670025188917e-05, "loss": 0.253, "step": 554 }, { "epoch": 1.4, "grad_norm": 7.71247673034668, "learning_rate": 2.3010075566750633e-05, "loss": 0.1637, "step": 555 }, { "epoch": 1.4, "grad_norm": 9.804146766662598, "learning_rate": 2.2997481108312342e-05, "loss": 0.323, "step": 556 }, { "epoch": 1.4, "grad_norm": 14.511861801147461, "learning_rate": 2.2984886649874054e-05, "loss": 0.5289, "step": 557 }, { "epoch": 1.41, "grad_norm": 4.078545570373535, "learning_rate": 2.297229219143577e-05, "loss": 0.1315, "step": 558 }, { "epoch": 1.41, "grad_norm": 9.631327629089355, "learning_rate": 2.2959697732997483e-05, "loss": 0.2195, "step": 559 }, { "epoch": 1.41, "grad_norm": 8.578085899353027, "learning_rate": 2.2947103274559195e-05, "loss": 0.4047, "step": 560 }, { "epoch": 1.41, "grad_norm": 13.405282020568848, "learning_rate": 2.2934508816120907e-05, "loss": 0.9347, "step": 561 }, { "epoch": 1.42, "grad_norm": 11.71997356414795, "learning_rate": 2.292191435768262e-05, "loss": 0.4144, "step": 562 }, { "epoch": 1.42, "grad_norm": 5.992588043212891, "learning_rate": 2.2909319899244332e-05, "loss": 0.2267, "step": 563 }, { "epoch": 1.42, "grad_norm": 6.550726890563965, "learning_rate": 2.2896725440806048e-05, "loss": 0.213, "step": 564 }, { "epoch": 1.42, "grad_norm": 6.907693386077881, "learning_rate": 2.288413098236776e-05, "loss": 0.2641, "step": 565 }, { "epoch": 1.43, "grad_norm": 20.506135940551758, "learning_rate": 2.287153652392947e-05, "loss": 0.6865, "step": 566 }, { "epoch": 1.43, "grad_norm": 5.674739360809326, "learning_rate": 2.2858942065491185e-05, "loss": 0.5013, "step": 567 }, { "epoch": 1.43, "grad_norm": 7.887065887451172, "learning_rate": 2.2846347607052898e-05, "loss": 0.3375, "step": 568 }, { "epoch": 1.43, "grad_norm": 7.742120742797852, "learning_rate": 2.283375314861461e-05, "loss": 0.2199, "step": 569 }, { "epoch": 1.44, "grad_norm": 5.057669162750244, "learning_rate": 2.2821158690176323e-05, "loss": 0.2353, "step": 570 }, { "epoch": 1.44, "grad_norm": 11.653826713562012, "learning_rate": 2.2808564231738035e-05, "loss": 0.1651, "step": 571 }, { "epoch": 1.44, "grad_norm": 6.661721229553223, "learning_rate": 2.2795969773299747e-05, "loss": 0.1668, "step": 572 }, { "epoch": 1.44, "grad_norm": 13.69086742401123, "learning_rate": 2.2783375314861463e-05, "loss": 0.3931, "step": 573 }, { "epoch": 1.45, "grad_norm": 4.947517395019531, "learning_rate": 2.2770780856423176e-05, "loss": 0.1813, "step": 574 }, { "epoch": 1.45, "grad_norm": 9.176385879516602, "learning_rate": 2.2758186397984885e-05, "loss": 0.5455, "step": 575 }, { "epoch": 1.45, "grad_norm": 14.556722640991211, "learning_rate": 2.27455919395466e-05, "loss": 0.4582, "step": 576 }, { "epoch": 1.45, "grad_norm": 8.501097679138184, "learning_rate": 2.2732997481108313e-05, "loss": 0.4377, "step": 577 }, { "epoch": 1.46, "grad_norm": 8.8375244140625, "learning_rate": 2.2720403022670025e-05, "loss": 0.4837, "step": 578 }, { "epoch": 1.46, "grad_norm": 7.6127400398254395, "learning_rate": 2.270780856423174e-05, "loss": 0.2413, "step": 579 }, { "epoch": 1.46, "grad_norm": 5.552041530609131, "learning_rate": 2.269521410579345e-05, "loss": 0.3138, "step": 580 }, { "epoch": 1.46, "grad_norm": 11.838778495788574, "learning_rate": 2.2682619647355163e-05, "loss": 0.5564, "step": 581 }, { "epoch": 1.47, "grad_norm": 10.92528247833252, "learning_rate": 2.267002518891688e-05, "loss": 0.4416, "step": 582 }, { "epoch": 1.47, "grad_norm": 15.060375213623047, "learning_rate": 2.265743073047859e-05, "loss": 0.5191, "step": 583 }, { "epoch": 1.47, "grad_norm": 6.89783239364624, "learning_rate": 2.26448362720403e-05, "loss": 0.2395, "step": 584 }, { "epoch": 1.47, "grad_norm": 6.548165798187256, "learning_rate": 2.2632241813602016e-05, "loss": 0.16, "step": 585 }, { "epoch": 1.48, "grad_norm": 7.645367622375488, "learning_rate": 2.2619647355163728e-05, "loss": 0.2274, "step": 586 }, { "epoch": 1.48, "grad_norm": 6.4248247146606445, "learning_rate": 2.260705289672544e-05, "loss": 0.2073, "step": 587 }, { "epoch": 1.48, "grad_norm": 16.396223068237305, "learning_rate": 2.2594458438287157e-05, "loss": 0.4331, "step": 588 }, { "epoch": 1.48, "grad_norm": 15.62296199798584, "learning_rate": 2.2581863979848866e-05, "loss": 0.2843, "step": 589 }, { "epoch": 1.49, "grad_norm": 9.540681838989258, "learning_rate": 2.2569269521410578e-05, "loss": 0.5034, "step": 590 }, { "epoch": 1.49, "grad_norm": 6.93734073638916, "learning_rate": 2.2556675062972294e-05, "loss": 0.1416, "step": 591 }, { "epoch": 1.49, "grad_norm": 14.586864471435547, "learning_rate": 2.2544080604534006e-05, "loss": 0.6036, "step": 592 }, { "epoch": 1.49, "grad_norm": 10.943079948425293, "learning_rate": 2.253148614609572e-05, "loss": 0.3713, "step": 593 }, { "epoch": 1.5, "grad_norm": 10.702349662780762, "learning_rate": 2.251889168765743e-05, "loss": 0.4334, "step": 594 }, { "epoch": 1.5, "grad_norm": 5.976529121398926, "learning_rate": 2.2506297229219144e-05, "loss": 0.2093, "step": 595 }, { "epoch": 1.5, "grad_norm": 8.961341857910156, "learning_rate": 2.2493702770780856e-05, "loss": 0.2841, "step": 596 }, { "epoch": 1.5, "grad_norm": 23.375741958618164, "learning_rate": 2.2481108312342572e-05, "loss": 0.5659, "step": 597 }, { "epoch": 1.51, "grad_norm": 5.276069164276123, "learning_rate": 2.246851385390428e-05, "loss": 0.1247, "step": 598 }, { "epoch": 1.51, "grad_norm": 10.108563423156738, "learning_rate": 2.2455919395465993e-05, "loss": 0.2897, "step": 599 }, { "epoch": 1.51, "grad_norm": 9.869269371032715, "learning_rate": 2.244332493702771e-05, "loss": 0.2426, "step": 600 }, { "epoch": 1.51, "grad_norm": 3.3714253902435303, "learning_rate": 2.243073047858942e-05, "loss": 0.2662, "step": 601 }, { "epoch": 1.52, "grad_norm": 2.3957178592681885, "learning_rate": 2.2418136020151134e-05, "loss": 0.0724, "step": 602 }, { "epoch": 1.52, "grad_norm": 4.764224529266357, "learning_rate": 2.2405541561712846e-05, "loss": 0.0976, "step": 603 }, { "epoch": 1.52, "grad_norm": 12.71890640258789, "learning_rate": 2.239294710327456e-05, "loss": 0.3899, "step": 604 }, { "epoch": 1.52, "grad_norm": 13.903827667236328, "learning_rate": 2.238035264483627e-05, "loss": 0.566, "step": 605 }, { "epoch": 1.53, "grad_norm": 9.071674346923828, "learning_rate": 2.2367758186397987e-05, "loss": 0.5847, "step": 606 }, { "epoch": 1.53, "grad_norm": 21.84575843811035, "learning_rate": 2.23551637279597e-05, "loss": 0.4494, "step": 607 }, { "epoch": 1.53, "grad_norm": 10.90080451965332, "learning_rate": 2.234256926952141e-05, "loss": 0.2771, "step": 608 }, { "epoch": 1.53, "grad_norm": 7.766908168792725, "learning_rate": 2.2329974811083124e-05, "loss": 0.1046, "step": 609 }, { "epoch": 1.54, "grad_norm": 11.402420043945312, "learning_rate": 2.2317380352644837e-05, "loss": 0.4345, "step": 610 }, { "epoch": 1.54, "grad_norm": 6.0967607498168945, "learning_rate": 2.230478589420655e-05, "loss": 0.0931, "step": 611 }, { "epoch": 1.54, "grad_norm": 9.602745056152344, "learning_rate": 2.229219143576826e-05, "loss": 0.4619, "step": 612 }, { "epoch": 1.54, "grad_norm": 7.642356872558594, "learning_rate": 2.2279596977329974e-05, "loss": 0.0986, "step": 613 }, { "epoch": 1.55, "grad_norm": 11.6745023727417, "learning_rate": 2.226700251889169e-05, "loss": 0.2138, "step": 614 }, { "epoch": 1.55, "grad_norm": 4.516753673553467, "learning_rate": 2.2254408060453402e-05, "loss": 0.255, "step": 615 }, { "epoch": 1.55, "grad_norm": 6.529497146606445, "learning_rate": 2.2241813602015115e-05, "loss": 0.0866, "step": 616 }, { "epoch": 1.55, "grad_norm": 14.750448226928711, "learning_rate": 2.2229219143576827e-05, "loss": 0.4606, "step": 617 }, { "epoch": 1.56, "grad_norm": 9.433393478393555, "learning_rate": 2.221662468513854e-05, "loss": 0.2749, "step": 618 }, { "epoch": 1.56, "grad_norm": 13.138323783874512, "learning_rate": 2.2204030226700252e-05, "loss": 0.4306, "step": 619 }, { "epoch": 1.56, "grad_norm": 9.385525703430176, "learning_rate": 2.2191435768261968e-05, "loss": 0.435, "step": 620 }, { "epoch": 1.56, "grad_norm": 9.789994239807129, "learning_rate": 2.217884130982368e-05, "loss": 0.3421, "step": 621 }, { "epoch": 1.57, "grad_norm": 12.882686614990234, "learning_rate": 2.216624685138539e-05, "loss": 0.5238, "step": 622 }, { "epoch": 1.57, "grad_norm": 9.727291107177734, "learning_rate": 2.2153652392947105e-05, "loss": 0.2041, "step": 623 }, { "epoch": 1.57, "grad_norm": 9.766213417053223, "learning_rate": 2.2141057934508818e-05, "loss": 0.3109, "step": 624 }, { "epoch": 1.57, "grad_norm": 10.814486503601074, "learning_rate": 2.212846347607053e-05, "loss": 0.2085, "step": 625 }, { "epoch": 1.58, "grad_norm": 6.508476734161377, "learning_rate": 2.2115869017632242e-05, "loss": 0.3238, "step": 626 }, { "epoch": 1.58, "grad_norm": 13.337801933288574, "learning_rate": 2.2103274559193955e-05, "loss": 0.157, "step": 627 }, { "epoch": 1.58, "grad_norm": 5.9723405838012695, "learning_rate": 2.2090680100755667e-05, "loss": 0.4441, "step": 628 }, { "epoch": 1.58, "grad_norm": 6.157145023345947, "learning_rate": 2.2078085642317383e-05, "loss": 0.0856, "step": 629 }, { "epoch": 1.59, "grad_norm": 6.4899373054504395, "learning_rate": 2.2065491183879095e-05, "loss": 0.3309, "step": 630 }, { "epoch": 1.59, "grad_norm": 3.3382256031036377, "learning_rate": 2.2052896725440805e-05, "loss": 0.0748, "step": 631 }, { "epoch": 1.59, "grad_norm": 7.103469371795654, "learning_rate": 2.204030226700252e-05, "loss": 0.2311, "step": 632 }, { "epoch": 1.59, "grad_norm": 6.943920135498047, "learning_rate": 2.2027707808564233e-05, "loss": 0.1224, "step": 633 }, { "epoch": 1.6, "grad_norm": 7.204896450042725, "learning_rate": 2.2015113350125945e-05, "loss": 0.3382, "step": 634 }, { "epoch": 1.6, "grad_norm": 10.163846015930176, "learning_rate": 2.200251889168766e-05, "loss": 0.4018, "step": 635 }, { "epoch": 1.6, "grad_norm": 14.589173316955566, "learning_rate": 2.198992443324937e-05, "loss": 0.2431, "step": 636 }, { "epoch": 1.6, "grad_norm": 8.574972152709961, "learning_rate": 2.1977329974811082e-05, "loss": 0.15, "step": 637 }, { "epoch": 1.61, "grad_norm": 5.912990093231201, "learning_rate": 2.1964735516372798e-05, "loss": 0.2588, "step": 638 }, { "epoch": 1.61, "grad_norm": 9.575433731079102, "learning_rate": 2.195214105793451e-05, "loss": 0.3173, "step": 639 }, { "epoch": 1.61, "grad_norm": 6.566403865814209, "learning_rate": 2.193954659949622e-05, "loss": 0.2146, "step": 640 }, { "epoch": 1.61, "grad_norm": 15.933618545532227, "learning_rate": 2.1926952141057936e-05, "loss": 0.6077, "step": 641 }, { "epoch": 1.62, "grad_norm": 6.235144138336182, "learning_rate": 2.1914357682619648e-05, "loss": 0.1017, "step": 642 }, { "epoch": 1.62, "grad_norm": 14.474084854125977, "learning_rate": 2.190176322418136e-05, "loss": 0.4763, "step": 643 }, { "epoch": 1.62, "grad_norm": 10.936660766601562, "learning_rate": 2.1889168765743076e-05, "loss": 0.3788, "step": 644 }, { "epoch": 1.62, "grad_norm": 15.92046070098877, "learning_rate": 2.1876574307304785e-05, "loss": 0.5219, "step": 645 }, { "epoch": 1.63, "grad_norm": 10.802763938903809, "learning_rate": 2.1863979848866498e-05, "loss": 0.5354, "step": 646 }, { "epoch": 1.63, "grad_norm": 12.62859058380127, "learning_rate": 2.1851385390428214e-05, "loss": 0.2161, "step": 647 }, { "epoch": 1.63, "grad_norm": 11.541522026062012, "learning_rate": 2.1838790931989926e-05, "loss": 0.2553, "step": 648 }, { "epoch": 1.63, "grad_norm": 8.991678237915039, "learning_rate": 2.182619647355164e-05, "loss": 0.1912, "step": 649 }, { "epoch": 1.64, "grad_norm": 11.52978515625, "learning_rate": 2.181360201511335e-05, "loss": 0.8781, "step": 650 }, { "epoch": 1.64, "grad_norm": 15.376463890075684, "learning_rate": 2.1801007556675063e-05, "loss": 0.5357, "step": 651 }, { "epoch": 1.64, "grad_norm": 9.569962501525879, "learning_rate": 2.1788413098236776e-05, "loss": 0.2626, "step": 652 }, { "epoch": 1.64, "grad_norm": 11.40342903137207, "learning_rate": 2.177581863979849e-05, "loss": 0.5519, "step": 653 }, { "epoch": 1.65, "grad_norm": 15.95056438446045, "learning_rate": 2.17632241813602e-05, "loss": 0.6138, "step": 654 }, { "epoch": 1.65, "grad_norm": 8.395648002624512, "learning_rate": 2.1750629722921913e-05, "loss": 0.1949, "step": 655 }, { "epoch": 1.65, "grad_norm": 12.559154510498047, "learning_rate": 2.173803526448363e-05, "loss": 0.443, "step": 656 }, { "epoch": 1.65, "grad_norm": 9.407703399658203, "learning_rate": 2.172544080604534e-05, "loss": 0.3019, "step": 657 }, { "epoch": 1.66, "grad_norm": 6.948453426361084, "learning_rate": 2.1712846347607054e-05, "loss": 0.2083, "step": 658 }, { "epoch": 1.66, "grad_norm": 11.148189544677734, "learning_rate": 2.1700251889168766e-05, "loss": 0.3048, "step": 659 }, { "epoch": 1.66, "grad_norm": 7.73734712600708, "learning_rate": 2.168765743073048e-05, "loss": 0.2406, "step": 660 }, { "epoch": 1.66, "grad_norm": 8.320653915405273, "learning_rate": 2.167506297229219e-05, "loss": 0.2978, "step": 661 }, { "epoch": 1.67, "grad_norm": 6.541665077209473, "learning_rate": 2.1662468513853907e-05, "loss": 0.2485, "step": 662 }, { "epoch": 1.67, "grad_norm": 7.9342732429504395, "learning_rate": 2.1649874055415616e-05, "loss": 0.2664, "step": 663 }, { "epoch": 1.67, "grad_norm": 10.804299354553223, "learning_rate": 2.1637279596977328e-05, "loss": 0.3781, "step": 664 }, { "epoch": 1.68, "grad_norm": 8.100717544555664, "learning_rate": 2.1624685138539044e-05, "loss": 0.3833, "step": 665 }, { "epoch": 1.68, "grad_norm": 6.31338357925415, "learning_rate": 2.1612090680100756e-05, "loss": 0.326, "step": 666 }, { "epoch": 1.68, "grad_norm": 13.130090713500977, "learning_rate": 2.159949622166247e-05, "loss": 0.3128, "step": 667 }, { "epoch": 1.68, "grad_norm": 8.29228401184082, "learning_rate": 2.158690176322418e-05, "loss": 0.1476, "step": 668 }, { "epoch": 1.69, "grad_norm": 17.379844665527344, "learning_rate": 2.1574307304785894e-05, "loss": 0.325, "step": 669 }, { "epoch": 1.69, "grad_norm": 3.1806654930114746, "learning_rate": 2.1561712846347606e-05, "loss": 0.0877, "step": 670 }, { "epoch": 1.69, "grad_norm": 6.332939624786377, "learning_rate": 2.1549118387909322e-05, "loss": 0.3275, "step": 671 }, { "epoch": 1.69, "grad_norm": 8.985432624816895, "learning_rate": 2.1536523929471034e-05, "loss": 0.2804, "step": 672 }, { "epoch": 1.7, "grad_norm": 4.4867730140686035, "learning_rate": 2.1523929471032743e-05, "loss": 0.2085, "step": 673 }, { "epoch": 1.7, "grad_norm": 8.214529991149902, "learning_rate": 2.151133501259446e-05, "loss": 0.2132, "step": 674 }, { "epoch": 1.7, "grad_norm": 13.838793754577637, "learning_rate": 2.1498740554156172e-05, "loss": 0.3482, "step": 675 }, { "epoch": 1.7, "grad_norm": 5.29369592666626, "learning_rate": 2.1486146095717884e-05, "loss": 0.1623, "step": 676 }, { "epoch": 1.71, "grad_norm": 7.333521366119385, "learning_rate": 2.1473551637279597e-05, "loss": 0.1188, "step": 677 }, { "epoch": 1.71, "grad_norm": 14.213542938232422, "learning_rate": 2.146095717884131e-05, "loss": 0.2021, "step": 678 }, { "epoch": 1.71, "grad_norm": 12.950404167175293, "learning_rate": 2.144836272040302e-05, "loss": 0.3876, "step": 679 }, { "epoch": 1.71, "grad_norm": 4.482484817504883, "learning_rate": 2.1435768261964737e-05, "loss": 0.1161, "step": 680 }, { "epoch": 1.72, "grad_norm": 9.34411907196045, "learning_rate": 2.142317380352645e-05, "loss": 0.4212, "step": 681 }, { "epoch": 1.72, "grad_norm": 9.932082176208496, "learning_rate": 2.141057934508816e-05, "loss": 0.3787, "step": 682 }, { "epoch": 1.72, "grad_norm": 11.614188194274902, "learning_rate": 2.1397984886649875e-05, "loss": 0.4952, "step": 683 }, { "epoch": 1.72, "grad_norm": 15.483943939208984, "learning_rate": 2.1385390428211587e-05, "loss": 0.7035, "step": 684 }, { "epoch": 1.73, "grad_norm": 9.446450233459473, "learning_rate": 2.13727959697733e-05, "loss": 0.2137, "step": 685 }, { "epoch": 1.73, "grad_norm": 5.429463863372803, "learning_rate": 2.1360201511335015e-05, "loss": 0.0765, "step": 686 }, { "epoch": 1.73, "grad_norm": 7.334542274475098, "learning_rate": 2.1347607052896724e-05, "loss": 0.6548, "step": 687 }, { "epoch": 1.73, "grad_norm": 4.637013912200928, "learning_rate": 2.133501259445844e-05, "loss": 0.0695, "step": 688 }, { "epoch": 1.74, "grad_norm": 9.739296913146973, "learning_rate": 2.1322418136020153e-05, "loss": 0.8203, "step": 689 }, { "epoch": 1.74, "grad_norm": 12.037601470947266, "learning_rate": 2.1309823677581865e-05, "loss": 0.2714, "step": 690 }, { "epoch": 1.74, "grad_norm": 9.018967628479004, "learning_rate": 2.1297229219143577e-05, "loss": 0.1763, "step": 691 }, { "epoch": 1.74, "grad_norm": 10.764039993286133, "learning_rate": 2.128463476070529e-05, "loss": 0.29, "step": 692 }, { "epoch": 1.75, "grad_norm": 7.258769512176514, "learning_rate": 2.1272040302267002e-05, "loss": 0.2764, "step": 693 }, { "epoch": 1.75, "grad_norm": 5.780182361602783, "learning_rate": 2.1259445843828718e-05, "loss": 0.2163, "step": 694 }, { "epoch": 1.75, "grad_norm": 8.747794151306152, "learning_rate": 2.124685138539043e-05, "loss": 0.2305, "step": 695 }, { "epoch": 1.75, "grad_norm": 6.942091464996338, "learning_rate": 2.123425692695214e-05, "loss": 0.0966, "step": 696 }, { "epoch": 1.76, "grad_norm": 9.203085899353027, "learning_rate": 2.1221662468513855e-05, "loss": 0.1701, "step": 697 }, { "epoch": 1.76, "grad_norm": 12.736608505249023, "learning_rate": 2.1209068010075568e-05, "loss": 0.5168, "step": 698 }, { "epoch": 1.76, "grad_norm": 16.280555725097656, "learning_rate": 2.119647355163728e-05, "loss": 0.2754, "step": 699 }, { "epoch": 1.76, "grad_norm": 19.578857421875, "learning_rate": 2.1183879093198996e-05, "loss": 0.7287, "step": 700 }, { "epoch": 1.77, "grad_norm": 13.128076553344727, "learning_rate": 2.1171284634760705e-05, "loss": 0.5267, "step": 701 }, { "epoch": 1.77, "grad_norm": 9.765761375427246, "learning_rate": 2.1158690176322417e-05, "loss": 0.241, "step": 702 }, { "epoch": 1.77, "grad_norm": 4.671734809875488, "learning_rate": 2.1146095717884133e-05, "loss": 0.097, "step": 703 }, { "epoch": 1.77, "grad_norm": 4.864780426025391, "learning_rate": 2.1133501259445846e-05, "loss": 0.279, "step": 704 }, { "epoch": 1.78, "grad_norm": 9.086429595947266, "learning_rate": 2.1120906801007555e-05, "loss": 0.383, "step": 705 }, { "epoch": 1.78, "grad_norm": 8.034072875976562, "learning_rate": 2.110831234256927e-05, "loss": 0.3027, "step": 706 }, { "epoch": 1.78, "grad_norm": 17.50482177734375, "learning_rate": 2.1095717884130983e-05, "loss": 0.7063, "step": 707 }, { "epoch": 1.78, "grad_norm": 15.067398071289062, "learning_rate": 2.1083123425692695e-05, "loss": 0.2317, "step": 708 }, { "epoch": 1.79, "grad_norm": 12.081408500671387, "learning_rate": 2.107052896725441e-05, "loss": 0.6775, "step": 709 }, { "epoch": 1.79, "grad_norm": 8.175625801086426, "learning_rate": 2.105793450881612e-05, "loss": 0.2379, "step": 710 }, { "epoch": 1.79, "grad_norm": 11.485807418823242, "learning_rate": 2.1045340050377833e-05, "loss": 0.3477, "step": 711 }, { "epoch": 1.79, "grad_norm": 10.902219772338867, "learning_rate": 2.103274559193955e-05, "loss": 0.3242, "step": 712 }, { "epoch": 1.8, "grad_norm": 7.6868205070495605, "learning_rate": 2.102015113350126e-05, "loss": 0.2112, "step": 713 }, { "epoch": 1.8, "grad_norm": 4.256736755371094, "learning_rate": 2.1007556675062973e-05, "loss": 0.1194, "step": 714 }, { "epoch": 1.8, "grad_norm": 4.6595659255981445, "learning_rate": 2.0994962216624686e-05, "loss": 0.1548, "step": 715 }, { "epoch": 1.8, "grad_norm": 12.510821342468262, "learning_rate": 2.0982367758186398e-05, "loss": 0.2877, "step": 716 }, { "epoch": 1.81, "grad_norm": 5.698768138885498, "learning_rate": 2.096977329974811e-05, "loss": 0.3862, "step": 717 }, { "epoch": 1.81, "grad_norm": 8.279985427856445, "learning_rate": 2.0957178841309827e-05, "loss": 0.3283, "step": 718 }, { "epoch": 1.81, "grad_norm": 10.287877082824707, "learning_rate": 2.0944584382871536e-05, "loss": 0.3959, "step": 719 }, { "epoch": 1.81, "grad_norm": 8.201375961303711, "learning_rate": 2.0931989924433248e-05, "loss": 0.1885, "step": 720 }, { "epoch": 1.82, "grad_norm": 9.439096450805664, "learning_rate": 2.0919395465994964e-05, "loss": 0.1917, "step": 721 }, { "epoch": 1.82, "grad_norm": 10.867537498474121, "learning_rate": 2.0906801007556676e-05, "loss": 0.3561, "step": 722 }, { "epoch": 1.82, "grad_norm": 10.648015975952148, "learning_rate": 2.089420654911839e-05, "loss": 0.3826, "step": 723 }, { "epoch": 1.82, "grad_norm": 20.001815795898438, "learning_rate": 2.08816120906801e-05, "loss": 0.4748, "step": 724 }, { "epoch": 1.83, "grad_norm": 7.351145267486572, "learning_rate": 2.0869017632241814e-05, "loss": 0.0979, "step": 725 }, { "epoch": 1.83, "grad_norm": 7.364979267120361, "learning_rate": 2.0856423173803526e-05, "loss": 0.3825, "step": 726 }, { "epoch": 1.83, "grad_norm": 14.710403442382812, "learning_rate": 2.0843828715365242e-05, "loss": 0.472, "step": 727 }, { "epoch": 1.83, "grad_norm": 3.746670961380005, "learning_rate": 2.0831234256926954e-05, "loss": 0.2175, "step": 728 }, { "epoch": 1.84, "grad_norm": 9.130343437194824, "learning_rate": 2.0818639798488663e-05, "loss": 0.4235, "step": 729 }, { "epoch": 1.84, "grad_norm": 13.102235794067383, "learning_rate": 2.080604534005038e-05, "loss": 0.615, "step": 730 }, { "epoch": 1.84, "grad_norm": 6.6438822746276855, "learning_rate": 2.079345088161209e-05, "loss": 0.1462, "step": 731 }, { "epoch": 1.84, "grad_norm": 15.286580085754395, "learning_rate": 2.0780856423173804e-05, "loss": 0.3657, "step": 732 }, { "epoch": 1.85, "grad_norm": 12.244285583496094, "learning_rate": 2.0768261964735516e-05, "loss": 0.3817, "step": 733 }, { "epoch": 1.85, "grad_norm": 6.858506679534912, "learning_rate": 2.075566750629723e-05, "loss": 0.3086, "step": 734 }, { "epoch": 1.85, "grad_norm": 13.334952354431152, "learning_rate": 2.074307304785894e-05, "loss": 0.5208, "step": 735 }, { "epoch": 1.85, "grad_norm": 14.200685501098633, "learning_rate": 2.0730478589420657e-05, "loss": 0.5627, "step": 736 }, { "epoch": 1.86, "grad_norm": 5.623879909515381, "learning_rate": 2.071788413098237e-05, "loss": 0.1789, "step": 737 }, { "epoch": 1.86, "grad_norm": 13.440046310424805, "learning_rate": 2.070528967254408e-05, "loss": 0.3759, "step": 738 }, { "epoch": 1.86, "grad_norm": 8.08634090423584, "learning_rate": 2.0692695214105794e-05, "loss": 0.2264, "step": 739 }, { "epoch": 1.86, "grad_norm": 15.935822486877441, "learning_rate": 2.0680100755667507e-05, "loss": 0.2984, "step": 740 }, { "epoch": 1.87, "grad_norm": 8.345213890075684, "learning_rate": 2.066750629722922e-05, "loss": 0.1735, "step": 741 }, { "epoch": 1.87, "grad_norm": 4.334136486053467, "learning_rate": 2.0654911838790935e-05, "loss": 0.1256, "step": 742 }, { "epoch": 1.87, "grad_norm": 8.379855155944824, "learning_rate": 2.0642317380352644e-05, "loss": 0.1301, "step": 743 }, { "epoch": 1.87, "grad_norm": 6.75809383392334, "learning_rate": 2.0629722921914356e-05, "loss": 0.2219, "step": 744 }, { "epoch": 1.88, "grad_norm": 15.786321640014648, "learning_rate": 2.0617128463476072e-05, "loss": 0.2864, "step": 745 }, { "epoch": 1.88, "grad_norm": 10.396544456481934, "learning_rate": 2.0604534005037785e-05, "loss": 0.7485, "step": 746 }, { "epoch": 1.88, "grad_norm": 12.27066707611084, "learning_rate": 2.0591939546599494e-05, "loss": 0.2798, "step": 747 }, { "epoch": 1.88, "grad_norm": 10.269973754882812, "learning_rate": 2.057934508816121e-05, "loss": 0.2683, "step": 748 }, { "epoch": 1.89, "grad_norm": 11.86515998840332, "learning_rate": 2.0566750629722922e-05, "loss": 0.2667, "step": 749 }, { "epoch": 1.89, "grad_norm": 14.70499038696289, "learning_rate": 2.0554156171284634e-05, "loss": 0.6922, "step": 750 }, { "epoch": 1.89, "grad_norm": 8.275458335876465, "learning_rate": 2.054156171284635e-05, "loss": 0.2484, "step": 751 }, { "epoch": 1.89, "grad_norm": 17.040943145751953, "learning_rate": 2.052896725440806e-05, "loss": 0.8717, "step": 752 }, { "epoch": 1.9, "grad_norm": 13.45297622680664, "learning_rate": 2.0516372795969772e-05, "loss": 0.265, "step": 753 }, { "epoch": 1.9, "grad_norm": 17.105192184448242, "learning_rate": 2.0503778337531488e-05, "loss": 0.6025, "step": 754 }, { "epoch": 1.9, "grad_norm": 9.762206077575684, "learning_rate": 2.04911838790932e-05, "loss": 0.1654, "step": 755 }, { "epoch": 1.9, "grad_norm": 9.832480430603027, "learning_rate": 2.0478589420654912e-05, "loss": 0.2347, "step": 756 }, { "epoch": 1.91, "grad_norm": 4.055430889129639, "learning_rate": 2.0465994962216625e-05, "loss": 0.1007, "step": 757 }, { "epoch": 1.91, "grad_norm": 11.626514434814453, "learning_rate": 2.0453400503778337e-05, "loss": 0.2964, "step": 758 }, { "epoch": 1.91, "grad_norm": 4.129045009613037, "learning_rate": 2.0440806045340053e-05, "loss": 0.0867, "step": 759 }, { "epoch": 1.91, "grad_norm": 9.8018798828125, "learning_rate": 2.0428211586901765e-05, "loss": 0.1439, "step": 760 }, { "epoch": 1.92, "grad_norm": 17.625484466552734, "learning_rate": 2.0415617128463475e-05, "loss": 0.6171, "step": 761 }, { "epoch": 1.92, "grad_norm": 13.455811500549316, "learning_rate": 2.040302267002519e-05, "loss": 0.1957, "step": 762 }, { "epoch": 1.92, "grad_norm": 13.38533878326416, "learning_rate": 2.0390428211586903e-05, "loss": 0.4046, "step": 763 }, { "epoch": 1.92, "grad_norm": 7.502546787261963, "learning_rate": 2.0377833753148615e-05, "loss": 0.1991, "step": 764 }, { "epoch": 1.93, "grad_norm": 8.860077857971191, "learning_rate": 2.036523929471033e-05, "loss": 0.3378, "step": 765 }, { "epoch": 1.93, "grad_norm": 19.337848663330078, "learning_rate": 2.035264483627204e-05, "loss": 0.6095, "step": 766 }, { "epoch": 1.93, "grad_norm": 6.279582500457764, "learning_rate": 2.0340050377833752e-05, "loss": 0.3184, "step": 767 }, { "epoch": 1.93, "grad_norm": 4.984986305236816, "learning_rate": 2.032745591939547e-05, "loss": 0.3696, "step": 768 }, { "epoch": 1.94, "grad_norm": 11.388147354125977, "learning_rate": 2.031486146095718e-05, "loss": 0.4244, "step": 769 }, { "epoch": 1.94, "grad_norm": 8.785746574401855, "learning_rate": 2.030226700251889e-05, "loss": 0.1574, "step": 770 }, { "epoch": 1.94, "grad_norm": 17.234216690063477, "learning_rate": 2.0289672544080606e-05, "loss": 0.6394, "step": 771 }, { "epoch": 1.94, "grad_norm": 9.10766887664795, "learning_rate": 2.0277078085642318e-05, "loss": 0.7971, "step": 772 }, { "epoch": 1.95, "grad_norm": 10.063255310058594, "learning_rate": 2.026448362720403e-05, "loss": 0.1751, "step": 773 }, { "epoch": 1.95, "grad_norm": 8.564946174621582, "learning_rate": 2.0251889168765746e-05, "loss": 0.2599, "step": 774 }, { "epoch": 1.95, "grad_norm": 7.084752559661865, "learning_rate": 2.0239294710327455e-05, "loss": 0.2425, "step": 775 }, { "epoch": 1.95, "grad_norm": 12.423927307128906, "learning_rate": 2.0226700251889168e-05, "loss": 0.5981, "step": 776 }, { "epoch": 1.96, "grad_norm": 7.133967876434326, "learning_rate": 2.0214105793450884e-05, "loss": 0.4236, "step": 777 }, { "epoch": 1.96, "grad_norm": 12.152637481689453, "learning_rate": 2.0201511335012596e-05, "loss": 0.9084, "step": 778 }, { "epoch": 1.96, "grad_norm": 8.058826446533203, "learning_rate": 2.018891687657431e-05, "loss": 0.1888, "step": 779 }, { "epoch": 1.96, "grad_norm": 12.304523468017578, "learning_rate": 2.017632241813602e-05, "loss": 0.3763, "step": 780 }, { "epoch": 1.97, "grad_norm": 6.993175029754639, "learning_rate": 2.0163727959697733e-05, "loss": 0.3192, "step": 781 }, { "epoch": 1.97, "grad_norm": 3.840372323989868, "learning_rate": 2.0151133501259446e-05, "loss": 0.116, "step": 782 }, { "epoch": 1.97, "grad_norm": 6.5427446365356445, "learning_rate": 2.013853904282116e-05, "loss": 0.3154, "step": 783 }, { "epoch": 1.97, "grad_norm": 6.756482124328613, "learning_rate": 2.012594458438287e-05, "loss": 0.3972, "step": 784 }, { "epoch": 1.98, "grad_norm": 10.90357780456543, "learning_rate": 2.0113350125944583e-05, "loss": 0.2908, "step": 785 }, { "epoch": 1.98, "grad_norm": 17.345783233642578, "learning_rate": 2.01007556675063e-05, "loss": 0.4357, "step": 786 }, { "epoch": 1.98, "grad_norm": 4.750704288482666, "learning_rate": 2.008816120906801e-05, "loss": 0.212, "step": 787 }, { "epoch": 1.98, "grad_norm": 13.136444091796875, "learning_rate": 2.0075566750629724e-05, "loss": 0.2416, "step": 788 }, { "epoch": 1.99, "grad_norm": 6.8181376457214355, "learning_rate": 2.0062972292191436e-05, "loss": 0.1993, "step": 789 }, { "epoch": 1.99, "grad_norm": 17.32404899597168, "learning_rate": 2.005037783375315e-05, "loss": 0.7192, "step": 790 }, { "epoch": 1.99, "grad_norm": 10.616106986999512, "learning_rate": 2.003778337531486e-05, "loss": 0.3684, "step": 791 }, { "epoch": 1.99, "grad_norm": 6.223758220672607, "learning_rate": 2.0025188916876577e-05, "loss": 0.2985, "step": 792 }, { "epoch": 2.0, "grad_norm": 9.101003646850586, "learning_rate": 2.001259445843829e-05, "loss": 0.3622, "step": 793 }, { "epoch": 2.0, "grad_norm": 11.418810844421387, "learning_rate": 1.9999999999999998e-05, "loss": 0.0876, "step": 794 }, { "epoch": 2.0, "eval_accuracy": 0.804416403785489, "eval_f1": 0.8032698800120437, "eval_loss": 0.4402177035808563, "eval_precision": 0.8026196885367871, "eval_recall": 0.8048576887939576, "eval_runtime": 1213.5699, "eval_samples_per_second": 1.306, "eval_steps_per_second": 0.082, "step": 794 }, { "epoch": 2.0, "grad_norm": 5.976737976074219, "learning_rate": 1.9987405541561714e-05, "loss": 0.1186, "step": 795 }, { "epoch": 2.01, "grad_norm": 9.083830833435059, "learning_rate": 1.9974811083123426e-05, "loss": 0.1377, "step": 796 }, { "epoch": 2.01, "grad_norm": 5.250425338745117, "learning_rate": 1.996221662468514e-05, "loss": 0.1186, "step": 797 }, { "epoch": 2.01, "grad_norm": 5.4060797691345215, "learning_rate": 1.994962216624685e-05, "loss": 0.1543, "step": 798 }, { "epoch": 2.01, "grad_norm": 3.3773229122161865, "learning_rate": 1.9937027707808564e-05, "loss": 0.2486, "step": 799 }, { "epoch": 2.02, "grad_norm": 10.780204772949219, "learning_rate": 1.9924433249370276e-05, "loss": 0.3412, "step": 800 }, { "epoch": 2.02, "grad_norm": 7.569768905639648, "learning_rate": 1.9911838790931992e-05, "loss": 0.2412, "step": 801 }, { "epoch": 2.02, "grad_norm": 3.2330408096313477, "learning_rate": 1.9899244332493704e-05, "loss": 0.0898, "step": 802 }, { "epoch": 2.02, "grad_norm": 5.828372001647949, "learning_rate": 1.9886649874055413e-05, "loss": 0.419, "step": 803 }, { "epoch": 2.03, "grad_norm": 8.416691780090332, "learning_rate": 1.987405541561713e-05, "loss": 0.5116, "step": 804 }, { "epoch": 2.03, "grad_norm": 4.302773952484131, "learning_rate": 1.9861460957178842e-05, "loss": 0.0885, "step": 805 }, { "epoch": 2.03, "grad_norm": 21.357067108154297, "learning_rate": 1.9848866498740554e-05, "loss": 0.256, "step": 806 }, { "epoch": 2.03, "grad_norm": 12.905417442321777, "learning_rate": 1.983627204030227e-05, "loss": 0.409, "step": 807 }, { "epoch": 2.04, "grad_norm": 6.17220401763916, "learning_rate": 1.982367758186398e-05, "loss": 0.1705, "step": 808 }, { "epoch": 2.04, "grad_norm": 7.837858200073242, "learning_rate": 1.981108312342569e-05, "loss": 0.1145, "step": 809 }, { "epoch": 2.04, "grad_norm": 4.5199809074401855, "learning_rate": 1.9798488664987407e-05, "loss": 0.1877, "step": 810 }, { "epoch": 2.04, "grad_norm": 10.4535551071167, "learning_rate": 1.978589420654912e-05, "loss": 0.238, "step": 811 }, { "epoch": 2.05, "grad_norm": 4.682440280914307, "learning_rate": 1.977329974811083e-05, "loss": 0.2157, "step": 812 }, { "epoch": 2.05, "grad_norm": 3.6102426052093506, "learning_rate": 1.9760705289672545e-05, "loss": 0.0548, "step": 813 }, { "epoch": 2.05, "grad_norm": 8.517151832580566, "learning_rate": 1.9748110831234257e-05, "loss": 0.1414, "step": 814 }, { "epoch": 2.05, "grad_norm": 1.7788081169128418, "learning_rate": 1.973551637279597e-05, "loss": 0.0507, "step": 815 }, { "epoch": 2.06, "grad_norm": 11.008373260498047, "learning_rate": 1.9722921914357685e-05, "loss": 0.4553, "step": 816 }, { "epoch": 2.06, "grad_norm": 6.783651828765869, "learning_rate": 1.9710327455919394e-05, "loss": 0.1793, "step": 817 }, { "epoch": 2.06, "grad_norm": 4.473495960235596, "learning_rate": 1.9697732997481107e-05, "loss": 0.1944, "step": 818 }, { "epoch": 2.06, "grad_norm": 7.093316555023193, "learning_rate": 1.9685138539042823e-05, "loss": 0.3245, "step": 819 }, { "epoch": 2.07, "grad_norm": 8.107620239257812, "learning_rate": 1.9672544080604535e-05, "loss": 0.074, "step": 820 }, { "epoch": 2.07, "grad_norm": 7.021946907043457, "learning_rate": 1.9659949622166247e-05, "loss": 0.3628, "step": 821 }, { "epoch": 2.07, "grad_norm": 6.592014312744141, "learning_rate": 1.964735516372796e-05, "loss": 0.3116, "step": 822 }, { "epoch": 2.07, "grad_norm": 9.092830657958984, "learning_rate": 1.9634760705289672e-05, "loss": 0.3528, "step": 823 }, { "epoch": 2.08, "grad_norm": 4.819520950317383, "learning_rate": 1.9622166246851385e-05, "loss": 0.0542, "step": 824 }, { "epoch": 2.08, "grad_norm": 20.10268783569336, "learning_rate": 1.96095717884131e-05, "loss": 0.506, "step": 825 }, { "epoch": 2.08, "grad_norm": 15.67265796661377, "learning_rate": 1.959697732997481e-05, "loss": 0.1978, "step": 826 }, { "epoch": 2.08, "grad_norm": 6.482569217681885, "learning_rate": 1.9584382871536522e-05, "loss": 0.2595, "step": 827 }, { "epoch": 2.09, "grad_norm": 7.855728626251221, "learning_rate": 1.9571788413098238e-05, "loss": 0.0733, "step": 828 }, { "epoch": 2.09, "grad_norm": 14.004561424255371, "learning_rate": 1.955919395465995e-05, "loss": 0.2764, "step": 829 }, { "epoch": 2.09, "grad_norm": 5.431961536407471, "learning_rate": 1.9546599496221663e-05, "loss": 0.0654, "step": 830 }, { "epoch": 2.09, "grad_norm": 13.324678421020508, "learning_rate": 1.9534005037783375e-05, "loss": 0.3888, "step": 831 }, { "epoch": 2.1, "grad_norm": 14.44890022277832, "learning_rate": 1.9521410579345088e-05, "loss": 0.2096, "step": 832 }, { "epoch": 2.1, "grad_norm": 11.987161636352539, "learning_rate": 1.9508816120906803e-05, "loss": 0.119, "step": 833 }, { "epoch": 2.1, "grad_norm": 15.736373901367188, "learning_rate": 1.9496221662468516e-05, "loss": 0.2762, "step": 834 }, { "epoch": 2.1, "grad_norm": 8.252238273620605, "learning_rate": 1.9483627204030228e-05, "loss": 0.0847, "step": 835 }, { "epoch": 2.11, "grad_norm": 4.142892360687256, "learning_rate": 1.947103274559194e-05, "loss": 0.0768, "step": 836 }, { "epoch": 2.11, "grad_norm": 11.67130184173584, "learning_rate": 1.9458438287153653e-05, "loss": 0.2269, "step": 837 }, { "epoch": 2.11, "grad_norm": 8.394164085388184, "learning_rate": 1.9445843828715365e-05, "loss": 0.1033, "step": 838 }, { "epoch": 2.11, "grad_norm": 7.672865390777588, "learning_rate": 1.943324937027708e-05, "loss": 0.1247, "step": 839 }, { "epoch": 2.12, "grad_norm": 4.08453893661499, "learning_rate": 1.942065491183879e-05, "loss": 0.0456, "step": 840 }, { "epoch": 2.12, "grad_norm": 1.0883690118789673, "learning_rate": 1.9408060453400503e-05, "loss": 0.0326, "step": 841 }, { "epoch": 2.12, "grad_norm": 20.027908325195312, "learning_rate": 1.939546599496222e-05, "loss": 0.1281, "step": 842 }, { "epoch": 2.12, "grad_norm": 13.216873168945312, "learning_rate": 1.938287153652393e-05, "loss": 0.5115, "step": 843 }, { "epoch": 2.13, "grad_norm": 12.752791404724121, "learning_rate": 1.9370277078085643e-05, "loss": 0.6858, "step": 844 }, { "epoch": 2.13, "grad_norm": 18.067123413085938, "learning_rate": 1.9357682619647356e-05, "loss": 0.1749, "step": 845 }, { "epoch": 2.13, "grad_norm": 4.583468914031982, "learning_rate": 1.9345088161209068e-05, "loss": 0.0346, "step": 846 }, { "epoch": 2.13, "grad_norm": 12.90953540802002, "learning_rate": 1.933249370277078e-05, "loss": 0.2077, "step": 847 }, { "epoch": 2.14, "grad_norm": 5.2107110023498535, "learning_rate": 1.9319899244332497e-05, "loss": 0.0584, "step": 848 }, { "epoch": 2.14, "grad_norm": 8.783427238464355, "learning_rate": 1.930730478589421e-05, "loss": 0.5019, "step": 849 }, { "epoch": 2.14, "grad_norm": 6.2126288414001465, "learning_rate": 1.9294710327455918e-05, "loss": 0.0539, "step": 850 }, { "epoch": 2.14, "grad_norm": 13.92211627960205, "learning_rate": 1.9282115869017634e-05, "loss": 0.1497, "step": 851 }, { "epoch": 2.15, "grad_norm": 28.381793975830078, "learning_rate": 1.9269521410579346e-05, "loss": 0.5454, "step": 852 }, { "epoch": 2.15, "grad_norm": 16.814462661743164, "learning_rate": 1.925692695214106e-05, "loss": 0.2194, "step": 853 }, { "epoch": 2.15, "grad_norm": 7.227400302886963, "learning_rate": 1.924433249370277e-05, "loss": 0.0549, "step": 854 }, { "epoch": 2.15, "grad_norm": 15.914610862731934, "learning_rate": 1.9231738035264484e-05, "loss": 0.1357, "step": 855 }, { "epoch": 2.16, "grad_norm": 16.585025787353516, "learning_rate": 1.9219143576826196e-05, "loss": 0.378, "step": 856 }, { "epoch": 2.16, "grad_norm": 1.9572542905807495, "learning_rate": 1.9206549118387912e-05, "loss": 0.0286, "step": 857 }, { "epoch": 2.16, "grad_norm": 17.232242584228516, "learning_rate": 1.9193954659949624e-05, "loss": 0.2815, "step": 858 }, { "epoch": 2.16, "grad_norm": 3.981262683868408, "learning_rate": 1.9181360201511333e-05, "loss": 0.0536, "step": 859 }, { "epoch": 2.17, "grad_norm": 26.572742462158203, "learning_rate": 1.916876574307305e-05, "loss": 0.2345, "step": 860 }, { "epoch": 2.17, "grad_norm": 15.456450462341309, "learning_rate": 1.915617128463476e-05, "loss": 0.2391, "step": 861 }, { "epoch": 2.17, "grad_norm": 16.02436065673828, "learning_rate": 1.9143576826196474e-05, "loss": 0.1117, "step": 862 }, { "epoch": 2.17, "grad_norm": 3.077558994293213, "learning_rate": 1.913098236775819e-05, "loss": 0.0319, "step": 863 }, { "epoch": 2.18, "grad_norm": 24.19316864013672, "learning_rate": 1.91183879093199e-05, "loss": 0.1006, "step": 864 }, { "epoch": 2.18, "grad_norm": 14.604350090026855, "learning_rate": 1.910579345088161e-05, "loss": 0.2585, "step": 865 }, { "epoch": 2.18, "grad_norm": 10.65208911895752, "learning_rate": 1.9093198992443327e-05, "loss": 0.0927, "step": 866 }, { "epoch": 2.18, "grad_norm": 10.185251235961914, "learning_rate": 1.908060453400504e-05, "loss": 0.054, "step": 867 }, { "epoch": 2.19, "grad_norm": 4.103867053985596, "learning_rate": 1.906801007556675e-05, "loss": 0.0463, "step": 868 }, { "epoch": 2.19, "grad_norm": 17.59889030456543, "learning_rate": 1.9055415617128464e-05, "loss": 0.0794, "step": 869 }, { "epoch": 2.19, "grad_norm": 7.415209770202637, "learning_rate": 1.9042821158690177e-05, "loss": 0.0596, "step": 870 }, { "epoch": 2.19, "grad_norm": 11.936656951904297, "learning_rate": 1.903022670025189e-05, "loss": 0.1241, "step": 871 }, { "epoch": 2.2, "grad_norm": 12.43713665008545, "learning_rate": 1.9017632241813605e-05, "loss": 0.2765, "step": 872 }, { "epoch": 2.2, "grad_norm": 8.106919288635254, "learning_rate": 1.9005037783375314e-05, "loss": 0.2905, "step": 873 }, { "epoch": 2.2, "grad_norm": 6.52731466293335, "learning_rate": 1.8992443324937026e-05, "loss": 0.0641, "step": 874 }, { "epoch": 2.2, "grad_norm": 2.299429416656494, "learning_rate": 1.8979848866498742e-05, "loss": 0.0356, "step": 875 }, { "epoch": 2.21, "grad_norm": 23.403858184814453, "learning_rate": 1.8967254408060455e-05, "loss": 0.4673, "step": 876 }, { "epoch": 2.21, "grad_norm": 14.969141006469727, "learning_rate": 1.8954659949622164e-05, "loss": 0.1005, "step": 877 }, { "epoch": 2.21, "grad_norm": 7.990150451660156, "learning_rate": 1.894206549118388e-05, "loss": 0.2583, "step": 878 }, { "epoch": 2.21, "grad_norm": 4.63574743270874, "learning_rate": 1.8929471032745592e-05, "loss": 0.2944, "step": 879 }, { "epoch": 2.22, "grad_norm": 15.271684646606445, "learning_rate": 1.8916876574307304e-05, "loss": 0.2335, "step": 880 }, { "epoch": 2.22, "grad_norm": 18.186588287353516, "learning_rate": 1.890428211586902e-05, "loss": 0.1945, "step": 881 }, { "epoch": 2.22, "grad_norm": 12.56149959564209, "learning_rate": 1.889168765743073e-05, "loss": 0.1999, "step": 882 }, { "epoch": 2.22, "grad_norm": 2.1376261711120605, "learning_rate": 1.8879093198992442e-05, "loss": 0.0223, "step": 883 }, { "epoch": 2.23, "grad_norm": 22.14487075805664, "learning_rate": 1.8866498740554158e-05, "loss": 0.1304, "step": 884 }, { "epoch": 2.23, "grad_norm": 8.436043739318848, "learning_rate": 1.885390428211587e-05, "loss": 0.2673, "step": 885 }, { "epoch": 2.23, "grad_norm": 11.742157936096191, "learning_rate": 1.8841309823677582e-05, "loss": 0.1181, "step": 886 }, { "epoch": 2.23, "grad_norm": 38.36399459838867, "learning_rate": 1.8828715365239295e-05, "loss": 0.4267, "step": 887 }, { "epoch": 2.24, "grad_norm": 15.271390914916992, "learning_rate": 1.8816120906801007e-05, "loss": 0.1224, "step": 888 }, { "epoch": 2.24, "grad_norm": 11.08383846282959, "learning_rate": 1.880352644836272e-05, "loss": 0.106, "step": 889 }, { "epoch": 2.24, "grad_norm": 29.323549270629883, "learning_rate": 1.8790931989924436e-05, "loss": 0.155, "step": 890 }, { "epoch": 2.24, "grad_norm": 9.078880310058594, "learning_rate": 1.8778337531486145e-05, "loss": 0.0812, "step": 891 }, { "epoch": 2.25, "grad_norm": 25.136865615844727, "learning_rate": 1.8765743073047857e-05, "loss": 0.2506, "step": 892 }, { "epoch": 2.25, "grad_norm": 16.840970993041992, "learning_rate": 1.8753148614609573e-05, "loss": 0.1291, "step": 893 }, { "epoch": 2.25, "grad_norm": 17.921939849853516, "learning_rate": 1.8740554156171285e-05, "loss": 0.3605, "step": 894 }, { "epoch": 2.25, "grad_norm": 11.916851997375488, "learning_rate": 1.8727959697732998e-05, "loss": 0.1311, "step": 895 }, { "epoch": 2.26, "grad_norm": 9.328707695007324, "learning_rate": 1.871536523929471e-05, "loss": 0.1106, "step": 896 }, { "epoch": 2.26, "grad_norm": 13.462206840515137, "learning_rate": 1.8702770780856423e-05, "loss": 0.6508, "step": 897 }, { "epoch": 2.26, "grad_norm": 5.1701741218566895, "learning_rate": 1.8690176322418135e-05, "loss": 0.0745, "step": 898 }, { "epoch": 2.26, "grad_norm": 27.52861213684082, "learning_rate": 1.867758186397985e-05, "loss": 0.1263, "step": 899 }, { "epoch": 2.27, "grad_norm": 3.011115074157715, "learning_rate": 1.8664987405541563e-05, "loss": 0.0328, "step": 900 }, { "epoch": 2.27, "grad_norm": 20.727210998535156, "learning_rate": 1.8652392947103272e-05, "loss": 0.1041, "step": 901 }, { "epoch": 2.27, "grad_norm": 16.751338958740234, "learning_rate": 1.8639798488664988e-05, "loss": 0.1396, "step": 902 }, { "epoch": 2.27, "grad_norm": 11.478860855102539, "learning_rate": 1.86272040302267e-05, "loss": 0.4094, "step": 903 }, { "epoch": 2.28, "grad_norm": 2.986703395843506, "learning_rate": 1.8614609571788416e-05, "loss": 0.0394, "step": 904 }, { "epoch": 2.28, "grad_norm": 9.955742835998535, "learning_rate": 1.8602015113350125e-05, "loss": 0.2274, "step": 905 }, { "epoch": 2.28, "grad_norm": 3.6407182216644287, "learning_rate": 1.8589420654911838e-05, "loss": 0.2928, "step": 906 }, { "epoch": 2.28, "grad_norm": 21.52337074279785, "learning_rate": 1.8576826196473554e-05, "loss": 0.3569, "step": 907 }, { "epoch": 2.29, "grad_norm": 0.8511970639228821, "learning_rate": 1.8564231738035266e-05, "loss": 0.0187, "step": 908 }, { "epoch": 2.29, "grad_norm": 11.474825859069824, "learning_rate": 1.855163727959698e-05, "loss": 0.1168, "step": 909 }, { "epoch": 2.29, "grad_norm": 5.235669136047363, "learning_rate": 1.853904282115869e-05, "loss": 0.0495, "step": 910 }, { "epoch": 2.29, "grad_norm": 10.870506286621094, "learning_rate": 1.8526448362720403e-05, "loss": 0.2467, "step": 911 }, { "epoch": 2.3, "grad_norm": 31.614450454711914, "learning_rate": 1.8513853904282116e-05, "loss": 0.2447, "step": 912 }, { "epoch": 2.3, "grad_norm": 11.93704605102539, "learning_rate": 1.850125944584383e-05, "loss": 0.3164, "step": 913 }, { "epoch": 2.3, "grad_norm": 21.598426818847656, "learning_rate": 1.8488664987405544e-05, "loss": 0.0576, "step": 914 }, { "epoch": 2.3, "grad_norm": 5.371317386627197, "learning_rate": 1.8476070528967253e-05, "loss": 0.0423, "step": 915 }, { "epoch": 2.31, "grad_norm": 29.41996192932129, "learning_rate": 1.846347607052897e-05, "loss": 0.1859, "step": 916 }, { "epoch": 2.31, "grad_norm": 7.363552570343018, "learning_rate": 1.845088161209068e-05, "loss": 0.557, "step": 917 }, { "epoch": 2.31, "grad_norm": 32.41464614868164, "learning_rate": 1.8438287153652394e-05, "loss": 0.4631, "step": 918 }, { "epoch": 2.31, "grad_norm": 13.250725746154785, "learning_rate": 1.8425692695214106e-05, "loss": 0.1992, "step": 919 }, { "epoch": 2.32, "grad_norm": 34.75584030151367, "learning_rate": 1.841309823677582e-05, "loss": 0.4553, "step": 920 }, { "epoch": 2.32, "grad_norm": 7.2384233474731445, "learning_rate": 1.840050377833753e-05, "loss": 0.2492, "step": 921 }, { "epoch": 2.32, "grad_norm": 4.95755672454834, "learning_rate": 1.8387909319899247e-05, "loss": 0.2653, "step": 922 }, { "epoch": 2.32, "grad_norm": 44.550514221191406, "learning_rate": 1.837531486146096e-05, "loss": 0.5278, "step": 923 }, { "epoch": 2.33, "grad_norm": 13.902865409851074, "learning_rate": 1.8362720403022668e-05, "loss": 0.3654, "step": 924 }, { "epoch": 2.33, "grad_norm": 2.8645730018615723, "learning_rate": 1.8350125944584384e-05, "loss": 0.024, "step": 925 }, { "epoch": 2.33, "grad_norm": 2.747000217437744, "learning_rate": 1.8337531486146097e-05, "loss": 0.0349, "step": 926 }, { "epoch": 2.34, "grad_norm": 8.224370956420898, "learning_rate": 1.832493702770781e-05, "loss": 0.4737, "step": 927 }, { "epoch": 2.34, "grad_norm": 30.833358764648438, "learning_rate": 1.8312342569269525e-05, "loss": 0.094, "step": 928 }, { "epoch": 2.34, "grad_norm": 3.3213651180267334, "learning_rate": 1.8299748110831234e-05, "loss": 0.0378, "step": 929 }, { "epoch": 2.34, "grad_norm": 25.978818893432617, "learning_rate": 1.8287153652392946e-05, "loss": 0.2553, "step": 930 }, { "epoch": 2.35, "grad_norm": 34.23870086669922, "learning_rate": 1.8274559193954662e-05, "loss": 0.406, "step": 931 }, { "epoch": 2.35, "grad_norm": 24.33919906616211, "learning_rate": 1.8261964735516374e-05, "loss": 0.5579, "step": 932 }, { "epoch": 2.35, "grad_norm": 10.28875732421875, "learning_rate": 1.8249370277078084e-05, "loss": 0.2994, "step": 933 }, { "epoch": 2.35, "grad_norm": 18.206113815307617, "learning_rate": 1.82367758186398e-05, "loss": 0.7409, "step": 934 }, { "epoch": 2.36, "grad_norm": 7.450280666351318, "learning_rate": 1.8224181360201512e-05, "loss": 0.5185, "step": 935 }, { "epoch": 2.36, "grad_norm": 26.665252685546875, "learning_rate": 1.8211586901763224e-05, "loss": 0.279, "step": 936 }, { "epoch": 2.36, "grad_norm": 32.06397247314453, "learning_rate": 1.819899244332494e-05, "loss": 0.2519, "step": 937 }, { "epoch": 2.36, "grad_norm": 20.590030670166016, "learning_rate": 1.818639798488665e-05, "loss": 0.1598, "step": 938 }, { "epoch": 2.37, "grad_norm": 13.61314868927002, "learning_rate": 1.817380352644836e-05, "loss": 0.1431, "step": 939 }, { "epoch": 2.37, "grad_norm": 16.68808937072754, "learning_rate": 1.8161209068010077e-05, "loss": 0.0691, "step": 940 }, { "epoch": 2.37, "grad_norm": 21.29247283935547, "learning_rate": 1.814861460957179e-05, "loss": 0.2522, "step": 941 }, { "epoch": 2.37, "grad_norm": 12.311421394348145, "learning_rate": 1.8136020151133502e-05, "loss": 0.223, "step": 942 }, { "epoch": 2.38, "grad_norm": 10.444890975952148, "learning_rate": 1.8123425692695215e-05, "loss": 0.1636, "step": 943 }, { "epoch": 2.38, "grad_norm": 35.49920654296875, "learning_rate": 1.8110831234256927e-05, "loss": 0.8437, "step": 944 }, { "epoch": 2.38, "grad_norm": 7.902152061462402, "learning_rate": 1.809823677581864e-05, "loss": 0.1288, "step": 945 }, { "epoch": 2.38, "grad_norm": 4.148161888122559, "learning_rate": 1.8085642317380355e-05, "loss": 0.046, "step": 946 }, { "epoch": 2.39, "grad_norm": 14.699132919311523, "learning_rate": 1.8073047858942064e-05, "loss": 0.1743, "step": 947 }, { "epoch": 2.39, "grad_norm": 3.610971450805664, "learning_rate": 1.8060453400503777e-05, "loss": 0.0522, "step": 948 }, { "epoch": 2.39, "grad_norm": 4.12473201751709, "learning_rate": 1.8047858942065493e-05, "loss": 0.0616, "step": 949 }, { "epoch": 2.39, "grad_norm": 5.2726616859436035, "learning_rate": 1.8035264483627205e-05, "loss": 0.0438, "step": 950 }, { "epoch": 2.4, "grad_norm": 13.556662559509277, "learning_rate": 1.8022670025188917e-05, "loss": 0.1472, "step": 951 }, { "epoch": 2.4, "grad_norm": 45.24021530151367, "learning_rate": 1.801007556675063e-05, "loss": 0.1432, "step": 952 }, { "epoch": 2.4, "grad_norm": 19.90587043762207, "learning_rate": 1.7997481108312342e-05, "loss": 0.3582, "step": 953 }, { "epoch": 2.4, "grad_norm": 10.678705215454102, "learning_rate": 1.7984886649874055e-05, "loss": 0.0898, "step": 954 }, { "epoch": 2.41, "grad_norm": 12.450896263122559, "learning_rate": 1.797229219143577e-05, "loss": 0.1159, "step": 955 }, { "epoch": 2.41, "grad_norm": 4.022710800170898, "learning_rate": 1.7959697732997483e-05, "loss": 0.054, "step": 956 }, { "epoch": 2.41, "grad_norm": 18.864871978759766, "learning_rate": 1.7947103274559192e-05, "loss": 0.2458, "step": 957 }, { "epoch": 2.41, "grad_norm": 21.411914825439453, "learning_rate": 1.7934508816120908e-05, "loss": 0.1763, "step": 958 }, { "epoch": 2.42, "grad_norm": 14.408418655395508, "learning_rate": 1.792191435768262e-05, "loss": 0.2913, "step": 959 }, { "epoch": 2.42, "grad_norm": 10.474907875061035, "learning_rate": 1.7909319899244333e-05, "loss": 0.0939, "step": 960 }, { "epoch": 2.42, "grad_norm": 32.35608673095703, "learning_rate": 1.7896725440806045e-05, "loss": 0.4675, "step": 961 }, { "epoch": 2.42, "grad_norm": 0.9104124307632446, "learning_rate": 1.7884130982367758e-05, "loss": 0.0231, "step": 962 }, { "epoch": 2.43, "grad_norm": 0.9205286502838135, "learning_rate": 1.787153652392947e-05, "loss": 0.0203, "step": 963 }, { "epoch": 2.43, "grad_norm": 15.181899070739746, "learning_rate": 1.7858942065491186e-05, "loss": 0.1106, "step": 964 }, { "epoch": 2.43, "grad_norm": 4.9965033531188965, "learning_rate": 1.7846347607052898e-05, "loss": 0.0602, "step": 965 }, { "epoch": 2.43, "grad_norm": 4.398914813995361, "learning_rate": 1.7833753148614607e-05, "loss": 0.0526, "step": 966 }, { "epoch": 2.44, "grad_norm": 29.782489776611328, "learning_rate": 1.7821158690176323e-05, "loss": 0.1758, "step": 967 }, { "epoch": 2.44, "grad_norm": 11.924981117248535, "learning_rate": 1.7808564231738035e-05, "loss": 0.3013, "step": 968 }, { "epoch": 2.44, "grad_norm": 15.300978660583496, "learning_rate": 1.7795969773299748e-05, "loss": 0.0943, "step": 969 }, { "epoch": 2.44, "grad_norm": 8.09597396850586, "learning_rate": 1.7783375314861464e-05, "loss": 0.0648, "step": 970 }, { "epoch": 2.45, "grad_norm": 20.45931625366211, "learning_rate": 1.7770780856423173e-05, "loss": 0.4196, "step": 971 }, { "epoch": 2.45, "grad_norm": 5.484227180480957, "learning_rate": 1.7758186397984885e-05, "loss": 0.0419, "step": 972 }, { "epoch": 2.45, "grad_norm": 10.987439155578613, "learning_rate": 1.77455919395466e-05, "loss": 0.0789, "step": 973 }, { "epoch": 2.45, "grad_norm": 20.312944412231445, "learning_rate": 1.7732997481108313e-05, "loss": 0.5257, "step": 974 }, { "epoch": 2.46, "grad_norm": 25.339521408081055, "learning_rate": 1.7720403022670026e-05, "loss": 0.1372, "step": 975 }, { "epoch": 2.46, "grad_norm": 17.586299896240234, "learning_rate": 1.7707808564231738e-05, "loss": 0.1025, "step": 976 }, { "epoch": 2.46, "grad_norm": 18.70049476623535, "learning_rate": 1.769521410579345e-05, "loss": 0.055, "step": 977 }, { "epoch": 2.46, "grad_norm": 9.797844886779785, "learning_rate": 1.7682619647355167e-05, "loss": 0.6287, "step": 978 }, { "epoch": 2.47, "grad_norm": 1.4952588081359863, "learning_rate": 1.767002518891688e-05, "loss": 0.0154, "step": 979 }, { "epoch": 2.47, "grad_norm": 0.32095134258270264, "learning_rate": 1.7657430730478588e-05, "loss": 0.0106, "step": 980 }, { "epoch": 2.47, "grad_norm": 1.3020259141921997, "learning_rate": 1.7644836272040304e-05, "loss": 0.0136, "step": 981 }, { "epoch": 2.47, "grad_norm": 44.08338928222656, "learning_rate": 1.7632241813602016e-05, "loss": 0.8547, "step": 982 }, { "epoch": 2.48, "grad_norm": 21.83006477355957, "learning_rate": 1.761964735516373e-05, "loss": 0.3798, "step": 983 }, { "epoch": 2.48, "grad_norm": 43.243038177490234, "learning_rate": 1.760705289672544e-05, "loss": 0.692, "step": 984 }, { "epoch": 2.48, "grad_norm": 16.481504440307617, "learning_rate": 1.7594458438287154e-05, "loss": 0.548, "step": 985 }, { "epoch": 2.48, "grad_norm": 26.669479370117188, "learning_rate": 1.7581863979848866e-05, "loss": 1.2603, "step": 986 }, { "epoch": 2.49, "grad_norm": 17.41109275817871, "learning_rate": 1.7569269521410582e-05, "loss": 0.0898, "step": 987 }, { "epoch": 2.49, "grad_norm": 11.004029273986816, "learning_rate": 1.7556675062972294e-05, "loss": 0.0521, "step": 988 }, { "epoch": 2.49, "grad_norm": 16.72394561767578, "learning_rate": 1.7544080604534003e-05, "loss": 0.4888, "step": 989 }, { "epoch": 2.49, "grad_norm": 7.188633441925049, "learning_rate": 1.753148614609572e-05, "loss": 0.031, "step": 990 }, { "epoch": 2.5, "grad_norm": 0.7417293787002563, "learning_rate": 1.751889168765743e-05, "loss": 0.014, "step": 991 }, { "epoch": 2.5, "grad_norm": 4.745822906494141, "learning_rate": 1.7506297229219144e-05, "loss": 0.0298, "step": 992 }, { "epoch": 2.5, "grad_norm": 20.97715187072754, "learning_rate": 1.749370277078086e-05, "loss": 0.2558, "step": 993 }, { "epoch": 2.5, "grad_norm": 27.40770149230957, "learning_rate": 1.748110831234257e-05, "loss": 0.1127, "step": 994 }, { "epoch": 2.51, "grad_norm": 19.053752899169922, "learning_rate": 1.746851385390428e-05, "loss": 0.2859, "step": 995 }, { "epoch": 2.51, "grad_norm": 41.415008544921875, "learning_rate": 1.7455919395465997e-05, "loss": 0.3327, "step": 996 }, { "epoch": 2.51, "grad_norm": 10.322839736938477, "learning_rate": 1.744332493702771e-05, "loss": 0.2526, "step": 997 }, { "epoch": 2.51, "grad_norm": 12.210159301757812, "learning_rate": 1.743073047858942e-05, "loss": 0.2203, "step": 998 }, { "epoch": 2.52, "grad_norm": 21.103782653808594, "learning_rate": 1.7418136020151134e-05, "loss": 0.6672, "step": 999 }, { "epoch": 2.52, "grad_norm": 3.4252419471740723, "learning_rate": 1.7405541561712847e-05, "loss": 0.025, "step": 1000 }, { "epoch": 2.52, "grad_norm": 26.5319766998291, "learning_rate": 1.739294710327456e-05, "loss": 0.4177, "step": 1001 }, { "epoch": 2.52, "grad_norm": 11.68613338470459, "learning_rate": 1.7380352644836275e-05, "loss": 0.2262, "step": 1002 }, { "epoch": 2.53, "grad_norm": 14.039557456970215, "learning_rate": 1.7367758186397984e-05, "loss": 0.2382, "step": 1003 }, { "epoch": 2.53, "grad_norm": 20.03414535522461, "learning_rate": 1.7355163727959696e-05, "loss": 0.2475, "step": 1004 }, { "epoch": 2.53, "grad_norm": 29.128925323486328, "learning_rate": 1.7342569269521412e-05, "loss": 0.1902, "step": 1005 }, { "epoch": 2.53, "grad_norm": 7.585634231567383, "learning_rate": 1.7329974811083125e-05, "loss": 0.0359, "step": 1006 }, { "epoch": 2.54, "grad_norm": 0.3999019265174866, "learning_rate": 1.7317380352644837e-05, "loss": 0.0117, "step": 1007 }, { "epoch": 2.54, "grad_norm": 12.283281326293945, "learning_rate": 1.730478589420655e-05, "loss": 0.0946, "step": 1008 }, { "epoch": 2.54, "grad_norm": 9.220109939575195, "learning_rate": 1.7292191435768262e-05, "loss": 0.0397, "step": 1009 }, { "epoch": 2.54, "grad_norm": 3.8018109798431396, "learning_rate": 1.7279596977329974e-05, "loss": 0.0305, "step": 1010 }, { "epoch": 2.55, "grad_norm": 40.12253189086914, "learning_rate": 1.726700251889169e-05, "loss": 0.4056, "step": 1011 }, { "epoch": 2.55, "grad_norm": 13.994961738586426, "learning_rate": 1.72544080604534e-05, "loss": 0.3435, "step": 1012 }, { "epoch": 2.55, "grad_norm": 19.624113082885742, "learning_rate": 1.7241813602015112e-05, "loss": 0.4751, "step": 1013 }, { "epoch": 2.55, "grad_norm": 20.66043472290039, "learning_rate": 1.7229219143576828e-05, "loss": 0.1713, "step": 1014 }, { "epoch": 2.56, "grad_norm": 48.67464065551758, "learning_rate": 1.721662468513854e-05, "loss": 0.3976, "step": 1015 }, { "epoch": 2.56, "grad_norm": 30.28778839111328, "learning_rate": 1.7204030226700252e-05, "loss": 0.3882, "step": 1016 }, { "epoch": 2.56, "grad_norm": 31.674646377563477, "learning_rate": 1.7191435768261965e-05, "loss": 0.6509, "step": 1017 }, { "epoch": 2.56, "grad_norm": 15.582225799560547, "learning_rate": 1.7178841309823677e-05, "loss": 0.5633, "step": 1018 }, { "epoch": 2.57, "grad_norm": 8.46904182434082, "learning_rate": 1.716624685138539e-05, "loss": 0.0472, "step": 1019 }, { "epoch": 2.57, "grad_norm": 6.153964519500732, "learning_rate": 1.7153652392947106e-05, "loss": 0.2258, "step": 1020 }, { "epoch": 2.57, "grad_norm": 18.53000259399414, "learning_rate": 1.7141057934508818e-05, "loss": 0.2525, "step": 1021 }, { "epoch": 2.57, "grad_norm": 7.518229007720947, "learning_rate": 1.7128463476070527e-05, "loss": 0.052, "step": 1022 }, { "epoch": 2.58, "grad_norm": 16.217754364013672, "learning_rate": 1.7115869017632243e-05, "loss": 0.2544, "step": 1023 }, { "epoch": 2.58, "grad_norm": 21.273475646972656, "learning_rate": 1.7103274559193955e-05, "loss": 0.1248, "step": 1024 }, { "epoch": 2.58, "grad_norm": 10.482606887817383, "learning_rate": 1.7090680100755668e-05, "loss": 0.0675, "step": 1025 }, { "epoch": 2.58, "grad_norm": 18.854312896728516, "learning_rate": 1.707808564231738e-05, "loss": 0.1067, "step": 1026 }, { "epoch": 2.59, "grad_norm": 11.522710800170898, "learning_rate": 1.7065491183879093e-05, "loss": 0.2552, "step": 1027 }, { "epoch": 2.59, "grad_norm": 15.483019828796387, "learning_rate": 1.7052896725440805e-05, "loss": 0.3915, "step": 1028 }, { "epoch": 2.59, "grad_norm": 8.645544052124023, "learning_rate": 1.704030226700252e-05, "loss": 0.0815, "step": 1029 }, { "epoch": 2.59, "grad_norm": 18.710952758789062, "learning_rate": 1.7027707808564233e-05, "loss": 0.3915, "step": 1030 }, { "epoch": 2.6, "grad_norm": 14.992968559265137, "learning_rate": 1.7015113350125942e-05, "loss": 0.2023, "step": 1031 }, { "epoch": 2.6, "grad_norm": 14.996759414672852, "learning_rate": 1.7002518891687658e-05, "loss": 0.2511, "step": 1032 }, { "epoch": 2.6, "grad_norm": 6.331859111785889, "learning_rate": 1.698992443324937e-05, "loss": 0.0837, "step": 1033 }, { "epoch": 2.6, "grad_norm": 2.4125723838806152, "learning_rate": 1.6977329974811083e-05, "loss": 0.0321, "step": 1034 }, { "epoch": 2.61, "grad_norm": 14.080228805541992, "learning_rate": 1.69647355163728e-05, "loss": 0.8303, "step": 1035 }, { "epoch": 2.61, "grad_norm": 11.021700859069824, "learning_rate": 1.6952141057934508e-05, "loss": 0.3159, "step": 1036 }, { "epoch": 2.61, "grad_norm": 1.5516949892044067, "learning_rate": 1.693954659949622e-05, "loss": 0.0211, "step": 1037 }, { "epoch": 2.61, "grad_norm": 24.504554748535156, "learning_rate": 1.6926952141057936e-05, "loss": 0.2686, "step": 1038 }, { "epoch": 2.62, "grad_norm": 16.056262969970703, "learning_rate": 1.691435768261965e-05, "loss": 0.1862, "step": 1039 }, { "epoch": 2.62, "grad_norm": 23.649192810058594, "learning_rate": 1.6901763224181357e-05, "loss": 0.7134, "step": 1040 }, { "epoch": 2.62, "grad_norm": 4.870745658874512, "learning_rate": 1.6889168765743073e-05, "loss": 0.0574, "step": 1041 }, { "epoch": 2.62, "grad_norm": 4.79144287109375, "learning_rate": 1.6876574307304786e-05, "loss": 0.1917, "step": 1042 }, { "epoch": 2.63, "grad_norm": 6.743935585021973, "learning_rate": 1.6863979848866498e-05, "loss": 0.2646, "step": 1043 }, { "epoch": 2.63, "grad_norm": 19.601837158203125, "learning_rate": 1.6851385390428214e-05, "loss": 0.2423, "step": 1044 }, { "epoch": 2.63, "grad_norm": 2.839597225189209, "learning_rate": 1.6838790931989923e-05, "loss": 0.0257, "step": 1045 }, { "epoch": 2.63, "grad_norm": 5.611330509185791, "learning_rate": 1.6826196473551635e-05, "loss": 0.2994, "step": 1046 }, { "epoch": 2.64, "grad_norm": 9.278888702392578, "learning_rate": 1.681360201511335e-05, "loss": 0.311, "step": 1047 }, { "epoch": 2.64, "grad_norm": 14.311319351196289, "learning_rate": 1.6801007556675064e-05, "loss": 0.0693, "step": 1048 }, { "epoch": 2.64, "grad_norm": 22.729650497436523, "learning_rate": 1.678841309823678e-05, "loss": 0.2257, "step": 1049 }, { "epoch": 2.64, "grad_norm": 12.39243221282959, "learning_rate": 1.677581863979849e-05, "loss": 0.3833, "step": 1050 }, { "epoch": 2.65, "grad_norm": 14.148759841918945, "learning_rate": 1.67632241813602e-05, "loss": 0.2577, "step": 1051 }, { "epoch": 2.65, "grad_norm": 20.707704544067383, "learning_rate": 1.6750629722921917e-05, "loss": 0.4987, "step": 1052 }, { "epoch": 2.65, "grad_norm": 9.586409568786621, "learning_rate": 1.673803526448363e-05, "loss": 0.8248, "step": 1053 }, { "epoch": 2.65, "grad_norm": 6.405531406402588, "learning_rate": 1.6725440806045338e-05, "loss": 0.3748, "step": 1054 }, { "epoch": 2.66, "grad_norm": 9.89402961730957, "learning_rate": 1.6712846347607054e-05, "loss": 0.0682, "step": 1055 }, { "epoch": 2.66, "grad_norm": 2.5575547218322754, "learning_rate": 1.6700251889168767e-05, "loss": 0.0412, "step": 1056 }, { "epoch": 2.66, "grad_norm": 13.999021530151367, "learning_rate": 1.668765743073048e-05, "loss": 0.3433, "step": 1057 }, { "epoch": 2.66, "grad_norm": 17.20004653930664, "learning_rate": 1.6675062972292195e-05, "loss": 0.1682, "step": 1058 }, { "epoch": 2.67, "grad_norm": 15.041619300842285, "learning_rate": 1.6662468513853904e-05, "loss": 0.3703, "step": 1059 }, { "epoch": 2.67, "grad_norm": 19.173490524291992, "learning_rate": 1.6649874055415616e-05, "loss": 0.234, "step": 1060 }, { "epoch": 2.67, "grad_norm": 4.27054500579834, "learning_rate": 1.6637279596977332e-05, "loss": 0.0553, "step": 1061 }, { "epoch": 2.68, "grad_norm": 19.725751876831055, "learning_rate": 1.6624685138539044e-05, "loss": 0.3931, "step": 1062 }, { "epoch": 2.68, "grad_norm": 12.060813903808594, "learning_rate": 1.6612090680100757e-05, "loss": 0.1505, "step": 1063 }, { "epoch": 2.68, "grad_norm": 29.19813346862793, "learning_rate": 1.659949622166247e-05, "loss": 0.1956, "step": 1064 }, { "epoch": 2.68, "grad_norm": 5.451413154602051, "learning_rate": 1.6586901763224182e-05, "loss": 0.0523, "step": 1065 }, { "epoch": 2.69, "grad_norm": 21.697498321533203, "learning_rate": 1.6574307304785894e-05, "loss": 0.3233, "step": 1066 }, { "epoch": 2.69, "grad_norm": 16.664989471435547, "learning_rate": 1.656171284634761e-05, "loss": 0.1331, "step": 1067 }, { "epoch": 2.69, "grad_norm": 9.683480262756348, "learning_rate": 1.654911838790932e-05, "loss": 0.221, "step": 1068 }, { "epoch": 2.69, "grad_norm": 12.325887680053711, "learning_rate": 1.653652392947103e-05, "loss": 0.0978, "step": 1069 }, { "epoch": 2.7, "grad_norm": 6.554139614105225, "learning_rate": 1.6523929471032747e-05, "loss": 0.1858, "step": 1070 }, { "epoch": 2.7, "grad_norm": 18.414140701293945, "learning_rate": 1.651133501259446e-05, "loss": 0.4893, "step": 1071 }, { "epoch": 2.7, "grad_norm": 9.108610153198242, "learning_rate": 1.6498740554156172e-05, "loss": 0.0853, "step": 1072 }, { "epoch": 2.7, "grad_norm": 12.5266752243042, "learning_rate": 1.6486146095717885e-05, "loss": 0.478, "step": 1073 }, { "epoch": 2.71, "grad_norm": 1.9131548404693604, "learning_rate": 1.6473551637279597e-05, "loss": 0.0385, "step": 1074 }, { "epoch": 2.71, "grad_norm": 22.296390533447266, "learning_rate": 1.646095717884131e-05, "loss": 0.3907, "step": 1075 }, { "epoch": 2.71, "grad_norm": 7.380468368530273, "learning_rate": 1.6448362720403025e-05, "loss": 0.2112, "step": 1076 }, { "epoch": 2.71, "grad_norm": 8.03321647644043, "learning_rate": 1.6435768261964738e-05, "loss": 0.268, "step": 1077 }, { "epoch": 2.72, "grad_norm": 28.10780906677246, "learning_rate": 1.6423173803526447e-05, "loss": 0.1584, "step": 1078 }, { "epoch": 2.72, "grad_norm": 14.223876953125, "learning_rate": 1.6410579345088163e-05, "loss": 0.28, "step": 1079 }, { "epoch": 2.72, "grad_norm": 6.363096237182617, "learning_rate": 1.6397984886649875e-05, "loss": 0.3052, "step": 1080 }, { "epoch": 2.72, "grad_norm": 11.782586097717285, "learning_rate": 1.6385390428211587e-05, "loss": 0.2498, "step": 1081 }, { "epoch": 2.73, "grad_norm": 6.359012126922607, "learning_rate": 1.63727959697733e-05, "loss": 0.3136, "step": 1082 }, { "epoch": 2.73, "grad_norm": 6.206680774688721, "learning_rate": 1.6360201511335012e-05, "loss": 0.106, "step": 1083 }, { "epoch": 2.73, "grad_norm": 13.259072303771973, "learning_rate": 1.6347607052896725e-05, "loss": 0.4578, "step": 1084 }, { "epoch": 2.73, "grad_norm": 11.283618927001953, "learning_rate": 1.633501259445844e-05, "loss": 0.1433, "step": 1085 }, { "epoch": 2.74, "grad_norm": 14.00245475769043, "learning_rate": 1.6322418136020153e-05, "loss": 0.1216, "step": 1086 }, { "epoch": 2.74, "grad_norm": 20.484031677246094, "learning_rate": 1.6309823677581862e-05, "loss": 0.6095, "step": 1087 }, { "epoch": 2.74, "grad_norm": 4.319401264190674, "learning_rate": 1.6297229219143578e-05, "loss": 0.3035, "step": 1088 }, { "epoch": 2.74, "grad_norm": 8.893186569213867, "learning_rate": 1.628463476070529e-05, "loss": 0.2453, "step": 1089 }, { "epoch": 2.75, "grad_norm": 20.96607208251953, "learning_rate": 1.6272040302267003e-05, "loss": 0.3969, "step": 1090 }, { "epoch": 2.75, "grad_norm": 13.960694313049316, "learning_rate": 1.6259445843828715e-05, "loss": 0.2462, "step": 1091 }, { "epoch": 2.75, "grad_norm": 9.796236991882324, "learning_rate": 1.6246851385390428e-05, "loss": 0.2312, "step": 1092 }, { "epoch": 2.75, "grad_norm": 23.649677276611328, "learning_rate": 1.623425692695214e-05, "loss": 0.2756, "step": 1093 }, { "epoch": 2.76, "grad_norm": 11.236509323120117, "learning_rate": 1.6221662468513856e-05, "loss": 0.1333, "step": 1094 }, { "epoch": 2.76, "grad_norm": 19.394977569580078, "learning_rate": 1.6209068010075568e-05, "loss": 0.6215, "step": 1095 }, { "epoch": 2.76, "grad_norm": 5.315592288970947, "learning_rate": 1.6196473551637277e-05, "loss": 0.0702, "step": 1096 }, { "epoch": 2.76, "grad_norm": 18.263957977294922, "learning_rate": 1.6183879093198993e-05, "loss": 0.3489, "step": 1097 }, { "epoch": 2.77, "grad_norm": 12.144902229309082, "learning_rate": 1.6171284634760705e-05, "loss": 0.1314, "step": 1098 }, { "epoch": 2.77, "grad_norm": 10.130423545837402, "learning_rate": 1.6158690176322418e-05, "loss": 0.1701, "step": 1099 }, { "epoch": 2.77, "grad_norm": 15.319389343261719, "learning_rate": 1.6146095717884134e-05, "loss": 0.1384, "step": 1100 }, { "epoch": 2.77, "grad_norm": 3.53900146484375, "learning_rate": 1.6133501259445843e-05, "loss": 0.3098, "step": 1101 }, { "epoch": 2.78, "grad_norm": 2.5617988109588623, "learning_rate": 1.6120906801007555e-05, "loss": 0.0469, "step": 1102 }, { "epoch": 2.78, "grad_norm": 7.703180313110352, "learning_rate": 1.610831234256927e-05, "loss": 0.2913, "step": 1103 }, { "epoch": 2.78, "grad_norm": 9.907478332519531, "learning_rate": 1.6095717884130983e-05, "loss": 0.34, "step": 1104 }, { "epoch": 2.78, "grad_norm": 6.849803924560547, "learning_rate": 1.6083123425692692e-05, "loss": 0.253, "step": 1105 }, { "epoch": 2.79, "grad_norm": 13.280983924865723, "learning_rate": 1.607052896725441e-05, "loss": 0.3651, "step": 1106 }, { "epoch": 2.79, "grad_norm": 13.153647422790527, "learning_rate": 1.605793450881612e-05, "loss": 0.3285, "step": 1107 }, { "epoch": 2.79, "grad_norm": 8.686355590820312, "learning_rate": 1.6045340050377833e-05, "loss": 0.3197, "step": 1108 }, { "epoch": 2.79, "grad_norm": 7.679217338562012, "learning_rate": 1.603274559193955e-05, "loss": 0.3772, "step": 1109 }, { "epoch": 2.8, "grad_norm": 3.92250657081604, "learning_rate": 1.6020151133501258e-05, "loss": 0.0491, "step": 1110 }, { "epoch": 2.8, "grad_norm": 8.080659866333008, "learning_rate": 1.600755667506297e-05, "loss": 0.1022, "step": 1111 }, { "epoch": 2.8, "grad_norm": 10.759270668029785, "learning_rate": 1.5994962216624686e-05, "loss": 0.2858, "step": 1112 }, { "epoch": 2.8, "grad_norm": 2.1772842407226562, "learning_rate": 1.59823677581864e-05, "loss": 0.0408, "step": 1113 }, { "epoch": 2.81, "grad_norm": 17.511062622070312, "learning_rate": 1.596977329974811e-05, "loss": 0.2311, "step": 1114 }, { "epoch": 2.81, "grad_norm": 3.6058473587036133, "learning_rate": 1.5957178841309824e-05, "loss": 0.0536, "step": 1115 }, { "epoch": 2.81, "grad_norm": 12.71088981628418, "learning_rate": 1.5944584382871536e-05, "loss": 0.1877, "step": 1116 }, { "epoch": 2.81, "grad_norm": 4.7909345626831055, "learning_rate": 1.593198992443325e-05, "loss": 0.2639, "step": 1117 }, { "epoch": 2.82, "grad_norm": 19.712451934814453, "learning_rate": 1.5919395465994964e-05, "loss": 0.3709, "step": 1118 }, { "epoch": 2.82, "grad_norm": 25.3231258392334, "learning_rate": 1.5906801007556673e-05, "loss": 0.3071, "step": 1119 }, { "epoch": 2.82, "grad_norm": 23.117774963378906, "learning_rate": 1.589420654911839e-05, "loss": 0.3033, "step": 1120 }, { "epoch": 2.82, "grad_norm": 16.745607376098633, "learning_rate": 1.58816120906801e-05, "loss": 0.3242, "step": 1121 }, { "epoch": 2.83, "grad_norm": 12.918922424316406, "learning_rate": 1.5869017632241814e-05, "loss": 0.1945, "step": 1122 }, { "epoch": 2.83, "grad_norm": 12.439324378967285, "learning_rate": 1.585642317380353e-05, "loss": 0.3462, "step": 1123 }, { "epoch": 2.83, "grad_norm": 12.006231307983398, "learning_rate": 1.584382871536524e-05, "loss": 0.3037, "step": 1124 }, { "epoch": 2.83, "grad_norm": 19.369873046875, "learning_rate": 1.583123425692695e-05, "loss": 0.4054, "step": 1125 }, { "epoch": 2.84, "grad_norm": 7.270785331726074, "learning_rate": 1.5818639798488667e-05, "loss": 0.1068, "step": 1126 }, { "epoch": 2.84, "grad_norm": 16.919490814208984, "learning_rate": 1.580604534005038e-05, "loss": 0.1851, "step": 1127 }, { "epoch": 2.84, "grad_norm": 14.456689834594727, "learning_rate": 1.5793450881612092e-05, "loss": 0.2693, "step": 1128 }, { "epoch": 2.84, "grad_norm": 32.42156982421875, "learning_rate": 1.5780856423173804e-05, "loss": 0.4212, "step": 1129 }, { "epoch": 2.85, "grad_norm": 14.696311950683594, "learning_rate": 1.5768261964735517e-05, "loss": 0.3479, "step": 1130 }, { "epoch": 2.85, "grad_norm": 5.063379287719727, "learning_rate": 1.575566750629723e-05, "loss": 0.0615, "step": 1131 }, { "epoch": 2.85, "grad_norm": 18.039127349853516, "learning_rate": 1.5743073047858945e-05, "loss": 0.2822, "step": 1132 }, { "epoch": 2.85, "grad_norm": 10.71660327911377, "learning_rate": 1.5730478589420654e-05, "loss": 0.0876, "step": 1133 }, { "epoch": 2.86, "grad_norm": 13.763422012329102, "learning_rate": 1.5717884130982367e-05, "loss": 0.1789, "step": 1134 }, { "epoch": 2.86, "grad_norm": 18.239368438720703, "learning_rate": 1.5705289672544082e-05, "loss": 0.621, "step": 1135 }, { "epoch": 2.86, "grad_norm": 19.79184341430664, "learning_rate": 1.5692695214105795e-05, "loss": 0.3946, "step": 1136 }, { "epoch": 2.86, "grad_norm": 13.332741737365723, "learning_rate": 1.5680100755667507e-05, "loss": 0.3358, "step": 1137 }, { "epoch": 2.87, "grad_norm": 12.152838706970215, "learning_rate": 1.566750629722922e-05, "loss": 0.3625, "step": 1138 }, { "epoch": 2.87, "grad_norm": 19.56678009033203, "learning_rate": 1.5654911838790932e-05, "loss": 0.3858, "step": 1139 }, { "epoch": 2.87, "grad_norm": 25.041582107543945, "learning_rate": 1.5642317380352644e-05, "loss": 0.374, "step": 1140 }, { "epoch": 2.87, "grad_norm": 12.966362953186035, "learning_rate": 1.562972292191436e-05, "loss": 0.1982, "step": 1141 }, { "epoch": 2.88, "grad_norm": 4.819403648376465, "learning_rate": 1.5617128463476073e-05, "loss": 0.059, "step": 1142 }, { "epoch": 2.88, "grad_norm": 11.310750007629395, "learning_rate": 1.5604534005037782e-05, "loss": 0.2802, "step": 1143 }, { "epoch": 2.88, "grad_norm": 18.544479370117188, "learning_rate": 1.5591939546599498e-05, "loss": 0.3296, "step": 1144 }, { "epoch": 2.88, "grad_norm": 5.440631866455078, "learning_rate": 1.557934508816121e-05, "loss": 0.0817, "step": 1145 }, { "epoch": 2.89, "grad_norm": 17.53709602355957, "learning_rate": 1.5566750629722922e-05, "loss": 0.2045, "step": 1146 }, { "epoch": 2.89, "grad_norm": 5.029760360717773, "learning_rate": 1.5554156171284635e-05, "loss": 0.0772, "step": 1147 }, { "epoch": 2.89, "grad_norm": 15.92929744720459, "learning_rate": 1.5541561712846347e-05, "loss": 0.2221, "step": 1148 }, { "epoch": 2.89, "grad_norm": 14.654315948486328, "learning_rate": 1.552896725440806e-05, "loss": 0.194, "step": 1149 }, { "epoch": 2.9, "grad_norm": 10.082984924316406, "learning_rate": 1.5516372795969776e-05, "loss": 0.3038, "step": 1150 }, { "epoch": 2.9, "grad_norm": 12.362741470336914, "learning_rate": 1.5503778337531488e-05, "loss": 0.1464, "step": 1151 }, { "epoch": 2.9, "grad_norm": 11.521105766296387, "learning_rate": 1.5491183879093197e-05, "loss": 0.2223, "step": 1152 }, { "epoch": 2.9, "grad_norm": 3.9992454051971436, "learning_rate": 1.5478589420654913e-05, "loss": 0.0534, "step": 1153 }, { "epoch": 2.91, "grad_norm": 9.422941207885742, "learning_rate": 1.5465994962216625e-05, "loss": 0.1071, "step": 1154 }, { "epoch": 2.91, "grad_norm": 6.80059289932251, "learning_rate": 1.5453400503778338e-05, "loss": 0.3005, "step": 1155 }, { "epoch": 2.91, "grad_norm": 5.678284168243408, "learning_rate": 1.5440806045340053e-05, "loss": 0.0574, "step": 1156 }, { "epoch": 2.91, "grad_norm": 4.542283058166504, "learning_rate": 1.5428211586901763e-05, "loss": 0.2604, "step": 1157 }, { "epoch": 2.92, "grad_norm": 4.583611011505127, "learning_rate": 1.5415617128463475e-05, "loss": 0.0593, "step": 1158 }, { "epoch": 2.92, "grad_norm": 9.688311576843262, "learning_rate": 1.540302267002519e-05, "loss": 0.3942, "step": 1159 }, { "epoch": 2.92, "grad_norm": 13.455863952636719, "learning_rate": 1.5390428211586903e-05, "loss": 0.0816, "step": 1160 }, { "epoch": 2.92, "grad_norm": 18.235370635986328, "learning_rate": 1.5377833753148612e-05, "loss": 0.1245, "step": 1161 }, { "epoch": 2.93, "grad_norm": 11.47364330291748, "learning_rate": 1.5365239294710328e-05, "loss": 0.2704, "step": 1162 }, { "epoch": 2.93, "grad_norm": 9.850740432739258, "learning_rate": 1.535264483627204e-05, "loss": 0.2232, "step": 1163 }, { "epoch": 2.93, "grad_norm": 12.394991874694824, "learning_rate": 1.5340050377833753e-05, "loss": 0.0972, "step": 1164 }, { "epoch": 2.93, "grad_norm": 6.109569072723389, "learning_rate": 1.532745591939547e-05, "loss": 0.0611, "step": 1165 }, { "epoch": 2.94, "grad_norm": 15.410727500915527, "learning_rate": 1.5314861460957178e-05, "loss": 0.2336, "step": 1166 }, { "epoch": 2.94, "grad_norm": 9.841590881347656, "learning_rate": 1.530226700251889e-05, "loss": 0.2135, "step": 1167 }, { "epoch": 2.94, "grad_norm": 2.900602102279663, "learning_rate": 1.5289672544080606e-05, "loss": 0.2371, "step": 1168 }, { "epoch": 2.94, "grad_norm": 9.744848251342773, "learning_rate": 1.527707808564232e-05, "loss": 0.117, "step": 1169 }, { "epoch": 2.95, "grad_norm": 20.94938850402832, "learning_rate": 1.526448362720403e-05, "loss": 0.2436, "step": 1170 }, { "epoch": 2.95, "grad_norm": 4.247443199157715, "learning_rate": 1.5251889168765742e-05, "loss": 0.1908, "step": 1171 }, { "epoch": 2.95, "grad_norm": 11.222188949584961, "learning_rate": 1.5239294710327456e-05, "loss": 0.1171, "step": 1172 }, { "epoch": 2.95, "grad_norm": 0.8100425601005554, "learning_rate": 1.522670025188917e-05, "loss": 0.0229, "step": 1173 }, { "epoch": 2.96, "grad_norm": 13.735158920288086, "learning_rate": 1.5214105793450882e-05, "loss": 0.4922, "step": 1174 }, { "epoch": 2.96, "grad_norm": 40.11631774902344, "learning_rate": 1.5201511335012593e-05, "loss": 0.1388, "step": 1175 }, { "epoch": 2.96, "grad_norm": 29.13542938232422, "learning_rate": 1.5188916876574307e-05, "loss": 0.1924, "step": 1176 }, { "epoch": 2.96, "grad_norm": 7.6980767250061035, "learning_rate": 1.5176322418136021e-05, "loss": 0.0766, "step": 1177 }, { "epoch": 2.97, "grad_norm": 14.650445938110352, "learning_rate": 1.5163727959697734e-05, "loss": 0.211, "step": 1178 }, { "epoch": 2.97, "grad_norm": 15.319862365722656, "learning_rate": 1.5151133501259448e-05, "loss": 0.2816, "step": 1179 }, { "epoch": 2.97, "grad_norm": 5.754809856414795, "learning_rate": 1.5138539042821159e-05, "loss": 0.263, "step": 1180 }, { "epoch": 2.97, "grad_norm": 16.632143020629883, "learning_rate": 1.5125944584382871e-05, "loss": 0.2188, "step": 1181 }, { "epoch": 2.98, "grad_norm": 12.786446571350098, "learning_rate": 1.5113350125944585e-05, "loss": 0.235, "step": 1182 }, { "epoch": 2.98, "grad_norm": 16.24946403503418, "learning_rate": 1.51007556675063e-05, "loss": 0.3029, "step": 1183 }, { "epoch": 2.98, "grad_norm": 12.54625129699707, "learning_rate": 1.5088161209068012e-05, "loss": 0.2153, "step": 1184 }, { "epoch": 2.98, "grad_norm": 18.512767791748047, "learning_rate": 1.5075566750629722e-05, "loss": 0.6106, "step": 1185 }, { "epoch": 2.99, "grad_norm": 22.6643009185791, "learning_rate": 1.5062972292191437e-05, "loss": 0.3158, "step": 1186 }, { "epoch": 2.99, "grad_norm": 3.4380409717559814, "learning_rate": 1.5050377833753149e-05, "loss": 0.2336, "step": 1187 }, { "epoch": 2.99, "grad_norm": 18.3015079498291, "learning_rate": 1.5037783375314863e-05, "loss": 0.3209, "step": 1188 }, { "epoch": 2.99, "grad_norm": 10.597107887268066, "learning_rate": 1.5025188916876574e-05, "loss": 0.0967, "step": 1189 }, { "epoch": 3.0, "grad_norm": 10.336237907409668, "learning_rate": 1.5012594458438286e-05, "loss": 0.1293, "step": 1190 }, { "epoch": 3.0, "grad_norm": 2.06474232673645, "learning_rate": 1.5e-05, "loss": 0.0459, "step": 1191 }, { "epoch": 3.0, "eval_accuracy": 0.8088328075709779, "eval_f1": 0.8080378252480893, "eval_loss": 0.5197204947471619, "eval_precision": 0.8076567604174967, "eval_recall": 0.8105337313586474, "eval_runtime": 1211.4212, "eval_samples_per_second": 1.308, "eval_steps_per_second": 0.083, "step": 1191 } ], "logging_steps": 1, "max_steps": 2382, "num_input_tokens_seen": 0, "num_train_epochs": 6, "save_steps": 500, "total_flos": 5002004273448960.0, "train_batch_size": 16, "trial_name": null, "trial_params": null }