diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,8577 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.352018816150529, + "eval_steps": 700, + "global_step": 12000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.001960015680125441, + "grad_norm": 7.928689002990723, + "learning_rate": 8.333333333333334e-08, + "loss": 1.4388, + "step": 10 + }, + { + "epoch": 0.003920031360250882, + "grad_norm": 9.54440689086914, + "learning_rate": 1.6666666666666668e-07, + "loss": 1.4362, + "step": 20 + }, + { + "epoch": 0.005880047040376323, + "grad_norm": 9.008377075195312, + "learning_rate": 2.5000000000000004e-07, + "loss": 1.4445, + "step": 30 + }, + { + "epoch": 0.007840062720501764, + "grad_norm": 7.416823863983154, + "learning_rate": 3.3333333333333335e-07, + "loss": 1.4217, + "step": 40 + }, + { + "epoch": 0.009800078400627205, + "grad_norm": 7.977886199951172, + "learning_rate": 4.1666666666666667e-07, + "loss": 1.4243, + "step": 50 + }, + { + "epoch": 0.011760094080752646, + "grad_norm": 8.24797248840332, + "learning_rate": 5.000000000000001e-07, + "loss": 1.3976, + "step": 60 + }, + { + "epoch": 0.013720109760878087, + "grad_norm": 7.240636348724365, + "learning_rate": 5.833333333333334e-07, + "loss": 1.3853, + "step": 70 + }, + { + "epoch": 0.01568012544100353, + "grad_norm": 5.709108352661133, + "learning_rate": 6.666666666666667e-07, + "loss": 1.3353, + "step": 80 + }, + { + "epoch": 0.01764014112112897, + "grad_norm": 9.35693073272705, + "learning_rate": 7.5e-07, + "loss": 1.3051, + "step": 90 + }, + { + "epoch": 0.01960015680125441, + "grad_norm": 47.39125442504883, + "learning_rate": 8.333333333333333e-07, + "loss": 1.265, + "step": 100 + }, + { + "epoch": 0.02156017248137985, + "grad_norm": 5.6352434158325195, + "learning_rate": 9.166666666666666e-07, + "loss": 1.2583, + "step": 110 + }, + { + "epoch": 0.023520188161505293, + "grad_norm": 6.022407531738281, + "learning_rate": 1.0000000000000002e-06, + "loss": 1.2398, + "step": 120 + }, + { + "epoch": 0.025480203841630734, + "grad_norm": 6.143012046813965, + "learning_rate": 1.0833333333333335e-06, + "loss": 1.2237, + "step": 130 + }, + { + "epoch": 0.027440219521756175, + "grad_norm": 6.173267841339111, + "learning_rate": 1.1666666666666668e-06, + "loss": 1.2246, + "step": 140 + }, + { + "epoch": 0.029400235201881616, + "grad_norm": 5.4503092765808105, + "learning_rate": 1.25e-06, + "loss": 1.1993, + "step": 150 + }, + { + "epoch": 0.03136025088200706, + "grad_norm": 17.596582412719727, + "learning_rate": 1.3333333333333334e-06, + "loss": 1.1889, + "step": 160 + }, + { + "epoch": 0.033320266562132494, + "grad_norm": 6.131169319152832, + "learning_rate": 1.4166666666666667e-06, + "loss": 1.1884, + "step": 170 + }, + { + "epoch": 0.03528028224225794, + "grad_norm": 5.450586795806885, + "learning_rate": 1.5e-06, + "loss": 1.1622, + "step": 180 + }, + { + "epoch": 0.03724029792238338, + "grad_norm": 14.713933944702148, + "learning_rate": 1.5833333333333333e-06, + "loss": 1.1675, + "step": 190 + }, + { + "epoch": 0.03920031360250882, + "grad_norm": 8.510261535644531, + "learning_rate": 1.6666666666666667e-06, + "loss": 1.1584, + "step": 200 + }, + { + "epoch": 0.04116032928263426, + "grad_norm": 6.9118971824646, + "learning_rate": 1.75e-06, + "loss": 1.1696, + "step": 210 + }, + { + "epoch": 0.0431203449627597, + "grad_norm": 5.986424922943115, + "learning_rate": 1.8333333333333333e-06, + "loss": 1.1734, + "step": 220 + }, + { + "epoch": 0.04508036064288514, + "grad_norm": 7.179569244384766, + "learning_rate": 1.916666666666667e-06, + "loss": 1.1563, + "step": 230 + }, + { + "epoch": 0.047040376323010585, + "grad_norm": 7.15562629699707, + "learning_rate": 2.0000000000000003e-06, + "loss": 1.1443, + "step": 240 + }, + { + "epoch": 0.04900039200313602, + "grad_norm": 5.163909912109375, + "learning_rate": 2.0833333333333334e-06, + "loss": 1.1567, + "step": 250 + }, + { + "epoch": 0.05096040768326147, + "grad_norm": 5.861210346221924, + "learning_rate": 2.166666666666667e-06, + "loss": 1.1596, + "step": 260 + }, + { + "epoch": 0.052920423363386905, + "grad_norm": 5.1880292892456055, + "learning_rate": 2.25e-06, + "loss": 1.1428, + "step": 270 + }, + { + "epoch": 0.05488043904351235, + "grad_norm": 6.203003406524658, + "learning_rate": 2.3333333333333336e-06, + "loss": 1.1463, + "step": 280 + }, + { + "epoch": 0.05684045472363779, + "grad_norm": 6.191583156585693, + "learning_rate": 2.4166666666666667e-06, + "loss": 1.1538, + "step": 290 + }, + { + "epoch": 0.05880047040376323, + "grad_norm": 9.0908203125, + "learning_rate": 2.5e-06, + "loss": 1.1277, + "step": 300 + }, + { + "epoch": 0.06076048608388867, + "grad_norm": 5.685515880584717, + "learning_rate": 2.5833333333333337e-06, + "loss": 1.1471, + "step": 310 + }, + { + "epoch": 0.06272050176401411, + "grad_norm": 6.010474681854248, + "learning_rate": 2.666666666666667e-06, + "loss": 1.1232, + "step": 320 + }, + { + "epoch": 0.06468051744413955, + "grad_norm": 9.0894193649292, + "learning_rate": 2.7500000000000004e-06, + "loss": 1.1286, + "step": 330 + }, + { + "epoch": 0.06664053312426499, + "grad_norm": 5.936020851135254, + "learning_rate": 2.8333333333333335e-06, + "loss": 1.1116, + "step": 340 + }, + { + "epoch": 0.06860054880439044, + "grad_norm": 6.055352210998535, + "learning_rate": 2.916666666666667e-06, + "loss": 1.1167, + "step": 350 + }, + { + "epoch": 0.07056056448451588, + "grad_norm": 5.160061836242676, + "learning_rate": 3e-06, + "loss": 1.1225, + "step": 360 + }, + { + "epoch": 0.07252058016464132, + "grad_norm": 5.748876571655273, + "learning_rate": 3.0833333333333336e-06, + "loss": 1.1298, + "step": 370 + }, + { + "epoch": 0.07448059584476675, + "grad_norm": 5.9786057472229, + "learning_rate": 3.1666666666666667e-06, + "loss": 1.1394, + "step": 380 + }, + { + "epoch": 0.0764406115248922, + "grad_norm": 5.474339008331299, + "learning_rate": 3.2500000000000002e-06, + "loss": 1.138, + "step": 390 + }, + { + "epoch": 0.07840062720501764, + "grad_norm": 6.028172016143799, + "learning_rate": 3.3333333333333333e-06, + "loss": 1.1243, + "step": 400 + }, + { + "epoch": 0.08036064288514308, + "grad_norm": 11.877293586730957, + "learning_rate": 3.416666666666667e-06, + "loss": 1.1086, + "step": 410 + }, + { + "epoch": 0.08232065856526852, + "grad_norm": 6.457036972045898, + "learning_rate": 3.5e-06, + "loss": 1.1113, + "step": 420 + }, + { + "epoch": 0.08428067424539397, + "grad_norm": 8.589061737060547, + "learning_rate": 3.5833333333333335e-06, + "loss": 1.1207, + "step": 430 + }, + { + "epoch": 0.0862406899255194, + "grad_norm": 6.367908000946045, + "learning_rate": 3.6666666666666666e-06, + "loss": 1.1192, + "step": 440 + }, + { + "epoch": 0.08820070560564484, + "grad_norm": 6.087027549743652, + "learning_rate": 3.7500000000000005e-06, + "loss": 1.114, + "step": 450 + }, + { + "epoch": 0.09016072128577028, + "grad_norm": 6.365817546844482, + "learning_rate": 3.833333333333334e-06, + "loss": 1.1104, + "step": 460 + }, + { + "epoch": 0.09212073696589573, + "grad_norm": 6.38438606262207, + "learning_rate": 3.916666666666667e-06, + "loss": 1.1222, + "step": 470 + }, + { + "epoch": 0.09408075264602117, + "grad_norm": 5.983357906341553, + "learning_rate": 4.000000000000001e-06, + "loss": 1.1082, + "step": 480 + }, + { + "epoch": 0.09604076832614661, + "grad_norm": 7.514024257659912, + "learning_rate": 4.083333333333334e-06, + "loss": 1.1305, + "step": 490 + }, + { + "epoch": 0.09800078400627205, + "grad_norm": 5.6476922035217285, + "learning_rate": 4.166666666666667e-06, + "loss": 1.1183, + "step": 500 + }, + { + "epoch": 0.0999607996863975, + "grad_norm": 6.5463128089904785, + "learning_rate": 4.25e-06, + "loss": 1.1126, + "step": 510 + }, + { + "epoch": 0.10192081536652293, + "grad_norm": 6.156862258911133, + "learning_rate": 4.333333333333334e-06, + "loss": 1.1161, + "step": 520 + }, + { + "epoch": 0.10388083104664837, + "grad_norm": 5.682535648345947, + "learning_rate": 4.416666666666667e-06, + "loss": 1.1096, + "step": 530 + }, + { + "epoch": 0.10584084672677381, + "grad_norm": 7.1881489753723145, + "learning_rate": 4.5e-06, + "loss": 1.1136, + "step": 540 + }, + { + "epoch": 0.10780086240689926, + "grad_norm": 5.990985870361328, + "learning_rate": 4.583333333333333e-06, + "loss": 1.1224, + "step": 550 + }, + { + "epoch": 0.1097608780870247, + "grad_norm": 5.474586009979248, + "learning_rate": 4.666666666666667e-06, + "loss": 1.1078, + "step": 560 + }, + { + "epoch": 0.11172089376715014, + "grad_norm": 5.7349371910095215, + "learning_rate": 4.75e-06, + "loss": 1.1238, + "step": 570 + }, + { + "epoch": 0.11368090944727557, + "grad_norm": 7.513430595397949, + "learning_rate": 4.833333333333333e-06, + "loss": 1.1158, + "step": 580 + }, + { + "epoch": 0.11564092512740103, + "grad_norm": 5.410061836242676, + "learning_rate": 4.9166666666666665e-06, + "loss": 1.1102, + "step": 590 + }, + { + "epoch": 0.11760094080752646, + "grad_norm": 6.409933567047119, + "learning_rate": 5e-06, + "loss": 1.1035, + "step": 600 + }, + { + "epoch": 0.1195609564876519, + "grad_norm": 14.141221046447754, + "learning_rate": 5.0833333333333335e-06, + "loss": 1.1066, + "step": 610 + }, + { + "epoch": 0.12152097216777734, + "grad_norm": 5.547974586486816, + "learning_rate": 5.1666666666666675e-06, + "loss": 1.1218, + "step": 620 + }, + { + "epoch": 0.12348098784790279, + "grad_norm": 5.946651935577393, + "learning_rate": 5.2500000000000006e-06, + "loss": 1.112, + "step": 630 + }, + { + "epoch": 0.12544100352802823, + "grad_norm": 6.724623680114746, + "learning_rate": 5.333333333333334e-06, + "loss": 1.0986, + "step": 640 + }, + { + "epoch": 0.12740101920815367, + "grad_norm": 8.897869110107422, + "learning_rate": 5.416666666666667e-06, + "loss": 1.1077, + "step": 650 + }, + { + "epoch": 0.1293610348882791, + "grad_norm": 6.75730037689209, + "learning_rate": 5.500000000000001e-06, + "loss": 1.1018, + "step": 660 + }, + { + "epoch": 0.13132105056840454, + "grad_norm": 5.3133111000061035, + "learning_rate": 5.583333333333334e-06, + "loss": 1.1089, + "step": 670 + }, + { + "epoch": 0.13328106624852998, + "grad_norm": 5.644856929779053, + "learning_rate": 5.666666666666667e-06, + "loss": 1.1159, + "step": 680 + }, + { + "epoch": 0.13524108192865542, + "grad_norm": 5.264536380767822, + "learning_rate": 5.75e-06, + "loss": 1.1068, + "step": 690 + }, + { + "epoch": 0.13720109760878088, + "grad_norm": 5.970139026641846, + "learning_rate": 5.833333333333334e-06, + "loss": 1.1214, + "step": 700 + }, + { + "epoch": 0.13720109760878088, + "eval_loss": 1.0996023416519165, + "eval_runtime": 14.1011, + "eval_samples_per_second": 46.379, + "eval_steps_per_second": 5.815, + "step": 700 + }, + { + "epoch": 0.13916111328890632, + "grad_norm": 5.035984039306641, + "learning_rate": 5.916666666666667e-06, + "loss": 1.1078, + "step": 710 + }, + { + "epoch": 0.14112112896903176, + "grad_norm": 5.657273292541504, + "learning_rate": 6e-06, + "loss": 1.0998, + "step": 720 + }, + { + "epoch": 0.1430811446491572, + "grad_norm": 5.413987636566162, + "learning_rate": 6.083333333333333e-06, + "loss": 1.1023, + "step": 730 + }, + { + "epoch": 0.14504116032928263, + "grad_norm": 7.05068302154541, + "learning_rate": 6.166666666666667e-06, + "loss": 1.1053, + "step": 740 + }, + { + "epoch": 0.14700117600940807, + "grad_norm": 6.170547008514404, + "learning_rate": 6.25e-06, + "loss": 1.0998, + "step": 750 + }, + { + "epoch": 0.1489611916895335, + "grad_norm": 5.373660087585449, + "learning_rate": 6.333333333333333e-06, + "loss": 1.0899, + "step": 760 + }, + { + "epoch": 0.15092120736965894, + "grad_norm": 6.981970310211182, + "learning_rate": 6.416666666666667e-06, + "loss": 1.113, + "step": 770 + }, + { + "epoch": 0.1528812230497844, + "grad_norm": 4.918264865875244, + "learning_rate": 6.5000000000000004e-06, + "loss": 1.0884, + "step": 780 + }, + { + "epoch": 0.15484123872990985, + "grad_norm": 13.080545425415039, + "learning_rate": 6.5833333333333335e-06, + "loss": 1.1012, + "step": 790 + }, + { + "epoch": 0.15680125441003528, + "grad_norm": 5.957456111907959, + "learning_rate": 6.666666666666667e-06, + "loss": 1.1044, + "step": 800 + }, + { + "epoch": 0.15876127009016072, + "grad_norm": 5.442764759063721, + "learning_rate": 6.750000000000001e-06, + "loss": 1.1051, + "step": 810 + }, + { + "epoch": 0.16072128577028616, + "grad_norm": 5.379378318786621, + "learning_rate": 6.833333333333334e-06, + "loss": 1.1092, + "step": 820 + }, + { + "epoch": 0.1626813014504116, + "grad_norm": 6.137138366699219, + "learning_rate": 6.916666666666667e-06, + "loss": 1.0835, + "step": 830 + }, + { + "epoch": 0.16464131713053703, + "grad_norm": 5.297956943511963, + "learning_rate": 7e-06, + "loss": 1.1089, + "step": 840 + }, + { + "epoch": 0.16660133281066247, + "grad_norm": 5.63748025894165, + "learning_rate": 7.083333333333335e-06, + "loss": 1.1024, + "step": 850 + }, + { + "epoch": 0.16856134849078794, + "grad_norm": 5.432644844055176, + "learning_rate": 7.166666666666667e-06, + "loss": 1.0712, + "step": 860 + }, + { + "epoch": 0.17052136417091338, + "grad_norm": 5.4105305671691895, + "learning_rate": 7.25e-06, + "loss": 1.1122, + "step": 870 + }, + { + "epoch": 0.1724813798510388, + "grad_norm": 6.026330471038818, + "learning_rate": 7.333333333333333e-06, + "loss": 1.1049, + "step": 880 + }, + { + "epoch": 0.17444139553116425, + "grad_norm": 6.143797397613525, + "learning_rate": 7.416666666666668e-06, + "loss": 1.1038, + "step": 890 + }, + { + "epoch": 0.1764014112112897, + "grad_norm": 5.558448791503906, + "learning_rate": 7.500000000000001e-06, + "loss": 1.095, + "step": 900 + }, + { + "epoch": 0.17836142689141513, + "grad_norm": 6.965857028961182, + "learning_rate": 7.583333333333333e-06, + "loss": 1.0979, + "step": 910 + }, + { + "epoch": 0.18032144257154056, + "grad_norm": 6.795146942138672, + "learning_rate": 7.666666666666667e-06, + "loss": 1.103, + "step": 920 + }, + { + "epoch": 0.182281458251666, + "grad_norm": 5.742541313171387, + "learning_rate": 7.75e-06, + "loss": 1.1031, + "step": 930 + }, + { + "epoch": 0.18424147393179147, + "grad_norm": 5.67035436630249, + "learning_rate": 7.833333333333333e-06, + "loss": 1.1018, + "step": 940 + }, + { + "epoch": 0.1862014896119169, + "grad_norm": 8.501360893249512, + "learning_rate": 7.916666666666667e-06, + "loss": 1.0916, + "step": 950 + }, + { + "epoch": 0.18816150529204234, + "grad_norm": 5.092253684997559, + "learning_rate": 8.000000000000001e-06, + "loss": 1.0886, + "step": 960 + }, + { + "epoch": 0.19012152097216778, + "grad_norm": 6.499626636505127, + "learning_rate": 8.083333333333334e-06, + "loss": 1.1008, + "step": 970 + }, + { + "epoch": 0.19208153665229322, + "grad_norm": 5.9327216148376465, + "learning_rate": 8.166666666666668e-06, + "loss": 1.0911, + "step": 980 + }, + { + "epoch": 0.19404155233241865, + "grad_norm": 5.990231990814209, + "learning_rate": 8.25e-06, + "loss": 1.0955, + "step": 990 + }, + { + "epoch": 0.1960015680125441, + "grad_norm": 32.17470932006836, + "learning_rate": 8.333333333333334e-06, + "loss": 1.1087, + "step": 1000 + }, + { + "epoch": 0.19796158369266953, + "grad_norm": 5.19512414932251, + "learning_rate": 8.416666666666667e-06, + "loss": 1.1302, + "step": 1010 + }, + { + "epoch": 0.199921599372795, + "grad_norm": 5.765759468078613, + "learning_rate": 8.5e-06, + "loss": 1.1089, + "step": 1020 + }, + { + "epoch": 0.20188161505292043, + "grad_norm": 5.5894927978515625, + "learning_rate": 8.583333333333333e-06, + "loss": 1.0914, + "step": 1030 + }, + { + "epoch": 0.20384163073304587, + "grad_norm": 6.387049198150635, + "learning_rate": 8.666666666666668e-06, + "loss": 1.095, + "step": 1040 + }, + { + "epoch": 0.2058016464131713, + "grad_norm": 5.629969596862793, + "learning_rate": 8.750000000000001e-06, + "loss": 1.0999, + "step": 1050 + }, + { + "epoch": 0.20776166209329675, + "grad_norm": 22.919910430908203, + "learning_rate": 8.833333333333334e-06, + "loss": 1.096, + "step": 1060 + }, + { + "epoch": 0.20972167777342218, + "grad_norm": 5.267335414886475, + "learning_rate": 8.916666666666667e-06, + "loss": 1.1006, + "step": 1070 + }, + { + "epoch": 0.21168169345354762, + "grad_norm": 53.39712905883789, + "learning_rate": 9e-06, + "loss": 1.106, + "step": 1080 + }, + { + "epoch": 0.21364170913367306, + "grad_norm": 5.858781814575195, + "learning_rate": 9.083333333333333e-06, + "loss": 1.2549, + "step": 1090 + }, + { + "epoch": 0.21560172481379852, + "grad_norm": 6.5703935623168945, + "learning_rate": 9.166666666666666e-06, + "loss": 1.094, + "step": 1100 + }, + { + "epoch": 0.21756174049392396, + "grad_norm": 5.414449214935303, + "learning_rate": 9.250000000000001e-06, + "loss": 1.0891, + "step": 1110 + }, + { + "epoch": 0.2195217561740494, + "grad_norm": 7.185405731201172, + "learning_rate": 9.333333333333334e-06, + "loss": 1.089, + "step": 1120 + }, + { + "epoch": 0.22148177185417484, + "grad_norm": 6.4889092445373535, + "learning_rate": 9.416666666666667e-06, + "loss": 1.0874, + "step": 1130 + }, + { + "epoch": 0.22344178753430027, + "grad_norm": 6.4478759765625, + "learning_rate": 9.5e-06, + "loss": 1.1111, + "step": 1140 + }, + { + "epoch": 0.2254018032144257, + "grad_norm": 6.49618673324585, + "learning_rate": 9.583333333333335e-06, + "loss": 1.1001, + "step": 1150 + }, + { + "epoch": 0.22736181889455115, + "grad_norm": 7.893542766571045, + "learning_rate": 9.666666666666667e-06, + "loss": 1.097, + "step": 1160 + }, + { + "epoch": 0.2293218345746766, + "grad_norm": 11.165901184082031, + "learning_rate": 9.75e-06, + "loss": 1.092, + "step": 1170 + }, + { + "epoch": 0.23128185025480205, + "grad_norm": 21.967939376831055, + "learning_rate": 9.833333333333333e-06, + "loss": 1.1216, + "step": 1180 + }, + { + "epoch": 0.2332418659349275, + "grad_norm": 14.317709922790527, + "learning_rate": 9.916666666666668e-06, + "loss": 1.1047, + "step": 1190 + }, + { + "epoch": 0.23520188161505293, + "grad_norm": 6.240993499755859, + "learning_rate": 1e-05, + "loss": 1.1045, + "step": 1200 + }, + { + "epoch": 0.23716189729517836, + "grad_norm": 5.804283142089844, + "learning_rate": 9.990740740740741e-06, + "loss": 1.0933, + "step": 1210 + }, + { + "epoch": 0.2391219129753038, + "grad_norm": 5.778935432434082, + "learning_rate": 9.981481481481482e-06, + "loss": 1.1098, + "step": 1220 + }, + { + "epoch": 0.24108192865542924, + "grad_norm": 6.391726016998291, + "learning_rate": 9.972222222222224e-06, + "loss": 1.0993, + "step": 1230 + }, + { + "epoch": 0.24304194433555468, + "grad_norm": 4.967405319213867, + "learning_rate": 9.962962962962964e-06, + "loss": 1.0835, + "step": 1240 + }, + { + "epoch": 0.24500196001568011, + "grad_norm": 11.337072372436523, + "learning_rate": 9.953703703703704e-06, + "loss": 1.1057, + "step": 1250 + }, + { + "epoch": 0.24696197569580558, + "grad_norm": 5.182584285736084, + "learning_rate": 9.944444444444445e-06, + "loss": 1.0913, + "step": 1260 + }, + { + "epoch": 0.24892199137593102, + "grad_norm": 6.305624961853027, + "learning_rate": 9.935185185185185e-06, + "loss": 1.0963, + "step": 1270 + }, + { + "epoch": 0.25088200705605646, + "grad_norm": 5.263504981994629, + "learning_rate": 9.925925925925927e-06, + "loss": 1.1097, + "step": 1280 + }, + { + "epoch": 0.2528420227361819, + "grad_norm": 5.2210493087768555, + "learning_rate": 9.916666666666668e-06, + "loss": 1.0934, + "step": 1290 + }, + { + "epoch": 0.25480203841630733, + "grad_norm": 6.707633972167969, + "learning_rate": 9.907407407407408e-06, + "loss": 1.1049, + "step": 1300 + }, + { + "epoch": 0.25676205409643277, + "grad_norm": 5.195355415344238, + "learning_rate": 9.898148148148148e-06, + "loss": 1.0866, + "step": 1310 + }, + { + "epoch": 0.2587220697765582, + "grad_norm": 7.137194633483887, + "learning_rate": 9.88888888888889e-06, + "loss": 1.1032, + "step": 1320 + }, + { + "epoch": 0.26068208545668364, + "grad_norm": 5.271111488342285, + "learning_rate": 9.87962962962963e-06, + "loss": 1.0925, + "step": 1330 + }, + { + "epoch": 0.2626421011368091, + "grad_norm": 6.781525135040283, + "learning_rate": 9.870370370370371e-06, + "loss": 1.0966, + "step": 1340 + }, + { + "epoch": 0.2646021168169345, + "grad_norm": 8.270143508911133, + "learning_rate": 9.861111111111112e-06, + "loss": 1.0991, + "step": 1350 + }, + { + "epoch": 0.26656213249705996, + "grad_norm": 6.368267059326172, + "learning_rate": 9.851851851851852e-06, + "loss": 1.1005, + "step": 1360 + }, + { + "epoch": 0.2685221481771854, + "grad_norm": 5.750402927398682, + "learning_rate": 9.842592592592594e-06, + "loss": 1.0907, + "step": 1370 + }, + { + "epoch": 0.27048216385731083, + "grad_norm": 5.706563472747803, + "learning_rate": 9.833333333333333e-06, + "loss": 1.1133, + "step": 1380 + }, + { + "epoch": 0.2724421795374363, + "grad_norm": 5.816508769989014, + "learning_rate": 9.824074074074075e-06, + "loss": 1.1074, + "step": 1390 + }, + { + "epoch": 0.27440219521756176, + "grad_norm": 4.7041449546813965, + "learning_rate": 9.814814814814815e-06, + "loss": 1.0944, + "step": 1400 + }, + { + "epoch": 0.27440219521756176, + "eval_loss": 1.089659571647644, + "eval_runtime": 14.1203, + "eval_samples_per_second": 46.316, + "eval_steps_per_second": 5.807, + "step": 1400 + }, + { + "epoch": 0.2763622108976872, + "grad_norm": 4.5806355476379395, + "learning_rate": 9.805555555555556e-06, + "loss": 1.0868, + "step": 1410 + }, + { + "epoch": 0.27832222657781264, + "grad_norm": 6.664804458618164, + "learning_rate": 9.796296296296298e-06, + "loss": 1.1031, + "step": 1420 + }, + { + "epoch": 0.2802822422579381, + "grad_norm": 5.814957141876221, + "learning_rate": 9.787037037037038e-06, + "loss": 1.0818, + "step": 1430 + }, + { + "epoch": 0.2822422579380635, + "grad_norm": 11.043229103088379, + "learning_rate": 9.777777777777779e-06, + "loss": 1.1147, + "step": 1440 + }, + { + "epoch": 0.28420227361818895, + "grad_norm": 5.907972812652588, + "learning_rate": 9.768518518518519e-06, + "loss": 1.0949, + "step": 1450 + }, + { + "epoch": 0.2861622892983144, + "grad_norm": 8.000860214233398, + "learning_rate": 9.759259259259261e-06, + "loss": 1.107, + "step": 1460 + }, + { + "epoch": 0.2881223049784398, + "grad_norm": 5.279980182647705, + "learning_rate": 9.75e-06, + "loss": 1.0988, + "step": 1470 + }, + { + "epoch": 0.29008232065856526, + "grad_norm": 6.427779674530029, + "learning_rate": 9.740740740740742e-06, + "loss": 1.0969, + "step": 1480 + }, + { + "epoch": 0.2920423363386907, + "grad_norm": 5.733342170715332, + "learning_rate": 9.731481481481482e-06, + "loss": 1.0832, + "step": 1490 + }, + { + "epoch": 0.29400235201881614, + "grad_norm": 5.370787620544434, + "learning_rate": 9.722222222222223e-06, + "loss": 1.0724, + "step": 1500 + }, + { + "epoch": 0.2959623676989416, + "grad_norm": 4.780156135559082, + "learning_rate": 9.712962962962965e-06, + "loss": 1.0911, + "step": 1510 + }, + { + "epoch": 0.297922383379067, + "grad_norm": 5.770766258239746, + "learning_rate": 9.703703703703703e-06, + "loss": 1.0913, + "step": 1520 + }, + { + "epoch": 0.29988239905919245, + "grad_norm": 5.799376487731934, + "learning_rate": 9.694444444444446e-06, + "loss": 1.1072, + "step": 1530 + }, + { + "epoch": 0.3018424147393179, + "grad_norm": 4.571477890014648, + "learning_rate": 9.685185185185186e-06, + "loss": 1.0977, + "step": 1540 + }, + { + "epoch": 0.3038024304194434, + "grad_norm": 6.020542144775391, + "learning_rate": 9.675925925925926e-06, + "loss": 1.1056, + "step": 1550 + }, + { + "epoch": 0.3057624460995688, + "grad_norm": 6.760439395904541, + "learning_rate": 9.666666666666667e-06, + "loss": 1.0874, + "step": 1560 + }, + { + "epoch": 0.30772246177969426, + "grad_norm": 4.900004863739014, + "learning_rate": 9.657407407407409e-06, + "loss": 1.0946, + "step": 1570 + }, + { + "epoch": 0.3096824774598197, + "grad_norm": 4.82609748840332, + "learning_rate": 9.64814814814815e-06, + "loss": 1.1017, + "step": 1580 + }, + { + "epoch": 0.31164249313994513, + "grad_norm": 5.115661144256592, + "learning_rate": 9.63888888888889e-06, + "loss": 1.1073, + "step": 1590 + }, + { + "epoch": 0.31360250882007057, + "grad_norm": 4.987372875213623, + "learning_rate": 9.62962962962963e-06, + "loss": 1.084, + "step": 1600 + }, + { + "epoch": 0.315562524500196, + "grad_norm": 5.010407447814941, + "learning_rate": 9.62037037037037e-06, + "loss": 1.1031, + "step": 1610 + }, + { + "epoch": 0.31752254018032144, + "grad_norm": 4.8331122398376465, + "learning_rate": 9.611111111111112e-06, + "loss": 1.0995, + "step": 1620 + }, + { + "epoch": 0.3194825558604469, + "grad_norm": 5.5287909507751465, + "learning_rate": 9.601851851851853e-06, + "loss": 1.0884, + "step": 1630 + }, + { + "epoch": 0.3214425715405723, + "grad_norm": 7.749701023101807, + "learning_rate": 9.592592592592593e-06, + "loss": 1.0786, + "step": 1640 + }, + { + "epoch": 0.32340258722069776, + "grad_norm": 4.908470153808594, + "learning_rate": 9.583333333333335e-06, + "loss": 1.1049, + "step": 1650 + }, + { + "epoch": 0.3253626029008232, + "grad_norm": 8.760143280029297, + "learning_rate": 9.574074074074074e-06, + "loss": 1.1, + "step": 1660 + }, + { + "epoch": 0.32732261858094863, + "grad_norm": 4.7910590171813965, + "learning_rate": 9.564814814814816e-06, + "loss": 1.0936, + "step": 1670 + }, + { + "epoch": 0.32928263426107407, + "grad_norm": 10.59984302520752, + "learning_rate": 9.555555555555556e-06, + "loss": 1.0888, + "step": 1680 + }, + { + "epoch": 0.3312426499411995, + "grad_norm": 5.18034029006958, + "learning_rate": 9.546296296296297e-06, + "loss": 1.1015, + "step": 1690 + }, + { + "epoch": 0.33320266562132494, + "grad_norm": 4.835277557373047, + "learning_rate": 9.537037037037037e-06, + "loss": 1.0934, + "step": 1700 + }, + { + "epoch": 0.33516268130145044, + "grad_norm": 5.690189838409424, + "learning_rate": 9.527777777777778e-06, + "loss": 1.0981, + "step": 1710 + }, + { + "epoch": 0.3371226969815759, + "grad_norm": 5.126386642456055, + "learning_rate": 9.51851851851852e-06, + "loss": 1.0731, + "step": 1720 + }, + { + "epoch": 0.3390827126617013, + "grad_norm": 5.041740894317627, + "learning_rate": 9.50925925925926e-06, + "loss": 1.0822, + "step": 1730 + }, + { + "epoch": 0.34104272834182675, + "grad_norm": 5.251774787902832, + "learning_rate": 9.5e-06, + "loss": 1.0914, + "step": 1740 + }, + { + "epoch": 0.3430027440219522, + "grad_norm": 4.976339817047119, + "learning_rate": 9.490740740740741e-06, + "loss": 1.0994, + "step": 1750 + }, + { + "epoch": 0.3449627597020776, + "grad_norm": 4.93164587020874, + "learning_rate": 9.481481481481483e-06, + "loss": 1.1029, + "step": 1760 + }, + { + "epoch": 0.34692277538220306, + "grad_norm": 4.964654445648193, + "learning_rate": 9.472222222222223e-06, + "loss": 1.0846, + "step": 1770 + }, + { + "epoch": 0.3488827910623285, + "grad_norm": 5.763545036315918, + "learning_rate": 9.462962962962964e-06, + "loss": 1.0861, + "step": 1780 + }, + { + "epoch": 0.35084280674245394, + "grad_norm": 7.256554126739502, + "learning_rate": 9.453703703703704e-06, + "loss": 1.0878, + "step": 1790 + }, + { + "epoch": 0.3528028224225794, + "grad_norm": 5.118113994598389, + "learning_rate": 9.444444444444445e-06, + "loss": 1.0814, + "step": 1800 + }, + { + "epoch": 0.3547628381027048, + "grad_norm": 5.370246410369873, + "learning_rate": 9.435185185185187e-06, + "loss": 1.0977, + "step": 1810 + }, + { + "epoch": 0.35672285378283025, + "grad_norm": 4.952902793884277, + "learning_rate": 9.425925925925925e-06, + "loss": 1.0889, + "step": 1820 + }, + { + "epoch": 0.3586828694629557, + "grad_norm": 4.842602252960205, + "learning_rate": 9.416666666666667e-06, + "loss": 1.0926, + "step": 1830 + }, + { + "epoch": 0.3606428851430811, + "grad_norm": 4.805694103240967, + "learning_rate": 9.407407407407408e-06, + "loss": 1.1087, + "step": 1840 + }, + { + "epoch": 0.36260290082320656, + "grad_norm": 5.138493061065674, + "learning_rate": 9.398148148148148e-06, + "loss": 1.0862, + "step": 1850 + }, + { + "epoch": 0.364562916503332, + "grad_norm": 7.099350929260254, + "learning_rate": 9.38888888888889e-06, + "loss": 1.0887, + "step": 1860 + }, + { + "epoch": 0.3665229321834575, + "grad_norm": 5.295989990234375, + "learning_rate": 9.37962962962963e-06, + "loss": 1.0715, + "step": 1870 + }, + { + "epoch": 0.36848294786358293, + "grad_norm": 5.222052097320557, + "learning_rate": 9.370370370370371e-06, + "loss": 1.0979, + "step": 1880 + }, + { + "epoch": 0.37044296354370837, + "grad_norm": 4.347951889038086, + "learning_rate": 9.361111111111111e-06, + "loss": 1.1036, + "step": 1890 + }, + { + "epoch": 0.3724029792238338, + "grad_norm": 5.062961101531982, + "learning_rate": 9.351851851851854e-06, + "loss": 1.0864, + "step": 1900 + }, + { + "epoch": 0.37436299490395925, + "grad_norm": 5.308873176574707, + "learning_rate": 9.342592592592594e-06, + "loss": 1.0877, + "step": 1910 + }, + { + "epoch": 0.3763230105840847, + "grad_norm": 4.578522682189941, + "learning_rate": 9.333333333333334e-06, + "loss": 1.0961, + "step": 1920 + }, + { + "epoch": 0.3782830262642101, + "grad_norm": 4.571235179901123, + "learning_rate": 9.324074074074075e-06, + "loss": 1.0838, + "step": 1930 + }, + { + "epoch": 0.38024304194433556, + "grad_norm": 4.522839069366455, + "learning_rate": 9.314814814814815e-06, + "loss": 1.0786, + "step": 1940 + }, + { + "epoch": 0.382203057624461, + "grad_norm": 4.68695068359375, + "learning_rate": 9.305555555555557e-06, + "loss": 1.1146, + "step": 1950 + }, + { + "epoch": 0.38416307330458643, + "grad_norm": 4.976430416107178, + "learning_rate": 9.296296296296296e-06, + "loss": 1.0937, + "step": 1960 + }, + { + "epoch": 0.38612308898471187, + "grad_norm": 5.086645126342773, + "learning_rate": 9.287037037037038e-06, + "loss": 1.097, + "step": 1970 + }, + { + "epoch": 0.3880831046648373, + "grad_norm": 4.946636199951172, + "learning_rate": 9.277777777777778e-06, + "loss": 1.0986, + "step": 1980 + }, + { + "epoch": 0.39004312034496275, + "grad_norm": 4.823138236999512, + "learning_rate": 9.268518518518519e-06, + "loss": 1.1034, + "step": 1990 + }, + { + "epoch": 0.3920031360250882, + "grad_norm": 5.26657772064209, + "learning_rate": 9.25925925925926e-06, + "loss": 1.0798, + "step": 2000 + }, + { + "epoch": 0.3939631517052136, + "grad_norm": 5.231563568115234, + "learning_rate": 9.250000000000001e-06, + "loss": 1.0944, + "step": 2010 + }, + { + "epoch": 0.39592316738533906, + "grad_norm": 4.791154384613037, + "learning_rate": 9.240740740740742e-06, + "loss": 1.0949, + "step": 2020 + }, + { + "epoch": 0.3978831830654645, + "grad_norm": 4.691703796386719, + "learning_rate": 9.231481481481482e-06, + "loss": 1.0856, + "step": 2030 + }, + { + "epoch": 0.39984319874559, + "grad_norm": 7.213362693786621, + "learning_rate": 9.222222222222224e-06, + "loss": 1.0922, + "step": 2040 + }, + { + "epoch": 0.4018032144257154, + "grad_norm": 6.0655646324157715, + "learning_rate": 9.212962962962963e-06, + "loss": 1.0948, + "step": 2050 + }, + { + "epoch": 0.40376323010584086, + "grad_norm": 4.249541759490967, + "learning_rate": 9.203703703703705e-06, + "loss": 1.091, + "step": 2060 + }, + { + "epoch": 0.4057232457859663, + "grad_norm": 4.832751750946045, + "learning_rate": 9.194444444444445e-06, + "loss": 1.0907, + "step": 2070 + }, + { + "epoch": 0.40768326146609174, + "grad_norm": 5.216608047485352, + "learning_rate": 9.185185185185186e-06, + "loss": 1.0802, + "step": 2080 + }, + { + "epoch": 0.4096432771462172, + "grad_norm": 5.988987445831299, + "learning_rate": 9.175925925925928e-06, + "loss": 1.0938, + "step": 2090 + }, + { + "epoch": 0.4116032928263426, + "grad_norm": 4.744123458862305, + "learning_rate": 9.166666666666666e-06, + "loss": 1.0889, + "step": 2100 + }, + { + "epoch": 0.4116032928263426, + "eval_loss": 1.08055579662323, + "eval_runtime": 14.0694, + "eval_samples_per_second": 46.484, + "eval_steps_per_second": 5.828, + "step": 2100 + }, + { + "epoch": 0.41356330850646805, + "grad_norm": 5.0730485916137695, + "learning_rate": 9.157407407407409e-06, + "loss": 1.0956, + "step": 2110 + }, + { + "epoch": 0.4155233241865935, + "grad_norm": 4.156872749328613, + "learning_rate": 9.148148148148149e-06, + "loss": 1.0899, + "step": 2120 + }, + { + "epoch": 0.4174833398667189, + "grad_norm": 5.194650173187256, + "learning_rate": 9.13888888888889e-06, + "loss": 1.0772, + "step": 2130 + }, + { + "epoch": 0.41944335554684437, + "grad_norm": 5.057788848876953, + "learning_rate": 9.12962962962963e-06, + "loss": 1.0886, + "step": 2140 + }, + { + "epoch": 0.4214033712269698, + "grad_norm": 4.849510192871094, + "learning_rate": 9.120370370370372e-06, + "loss": 1.1027, + "step": 2150 + }, + { + "epoch": 0.42336338690709524, + "grad_norm": 4.8577704429626465, + "learning_rate": 9.111111111111112e-06, + "loss": 1.0809, + "step": 2160 + }, + { + "epoch": 0.4253234025872207, + "grad_norm": 4.937686443328857, + "learning_rate": 9.101851851851853e-06, + "loss": 1.0819, + "step": 2170 + }, + { + "epoch": 0.4272834182673461, + "grad_norm": 5.295153617858887, + "learning_rate": 9.092592592592593e-06, + "loss": 1.0734, + "step": 2180 + }, + { + "epoch": 0.42924343394747155, + "grad_norm": 5.876978874206543, + "learning_rate": 9.083333333333333e-06, + "loss": 1.078, + "step": 2190 + }, + { + "epoch": 0.43120344962759705, + "grad_norm": 4.725461006164551, + "learning_rate": 9.074074074074075e-06, + "loss": 1.0787, + "step": 2200 + }, + { + "epoch": 0.4331634653077225, + "grad_norm": 4.425513744354248, + "learning_rate": 9.064814814814816e-06, + "loss": 1.1066, + "step": 2210 + }, + { + "epoch": 0.4351234809878479, + "grad_norm": 4.781891822814941, + "learning_rate": 9.055555555555556e-06, + "loss": 1.083, + "step": 2220 + }, + { + "epoch": 0.43708349666797336, + "grad_norm": 5.292329788208008, + "learning_rate": 9.046296296296298e-06, + "loss": 1.094, + "step": 2230 + }, + { + "epoch": 0.4390435123480988, + "grad_norm": 5.246599197387695, + "learning_rate": 9.037037037037037e-06, + "loss": 1.0903, + "step": 2240 + }, + { + "epoch": 0.44100352802822423, + "grad_norm": 4.701263427734375, + "learning_rate": 9.027777777777779e-06, + "loss": 1.0811, + "step": 2250 + }, + { + "epoch": 0.44296354370834967, + "grad_norm": 5.434783458709717, + "learning_rate": 9.01851851851852e-06, + "loss": 1.0862, + "step": 2260 + }, + { + "epoch": 0.4449235593884751, + "grad_norm": 5.898828506469727, + "learning_rate": 9.00925925925926e-06, + "loss": 1.0972, + "step": 2270 + }, + { + "epoch": 0.44688357506860055, + "grad_norm": 5.089272975921631, + "learning_rate": 9e-06, + "loss": 1.091, + "step": 2280 + }, + { + "epoch": 0.448843590748726, + "grad_norm": 6.887242317199707, + "learning_rate": 8.99074074074074e-06, + "loss": 1.0785, + "step": 2290 + }, + { + "epoch": 0.4508036064288514, + "grad_norm": 5.111120700836182, + "learning_rate": 8.981481481481483e-06, + "loss": 1.0682, + "step": 2300 + }, + { + "epoch": 0.45276362210897686, + "grad_norm": 5.331275939941406, + "learning_rate": 8.972222222222223e-06, + "loss": 1.0854, + "step": 2310 + }, + { + "epoch": 0.4547236377891023, + "grad_norm": 5.59519100189209, + "learning_rate": 8.962962962962963e-06, + "loss": 1.0883, + "step": 2320 + }, + { + "epoch": 0.45668365346922773, + "grad_norm": 5.6016364097595215, + "learning_rate": 8.953703703703704e-06, + "loss": 1.0799, + "step": 2330 + }, + { + "epoch": 0.4586436691493532, + "grad_norm": 4.625690460205078, + "learning_rate": 8.944444444444446e-06, + "loss": 1.0716, + "step": 2340 + }, + { + "epoch": 0.4606036848294786, + "grad_norm": 4.6536688804626465, + "learning_rate": 8.935185185185186e-06, + "loss": 1.0624, + "step": 2350 + }, + { + "epoch": 0.4625637005096041, + "grad_norm": 4.807366847991943, + "learning_rate": 8.925925925925927e-06, + "loss": 1.0829, + "step": 2360 + }, + { + "epoch": 0.46452371618972954, + "grad_norm": 4.950624942779541, + "learning_rate": 8.916666666666667e-06, + "loss": 1.0861, + "step": 2370 + }, + { + "epoch": 0.466483731869855, + "grad_norm": 4.980051517486572, + "learning_rate": 8.907407407407408e-06, + "loss": 1.0772, + "step": 2380 + }, + { + "epoch": 0.4684437475499804, + "grad_norm": 5.208264350891113, + "learning_rate": 8.89814814814815e-06, + "loss": 1.0929, + "step": 2390 + }, + { + "epoch": 0.47040376323010585, + "grad_norm": 5.6281657218933105, + "learning_rate": 8.888888888888888e-06, + "loss": 1.0818, + "step": 2400 + }, + { + "epoch": 0.4723637789102313, + "grad_norm": 6.125396251678467, + "learning_rate": 8.87962962962963e-06, + "loss": 1.0865, + "step": 2410 + }, + { + "epoch": 0.47432379459035673, + "grad_norm": 5.989163398742676, + "learning_rate": 8.87037037037037e-06, + "loss": 1.0685, + "step": 2420 + }, + { + "epoch": 0.47628381027048217, + "grad_norm": 9.247689247131348, + "learning_rate": 8.861111111111111e-06, + "loss": 1.0879, + "step": 2430 + }, + { + "epoch": 0.4782438259506076, + "grad_norm": 4.848996639251709, + "learning_rate": 8.851851851851853e-06, + "loss": 1.0783, + "step": 2440 + }, + { + "epoch": 0.48020384163073304, + "grad_norm": 5.04819917678833, + "learning_rate": 8.842592592592594e-06, + "loss": 1.0679, + "step": 2450 + }, + { + "epoch": 0.4821638573108585, + "grad_norm": 5.755705833435059, + "learning_rate": 8.833333333333334e-06, + "loss": 1.0939, + "step": 2460 + }, + { + "epoch": 0.4841238729909839, + "grad_norm": 4.5991058349609375, + "learning_rate": 8.824074074074074e-06, + "loss": 1.0775, + "step": 2470 + }, + { + "epoch": 0.48608388867110935, + "grad_norm": 5.101771354675293, + "learning_rate": 8.814814814814817e-06, + "loss": 1.084, + "step": 2480 + }, + { + "epoch": 0.4880439043512348, + "grad_norm": 4.9321441650390625, + "learning_rate": 8.805555555555557e-06, + "loss": 1.0762, + "step": 2490 + }, + { + "epoch": 0.49000392003136023, + "grad_norm": 5.363485813140869, + "learning_rate": 8.796296296296297e-06, + "loss": 1.0595, + "step": 2500 + }, + { + "epoch": 0.49196393571148567, + "grad_norm": 5.353972434997559, + "learning_rate": 8.787037037037038e-06, + "loss": 1.0689, + "step": 2510 + }, + { + "epoch": 0.49392395139161116, + "grad_norm": 5.158483505249023, + "learning_rate": 8.777777777777778e-06, + "loss": 1.0711, + "step": 2520 + }, + { + "epoch": 0.4958839670717366, + "grad_norm": 5.160126686096191, + "learning_rate": 8.76851851851852e-06, + "loss": 1.0886, + "step": 2530 + }, + { + "epoch": 0.49784398275186204, + "grad_norm": 4.926945686340332, + "learning_rate": 8.759259259259259e-06, + "loss": 1.0856, + "step": 2540 + }, + { + "epoch": 0.4998039984319875, + "grad_norm": 5.253294467926025, + "learning_rate": 8.750000000000001e-06, + "loss": 1.0876, + "step": 2550 + }, + { + "epoch": 0.5017640141121129, + "grad_norm": 11.674654960632324, + "learning_rate": 8.740740740740741e-06, + "loss": 1.0784, + "step": 2560 + }, + { + "epoch": 0.5037240297922383, + "grad_norm": 5.032644271850586, + "learning_rate": 8.731481481481482e-06, + "loss": 1.07, + "step": 2570 + }, + { + "epoch": 0.5056840454723638, + "grad_norm": 5.1500372886657715, + "learning_rate": 8.722222222222224e-06, + "loss": 1.0932, + "step": 2580 + }, + { + "epoch": 0.5076440611524892, + "grad_norm": 5.761635780334473, + "learning_rate": 8.712962962962964e-06, + "loss": 1.0816, + "step": 2590 + }, + { + "epoch": 0.5096040768326147, + "grad_norm": 5.257477760314941, + "learning_rate": 8.703703703703705e-06, + "loss": 1.0712, + "step": 2600 + }, + { + "epoch": 0.5115640925127402, + "grad_norm": 5.4931416511535645, + "learning_rate": 8.694444444444445e-06, + "loss": 1.0832, + "step": 2610 + }, + { + "epoch": 0.5135241081928655, + "grad_norm": 6.102149963378906, + "learning_rate": 8.685185185185185e-06, + "loss": 1.076, + "step": 2620 + }, + { + "epoch": 0.515484123872991, + "grad_norm": 5.596585273742676, + "learning_rate": 8.675925925925926e-06, + "loss": 1.0972, + "step": 2630 + }, + { + "epoch": 0.5174441395531164, + "grad_norm": 4.6653523445129395, + "learning_rate": 8.666666666666668e-06, + "loss": 1.0872, + "step": 2640 + }, + { + "epoch": 0.5194041552332419, + "grad_norm": 4.938266277313232, + "learning_rate": 8.657407407407408e-06, + "loss": 1.0717, + "step": 2650 + }, + { + "epoch": 0.5213641709133673, + "grad_norm": 5.11916446685791, + "learning_rate": 8.648148148148149e-06, + "loss": 1.0955, + "step": 2660 + }, + { + "epoch": 0.5233241865934928, + "grad_norm": 4.867392539978027, + "learning_rate": 8.63888888888889e-06, + "loss": 1.0769, + "step": 2670 + }, + { + "epoch": 0.5252842022736182, + "grad_norm": 5.589994430541992, + "learning_rate": 8.62962962962963e-06, + "loss": 1.0808, + "step": 2680 + }, + { + "epoch": 0.5272442179537437, + "grad_norm": 5.221999645233154, + "learning_rate": 8.620370370370371e-06, + "loss": 1.0627, + "step": 2690 + }, + { + "epoch": 0.529204233633869, + "grad_norm": 4.844895839691162, + "learning_rate": 8.611111111111112e-06, + "loss": 1.0951, + "step": 2700 + }, + { + "epoch": 0.5311642493139945, + "grad_norm": 4.937107563018799, + "learning_rate": 8.601851851851852e-06, + "loss": 1.0898, + "step": 2710 + }, + { + "epoch": 0.5331242649941199, + "grad_norm": 4.758755207061768, + "learning_rate": 8.592592592592593e-06, + "loss": 1.1017, + "step": 2720 + }, + { + "epoch": 0.5350842806742454, + "grad_norm": 5.145150661468506, + "learning_rate": 8.583333333333333e-06, + "loss": 1.0702, + "step": 2730 + }, + { + "epoch": 0.5370442963543708, + "grad_norm": 5.257258415222168, + "learning_rate": 8.574074074074075e-06, + "loss": 1.0822, + "step": 2740 + }, + { + "epoch": 0.5390043120344963, + "grad_norm": 5.826773166656494, + "learning_rate": 8.564814814814816e-06, + "loss": 1.0714, + "step": 2750 + }, + { + "epoch": 0.5409643277146217, + "grad_norm": 4.5895304679870605, + "learning_rate": 8.555555555555556e-06, + "loss": 1.0866, + "step": 2760 + }, + { + "epoch": 0.5429243433947472, + "grad_norm": 6.960358619689941, + "learning_rate": 8.546296296296296e-06, + "loss": 1.0942, + "step": 2770 + }, + { + "epoch": 0.5448843590748726, + "grad_norm": 5.870516300201416, + "learning_rate": 8.537037037037038e-06, + "loss": 1.0937, + "step": 2780 + }, + { + "epoch": 0.546844374754998, + "grad_norm": 4.99727725982666, + "learning_rate": 8.527777777777779e-06, + "loss": 1.0851, + "step": 2790 + }, + { + "epoch": 0.5488043904351235, + "grad_norm": 4.9571003913879395, + "learning_rate": 8.518518518518519e-06, + "loss": 1.085, + "step": 2800 + }, + { + "epoch": 0.5488043904351235, + "eval_loss": 1.073391318321228, + "eval_runtime": 14.0283, + "eval_samples_per_second": 46.62, + "eval_steps_per_second": 5.845, + "step": 2800 + }, + { + "epoch": 0.5507644061152489, + "grad_norm": 4.457803726196289, + "learning_rate": 8.509259259259261e-06, + "loss": 1.0787, + "step": 2810 + }, + { + "epoch": 0.5527244217953744, + "grad_norm": 5.098880290985107, + "learning_rate": 8.5e-06, + "loss": 1.0672, + "step": 2820 + }, + { + "epoch": 0.5546844374754998, + "grad_norm": 5.3498735427856445, + "learning_rate": 8.490740740740742e-06, + "loss": 1.0852, + "step": 2830 + }, + { + "epoch": 0.5566444531556253, + "grad_norm": 5.1486921310424805, + "learning_rate": 8.481481481481482e-06, + "loss": 1.0564, + "step": 2840 + }, + { + "epoch": 0.5586044688357507, + "grad_norm": 5.613958358764648, + "learning_rate": 8.472222222222223e-06, + "loss": 1.0864, + "step": 2850 + }, + { + "epoch": 0.5605644845158761, + "grad_norm": 5.687408924102783, + "learning_rate": 8.462962962962963e-06, + "loss": 1.0871, + "step": 2860 + }, + { + "epoch": 0.5625245001960015, + "grad_norm": 5.822880268096924, + "learning_rate": 8.453703703703704e-06, + "loss": 1.0708, + "step": 2870 + }, + { + "epoch": 0.564484515876127, + "grad_norm": 5.982561111450195, + "learning_rate": 8.444444444444446e-06, + "loss": 1.0633, + "step": 2880 + }, + { + "epoch": 0.5664445315562524, + "grad_norm": 4.887526035308838, + "learning_rate": 8.435185185185186e-06, + "loss": 1.0823, + "step": 2890 + }, + { + "epoch": 0.5684045472363779, + "grad_norm": 5.810245037078857, + "learning_rate": 8.425925925925926e-06, + "loss": 1.0683, + "step": 2900 + }, + { + "epoch": 0.5703645629165033, + "grad_norm": 4.757987022399902, + "learning_rate": 8.416666666666667e-06, + "loss": 1.0703, + "step": 2910 + }, + { + "epoch": 0.5723245785966288, + "grad_norm": 5.597621440887451, + "learning_rate": 8.407407407407409e-06, + "loss": 1.0861, + "step": 2920 + }, + { + "epoch": 0.5742845942767543, + "grad_norm": 9.786075592041016, + "learning_rate": 8.39814814814815e-06, + "loss": 1.0755, + "step": 2930 + }, + { + "epoch": 0.5762446099568796, + "grad_norm": 4.969895362854004, + "learning_rate": 8.38888888888889e-06, + "loss": 1.0907, + "step": 2940 + }, + { + "epoch": 0.5782046256370051, + "grad_norm": 5.833816051483154, + "learning_rate": 8.37962962962963e-06, + "loss": 1.0679, + "step": 2950 + }, + { + "epoch": 0.5801646413171305, + "grad_norm": 5.07546329498291, + "learning_rate": 8.37037037037037e-06, + "loss": 1.0705, + "step": 2960 + }, + { + "epoch": 0.582124656997256, + "grad_norm": 5.735626220703125, + "learning_rate": 8.361111111111113e-06, + "loss": 1.0676, + "step": 2970 + }, + { + "epoch": 0.5840846726773814, + "grad_norm": 4.732352256774902, + "learning_rate": 8.351851851851851e-06, + "loss": 1.0809, + "step": 2980 + }, + { + "epoch": 0.5860446883575069, + "grad_norm": 6.006665229797363, + "learning_rate": 8.342592592592593e-06, + "loss": 1.0624, + "step": 2990 + }, + { + "epoch": 0.5880047040376323, + "grad_norm": 6.649717330932617, + "learning_rate": 8.333333333333334e-06, + "loss": 1.072, + "step": 3000 + }, + { + "epoch": 0.5899647197177578, + "grad_norm": 5.156081199645996, + "learning_rate": 8.324074074074074e-06, + "loss": 1.0791, + "step": 3010 + }, + { + "epoch": 0.5919247353978832, + "grad_norm": 5.489958763122559, + "learning_rate": 8.314814814814816e-06, + "loss": 1.0803, + "step": 3020 + }, + { + "epoch": 0.5938847510780086, + "grad_norm": 5.912964344024658, + "learning_rate": 8.305555555555557e-06, + "loss": 1.0888, + "step": 3030 + }, + { + "epoch": 0.595844766758134, + "grad_norm": 5.113432884216309, + "learning_rate": 8.296296296296297e-06, + "loss": 1.0728, + "step": 3040 + }, + { + "epoch": 0.5978047824382595, + "grad_norm": 4.774835109710693, + "learning_rate": 8.287037037037037e-06, + "loss": 1.0772, + "step": 3050 + }, + { + "epoch": 0.5997647981183849, + "grad_norm": 5.954660892486572, + "learning_rate": 8.277777777777778e-06, + "loss": 1.0805, + "step": 3060 + }, + { + "epoch": 0.6017248137985104, + "grad_norm": 4.933722019195557, + "learning_rate": 8.26851851851852e-06, + "loss": 1.0678, + "step": 3070 + }, + { + "epoch": 0.6036848294786358, + "grad_norm": 5.152841567993164, + "learning_rate": 8.25925925925926e-06, + "loss": 1.0831, + "step": 3080 + }, + { + "epoch": 0.6056448451587613, + "grad_norm": 5.011537551879883, + "learning_rate": 8.25e-06, + "loss": 1.0841, + "step": 3090 + }, + { + "epoch": 0.6076048608388868, + "grad_norm": 4.196855545043945, + "learning_rate": 8.240740740740741e-06, + "loss": 1.0647, + "step": 3100 + }, + { + "epoch": 0.6095648765190121, + "grad_norm": 5.21035099029541, + "learning_rate": 8.231481481481483e-06, + "loss": 1.049, + "step": 3110 + }, + { + "epoch": 0.6115248921991376, + "grad_norm": 26.132904052734375, + "learning_rate": 8.222222222222222e-06, + "loss": 1.0799, + "step": 3120 + }, + { + "epoch": 0.613484907879263, + "grad_norm": 6.4696760177612305, + "learning_rate": 8.212962962962964e-06, + "loss": 1.0666, + "step": 3130 + }, + { + "epoch": 0.6154449235593885, + "grad_norm": 4.687003135681152, + "learning_rate": 8.203703703703704e-06, + "loss": 1.0732, + "step": 3140 + }, + { + "epoch": 0.6174049392395139, + "grad_norm": 5.188929557800293, + "learning_rate": 8.194444444444445e-06, + "loss": 1.0911, + "step": 3150 + }, + { + "epoch": 0.6193649549196394, + "grad_norm": 4.781430244445801, + "learning_rate": 8.185185185185187e-06, + "loss": 1.0705, + "step": 3160 + }, + { + "epoch": 0.6213249705997648, + "grad_norm": 4.717843055725098, + "learning_rate": 8.175925925925925e-06, + "loss": 1.0822, + "step": 3170 + }, + { + "epoch": 0.6232849862798903, + "grad_norm": 5.404903888702393, + "learning_rate": 8.166666666666668e-06, + "loss": 1.0917, + "step": 3180 + }, + { + "epoch": 0.6252450019600156, + "grad_norm": 5.684728145599365, + "learning_rate": 8.157407407407408e-06, + "loss": 1.0842, + "step": 3190 + }, + { + "epoch": 0.6272050176401411, + "grad_norm": 4.7895965576171875, + "learning_rate": 8.148148148148148e-06, + "loss": 1.0652, + "step": 3200 + }, + { + "epoch": 0.6291650333202665, + "grad_norm": 6.572015285491943, + "learning_rate": 8.138888888888889e-06, + "loss": 1.0959, + "step": 3210 + }, + { + "epoch": 0.631125049000392, + "grad_norm": 4.640923976898193, + "learning_rate": 8.12962962962963e-06, + "loss": 1.0896, + "step": 3220 + }, + { + "epoch": 0.6330850646805174, + "grad_norm": 4.732535362243652, + "learning_rate": 8.120370370370371e-06, + "loss": 1.0646, + "step": 3230 + }, + { + "epoch": 0.6350450803606429, + "grad_norm": 5.692999362945557, + "learning_rate": 8.111111111111112e-06, + "loss": 1.0774, + "step": 3240 + }, + { + "epoch": 0.6370050960407684, + "grad_norm": 5.28448486328125, + "learning_rate": 8.101851851851854e-06, + "loss": 1.0844, + "step": 3250 + }, + { + "epoch": 0.6389651117208938, + "grad_norm": 3.9686877727508545, + "learning_rate": 8.092592592592592e-06, + "loss": 1.0626, + "step": 3260 + }, + { + "epoch": 0.6409251274010193, + "grad_norm": 6.364278793334961, + "learning_rate": 8.083333333333334e-06, + "loss": 1.0696, + "step": 3270 + }, + { + "epoch": 0.6428851430811446, + "grad_norm": 4.789191722869873, + "learning_rate": 8.074074074074075e-06, + "loss": 1.0696, + "step": 3280 + }, + { + "epoch": 0.6448451587612701, + "grad_norm": 4.61089563369751, + "learning_rate": 8.064814814814815e-06, + "loss": 1.0701, + "step": 3290 + }, + { + "epoch": 0.6468051744413955, + "grad_norm": 4.959714889526367, + "learning_rate": 8.055555555555557e-06, + "loss": 1.0852, + "step": 3300 + }, + { + "epoch": 0.648765190121521, + "grad_norm": 4.559225082397461, + "learning_rate": 8.046296296296296e-06, + "loss": 1.0708, + "step": 3310 + }, + { + "epoch": 0.6507252058016464, + "grad_norm": 4.9751691818237305, + "learning_rate": 8.037037037037038e-06, + "loss": 1.0545, + "step": 3320 + }, + { + "epoch": 0.6526852214817719, + "grad_norm": 5.153299808502197, + "learning_rate": 8.027777777777778e-06, + "loss": 1.074, + "step": 3330 + }, + { + "epoch": 0.6546452371618973, + "grad_norm": 5.507195472717285, + "learning_rate": 8.018518518518519e-06, + "loss": 1.0772, + "step": 3340 + }, + { + "epoch": 0.6566052528420228, + "grad_norm": 4.8530449867248535, + "learning_rate": 8.00925925925926e-06, + "loss": 1.087, + "step": 3350 + }, + { + "epoch": 0.6585652685221481, + "grad_norm": 4.754231929779053, + "learning_rate": 8.000000000000001e-06, + "loss": 1.0792, + "step": 3360 + }, + { + "epoch": 0.6605252842022736, + "grad_norm": 4.642824172973633, + "learning_rate": 7.990740740740742e-06, + "loss": 1.0681, + "step": 3370 + }, + { + "epoch": 0.662485299882399, + "grad_norm": 5.338252067565918, + "learning_rate": 7.981481481481482e-06, + "loss": 1.0579, + "step": 3380 + }, + { + "epoch": 0.6644453155625245, + "grad_norm": 5.290416240692139, + "learning_rate": 7.972222222222224e-06, + "loss": 1.0523, + "step": 3390 + }, + { + "epoch": 0.6664053312426499, + "grad_norm": 4.5010294914245605, + "learning_rate": 7.962962962962963e-06, + "loss": 1.0694, + "step": 3400 + }, + { + "epoch": 0.6683653469227754, + "grad_norm": 5.253542900085449, + "learning_rate": 7.953703703703705e-06, + "loss": 1.0745, + "step": 3410 + }, + { + "epoch": 0.6703253626029009, + "grad_norm": 4.743661880493164, + "learning_rate": 7.944444444444445e-06, + "loss": 1.0586, + "step": 3420 + }, + { + "epoch": 0.6722853782830263, + "grad_norm": 5.1141839027404785, + "learning_rate": 7.935185185185186e-06, + "loss": 1.0767, + "step": 3430 + }, + { + "epoch": 0.6742453939631518, + "grad_norm": 4.729719638824463, + "learning_rate": 7.925925925925926e-06, + "loss": 1.0841, + "step": 3440 + }, + { + "epoch": 0.6762054096432771, + "grad_norm": 5.445465087890625, + "learning_rate": 7.916666666666667e-06, + "loss": 1.0901, + "step": 3450 + }, + { + "epoch": 0.6781654253234026, + "grad_norm": 4.362582683563232, + "learning_rate": 7.907407407407409e-06, + "loss": 1.0634, + "step": 3460 + }, + { + "epoch": 0.680125441003528, + "grad_norm": 4.604685306549072, + "learning_rate": 7.898148148148149e-06, + "loss": 1.0631, + "step": 3470 + }, + { + "epoch": 0.6820854566836535, + "grad_norm": 5.426617622375488, + "learning_rate": 7.88888888888889e-06, + "loss": 1.0691, + "step": 3480 + }, + { + "epoch": 0.6840454723637789, + "grad_norm": 4.578293323516846, + "learning_rate": 7.87962962962963e-06, + "loss": 1.0761, + "step": 3490 + }, + { + "epoch": 0.6860054880439044, + "grad_norm": 5.362606525421143, + "learning_rate": 7.870370370370372e-06, + "loss": 1.0836, + "step": 3500 + }, + { + "epoch": 0.6860054880439044, + "eval_loss": 1.0641947984695435, + "eval_runtime": 14.0879, + "eval_samples_per_second": 46.423, + "eval_steps_per_second": 5.821, + "step": 3500 + }, + { + "epoch": 0.6879655037240298, + "grad_norm": 4.338893890380859, + "learning_rate": 7.861111111111112e-06, + "loss": 1.0604, + "step": 3510 + }, + { + "epoch": 0.6899255194041553, + "grad_norm": 5.606161594390869, + "learning_rate": 7.851851851851853e-06, + "loss": 1.0869, + "step": 3520 + }, + { + "epoch": 0.6918855350842806, + "grad_norm": 4.934096336364746, + "learning_rate": 7.842592592592593e-06, + "loss": 1.07, + "step": 3530 + }, + { + "epoch": 0.6938455507644061, + "grad_norm": 4.351364612579346, + "learning_rate": 7.833333333333333e-06, + "loss": 1.0682, + "step": 3540 + }, + { + "epoch": 0.6958055664445315, + "grad_norm": 4.8932976722717285, + "learning_rate": 7.824074074074076e-06, + "loss": 1.0766, + "step": 3550 + }, + { + "epoch": 0.697765582124657, + "grad_norm": 4.501565456390381, + "learning_rate": 7.814814814814816e-06, + "loss": 1.0612, + "step": 3560 + }, + { + "epoch": 0.6997255978047824, + "grad_norm": 4.929937362670898, + "learning_rate": 7.805555555555556e-06, + "loss": 1.0609, + "step": 3570 + }, + { + "epoch": 0.7016856134849079, + "grad_norm": 5.501171112060547, + "learning_rate": 7.796296296296297e-06, + "loss": 1.0583, + "step": 3580 + }, + { + "epoch": 0.7036456291650334, + "grad_norm": 4.9919891357421875, + "learning_rate": 7.787037037037037e-06, + "loss": 1.0603, + "step": 3590 + }, + { + "epoch": 0.7056056448451588, + "grad_norm": 4.892634868621826, + "learning_rate": 7.77777777777778e-06, + "loss": 1.0718, + "step": 3600 + }, + { + "epoch": 0.7075656605252842, + "grad_norm": 5.0666632652282715, + "learning_rate": 7.76851851851852e-06, + "loss": 1.0733, + "step": 3610 + }, + { + "epoch": 0.7095256762054096, + "grad_norm": 4.797816753387451, + "learning_rate": 7.75925925925926e-06, + "loss": 1.0664, + "step": 3620 + }, + { + "epoch": 0.7114856918855351, + "grad_norm": 5.004721164703369, + "learning_rate": 7.75e-06, + "loss": 1.0847, + "step": 3630 + }, + { + "epoch": 0.7134457075656605, + "grad_norm": 4.80883264541626, + "learning_rate": 7.74074074074074e-06, + "loss": 1.075, + "step": 3640 + }, + { + "epoch": 0.715405723245786, + "grad_norm": 5.439119815826416, + "learning_rate": 7.731481481481483e-06, + "loss": 1.0517, + "step": 3650 + }, + { + "epoch": 0.7173657389259114, + "grad_norm": 4.244631290435791, + "learning_rate": 7.722222222222223e-06, + "loss": 1.0836, + "step": 3660 + }, + { + "epoch": 0.7193257546060369, + "grad_norm": 4.710737705230713, + "learning_rate": 7.712962962962964e-06, + "loss": 1.0748, + "step": 3670 + }, + { + "epoch": 0.7212857702861623, + "grad_norm": 5.23133659362793, + "learning_rate": 7.703703703703704e-06, + "loss": 1.0843, + "step": 3680 + }, + { + "epoch": 0.7232457859662877, + "grad_norm": 4.706710338592529, + "learning_rate": 7.694444444444446e-06, + "loss": 1.0637, + "step": 3690 + }, + { + "epoch": 0.7252058016464131, + "grad_norm": 4.4408955574035645, + "learning_rate": 7.685185185185185e-06, + "loss": 1.0945, + "step": 3700 + }, + { + "epoch": 0.7271658173265386, + "grad_norm": 5.1011199951171875, + "learning_rate": 7.675925925925927e-06, + "loss": 1.0573, + "step": 3710 + }, + { + "epoch": 0.729125833006664, + "grad_norm": 4.85494327545166, + "learning_rate": 7.666666666666667e-06, + "loss": 1.0686, + "step": 3720 + }, + { + "epoch": 0.7310858486867895, + "grad_norm": 4.990062236785889, + "learning_rate": 7.657407407407408e-06, + "loss": 1.052, + "step": 3730 + }, + { + "epoch": 0.733045864366915, + "grad_norm": 4.779956817626953, + "learning_rate": 7.64814814814815e-06, + "loss": 1.0619, + "step": 3740 + }, + { + "epoch": 0.7350058800470404, + "grad_norm": 4.921076774597168, + "learning_rate": 7.638888888888888e-06, + "loss": 1.0834, + "step": 3750 + }, + { + "epoch": 0.7369658957271659, + "grad_norm": 4.729534149169922, + "learning_rate": 7.62962962962963e-06, + "loss": 1.0643, + "step": 3760 + }, + { + "epoch": 0.7389259114072912, + "grad_norm": 4.321104049682617, + "learning_rate": 7.620370370370372e-06, + "loss": 1.0736, + "step": 3770 + }, + { + "epoch": 0.7408859270874167, + "grad_norm": 5.252387523651123, + "learning_rate": 7.611111111111111e-06, + "loss": 1.0515, + "step": 3780 + }, + { + "epoch": 0.7428459427675421, + "grad_norm": 4.770890712738037, + "learning_rate": 7.6018518518518525e-06, + "loss": 1.0865, + "step": 3790 + }, + { + "epoch": 0.7448059584476676, + "grad_norm": 4.786081790924072, + "learning_rate": 7.592592592592594e-06, + "loss": 1.0727, + "step": 3800 + }, + { + "epoch": 0.746765974127793, + "grad_norm": 5.215248107910156, + "learning_rate": 7.583333333333333e-06, + "loss": 1.0657, + "step": 3810 + }, + { + "epoch": 0.7487259898079185, + "grad_norm": 8.386008262634277, + "learning_rate": 7.5740740740740745e-06, + "loss": 1.0774, + "step": 3820 + }, + { + "epoch": 0.7506860054880439, + "grad_norm": 5.009541034698486, + "learning_rate": 7.564814814814816e-06, + "loss": 1.0517, + "step": 3830 + }, + { + "epoch": 0.7526460211681694, + "grad_norm": 5.512828826904297, + "learning_rate": 7.555555555555556e-06, + "loss": 1.0661, + "step": 3840 + }, + { + "epoch": 0.7546060368482947, + "grad_norm": 4.983898162841797, + "learning_rate": 7.546296296296297e-06, + "loss": 1.0561, + "step": 3850 + }, + { + "epoch": 0.7565660525284202, + "grad_norm": 4.687797546386719, + "learning_rate": 7.537037037037037e-06, + "loss": 1.0623, + "step": 3860 + }, + { + "epoch": 0.7585260682085456, + "grad_norm": 4.924409866333008, + "learning_rate": 7.527777777777778e-06, + "loss": 1.0557, + "step": 3870 + }, + { + "epoch": 0.7604860838886711, + "grad_norm": 4.704287528991699, + "learning_rate": 7.518518518518519e-06, + "loss": 1.0732, + "step": 3880 + }, + { + "epoch": 0.7624460995687965, + "grad_norm": 5.610473155975342, + "learning_rate": 7.50925925925926e-06, + "loss": 1.0827, + "step": 3890 + }, + { + "epoch": 0.764406115248922, + "grad_norm": 5.7728590965271, + "learning_rate": 7.500000000000001e-06, + "loss": 1.0738, + "step": 3900 + }, + { + "epoch": 0.7663661309290475, + "grad_norm": 4.7245659828186035, + "learning_rate": 7.4907407407407414e-06, + "loss": 1.0665, + "step": 3910 + }, + { + "epoch": 0.7683261466091729, + "grad_norm": 6.519947528839111, + "learning_rate": 7.481481481481482e-06, + "loss": 1.0631, + "step": 3920 + }, + { + "epoch": 0.7702861622892984, + "grad_norm": 4.989842891693115, + "learning_rate": 7.472222222222223e-06, + "loss": 1.053, + "step": 3930 + }, + { + "epoch": 0.7722461779694237, + "grad_norm": 4.687224864959717, + "learning_rate": 7.462962962962964e-06, + "loss": 1.0593, + "step": 3940 + }, + { + "epoch": 0.7742061936495492, + "grad_norm": 4.693930625915527, + "learning_rate": 7.453703703703704e-06, + "loss": 1.0665, + "step": 3950 + }, + { + "epoch": 0.7761662093296746, + "grad_norm": 5.265392303466797, + "learning_rate": 7.444444444444445e-06, + "loss": 1.042, + "step": 3960 + }, + { + "epoch": 0.7781262250098001, + "grad_norm": 5.122781753540039, + "learning_rate": 7.4351851851851855e-06, + "loss": 1.0748, + "step": 3970 + }, + { + "epoch": 0.7800862406899255, + "grad_norm": 4.934235095977783, + "learning_rate": 7.425925925925927e-06, + "loss": 1.0603, + "step": 3980 + }, + { + "epoch": 0.782046256370051, + "grad_norm": 4.4583587646484375, + "learning_rate": 7.416666666666668e-06, + "loss": 1.0724, + "step": 3990 + }, + { + "epoch": 0.7840062720501764, + "grad_norm": 5.3386030197143555, + "learning_rate": 7.4074074074074075e-06, + "loss": 1.0651, + "step": 4000 + }, + { + "epoch": 0.7859662877303019, + "grad_norm": 4.659664154052734, + "learning_rate": 7.398148148148149e-06, + "loss": 1.0525, + "step": 4010 + }, + { + "epoch": 0.7879263034104272, + "grad_norm": 4.794917106628418, + "learning_rate": 7.38888888888889e-06, + "loss": 1.0561, + "step": 4020 + }, + { + "epoch": 0.7898863190905527, + "grad_norm": 5.210708141326904, + "learning_rate": 7.3796296296296295e-06, + "loss": 1.0641, + "step": 4030 + }, + { + "epoch": 0.7918463347706781, + "grad_norm": 4.441596031188965, + "learning_rate": 7.370370370370371e-06, + "loss": 1.0696, + "step": 4040 + }, + { + "epoch": 0.7938063504508036, + "grad_norm": 5.239253044128418, + "learning_rate": 7.361111111111112e-06, + "loss": 1.0576, + "step": 4050 + }, + { + "epoch": 0.795766366130929, + "grad_norm": 5.195384502410889, + "learning_rate": 7.351851851851852e-06, + "loss": 1.0556, + "step": 4060 + }, + { + "epoch": 0.7977263818110545, + "grad_norm": 5.0446648597717285, + "learning_rate": 7.342592592592594e-06, + "loss": 1.067, + "step": 4070 + }, + { + "epoch": 0.79968639749118, + "grad_norm": 5.061046600341797, + "learning_rate": 7.333333333333333e-06, + "loss": 1.0587, + "step": 4080 + }, + { + "epoch": 0.8016464131713054, + "grad_norm": 7.661867141723633, + "learning_rate": 7.324074074074074e-06, + "loss": 1.0694, + "step": 4090 + }, + { + "epoch": 0.8036064288514309, + "grad_norm": 4.821944236755371, + "learning_rate": 7.314814814814816e-06, + "loss": 1.0675, + "step": 4100 + }, + { + "epoch": 0.8055664445315562, + "grad_norm": 4.810634613037109, + "learning_rate": 7.305555555555556e-06, + "loss": 1.0589, + "step": 4110 + }, + { + "epoch": 0.8075264602116817, + "grad_norm": 5.0643157958984375, + "learning_rate": 7.296296296296297e-06, + "loss": 1.0714, + "step": 4120 + }, + { + "epoch": 0.8094864758918071, + "grad_norm": 4.812649250030518, + "learning_rate": 7.287037037037038e-06, + "loss": 1.0635, + "step": 4130 + }, + { + "epoch": 0.8114464915719326, + "grad_norm": 4.7638840675354, + "learning_rate": 7.277777777777778e-06, + "loss": 1.064, + "step": 4140 + }, + { + "epoch": 0.813406507252058, + "grad_norm": 4.386782646179199, + "learning_rate": 7.268518518518519e-06, + "loss": 1.0726, + "step": 4150 + }, + { + "epoch": 0.8153665229321835, + "grad_norm": 4.593876838684082, + "learning_rate": 7.2592592592592605e-06, + "loss": 1.0542, + "step": 4160 + }, + { + "epoch": 0.8173265386123089, + "grad_norm": 4.822241306304932, + "learning_rate": 7.25e-06, + "loss": 1.0541, + "step": 4170 + }, + { + "epoch": 0.8192865542924344, + "grad_norm": 4.806110858917236, + "learning_rate": 7.240740740740741e-06, + "loss": 1.0522, + "step": 4180 + }, + { + "epoch": 0.8212465699725597, + "grad_norm": 4.5700788497924805, + "learning_rate": 7.231481481481482e-06, + "loss": 1.0672, + "step": 4190 + }, + { + "epoch": 0.8232065856526852, + "grad_norm": 4.629611492156982, + "learning_rate": 7.222222222222223e-06, + "loss": 1.072, + "step": 4200 + }, + { + "epoch": 0.8232065856526852, + "eval_loss": 1.0551141500473022, + "eval_runtime": 13.9585, + "eval_samples_per_second": 46.853, + "eval_steps_per_second": 5.875, + "step": 4200 + }, + { + "epoch": 0.8251666013328106, + "grad_norm": 5.06083345413208, + "learning_rate": 7.212962962962964e-06, + "loss": 1.0641, + "step": 4210 + }, + { + "epoch": 0.8271266170129361, + "grad_norm": 4.53200626373291, + "learning_rate": 7.203703703703704e-06, + "loss": 1.0764, + "step": 4220 + }, + { + "epoch": 0.8290866326930616, + "grad_norm": 4.572458744049072, + "learning_rate": 7.194444444444445e-06, + "loss": 1.0616, + "step": 4230 + }, + { + "epoch": 0.831046648373187, + "grad_norm": 21.508056640625, + "learning_rate": 7.185185185185186e-06, + "loss": 1.0606, + "step": 4240 + }, + { + "epoch": 0.8330066640533125, + "grad_norm": 5.574312210083008, + "learning_rate": 7.1759259259259266e-06, + "loss": 1.0736, + "step": 4250 + }, + { + "epoch": 0.8349666797334379, + "grad_norm": 4.778395175933838, + "learning_rate": 7.166666666666667e-06, + "loss": 1.0561, + "step": 4260 + }, + { + "epoch": 0.8369266954135633, + "grad_norm": 4.696834564208984, + "learning_rate": 7.157407407407408e-06, + "loss": 1.0753, + "step": 4270 + }, + { + "epoch": 0.8388867110936887, + "grad_norm": 4.8193769454956055, + "learning_rate": 7.1481481481481486e-06, + "loss": 1.0454, + "step": 4280 + }, + { + "epoch": 0.8408467267738142, + "grad_norm": 4.693731784820557, + "learning_rate": 7.13888888888889e-06, + "loss": 1.0634, + "step": 4290 + }, + { + "epoch": 0.8428067424539396, + "grad_norm": 5.045989990234375, + "learning_rate": 7.129629629629629e-06, + "loss": 1.057, + "step": 4300 + }, + { + "epoch": 0.8447667581340651, + "grad_norm": 5.245114326477051, + "learning_rate": 7.120370370370371e-06, + "loss": 1.0435, + "step": 4310 + }, + { + "epoch": 0.8467267738141905, + "grad_norm": 5.094894886016846, + "learning_rate": 7.111111111111112e-06, + "loss": 1.0657, + "step": 4320 + }, + { + "epoch": 0.848686789494316, + "grad_norm": 4.630204200744629, + "learning_rate": 7.101851851851852e-06, + "loss": 1.0729, + "step": 4330 + }, + { + "epoch": 0.8506468051744414, + "grad_norm": 4.463089942932129, + "learning_rate": 7.0925925925925935e-06, + "loss": 1.0573, + "step": 4340 + }, + { + "epoch": 0.8526068208545668, + "grad_norm": 4.479245185852051, + "learning_rate": 7.083333333333335e-06, + "loss": 1.0534, + "step": 4350 + }, + { + "epoch": 0.8545668365346922, + "grad_norm": 4.832218170166016, + "learning_rate": 7.074074074074074e-06, + "loss": 1.0499, + "step": 4360 + }, + { + "epoch": 0.8565268522148177, + "grad_norm": 4.618699550628662, + "learning_rate": 7.0648148148148155e-06, + "loss": 1.073, + "step": 4370 + }, + { + "epoch": 0.8584868678949431, + "grad_norm": 4.646695137023926, + "learning_rate": 7.055555555555557e-06, + "loss": 1.0662, + "step": 4380 + }, + { + "epoch": 0.8604468835750686, + "grad_norm": 4.74224853515625, + "learning_rate": 7.046296296296296e-06, + "loss": 1.0516, + "step": 4390 + }, + { + "epoch": 0.8624068992551941, + "grad_norm": 4.836453914642334, + "learning_rate": 7.0370370370370375e-06, + "loss": 1.0622, + "step": 4400 + }, + { + "epoch": 0.8643669149353195, + "grad_norm": 5.0554046630859375, + "learning_rate": 7.027777777777778e-06, + "loss": 1.0725, + "step": 4410 + }, + { + "epoch": 0.866326930615445, + "grad_norm": 4.8123579025268555, + "learning_rate": 7.018518518518519e-06, + "loss": 1.0451, + "step": 4420 + }, + { + "epoch": 0.8682869462955703, + "grad_norm": 4.815413951873779, + "learning_rate": 7.00925925925926e-06, + "loss": 1.057, + "step": 4430 + }, + { + "epoch": 0.8702469619756958, + "grad_norm": 5.103263854980469, + "learning_rate": 7e-06, + "loss": 1.073, + "step": 4440 + }, + { + "epoch": 0.8722069776558212, + "grad_norm": 5.115508556365967, + "learning_rate": 6.990740740740741e-06, + "loss": 1.058, + "step": 4450 + }, + { + "epoch": 0.8741669933359467, + "grad_norm": 4.976015567779541, + "learning_rate": 6.981481481481482e-06, + "loss": 1.0601, + "step": 4460 + }, + { + "epoch": 0.8761270090160721, + "grad_norm": 5.111723899841309, + "learning_rate": 6.972222222222223e-06, + "loss": 1.0735, + "step": 4470 + }, + { + "epoch": 0.8780870246961976, + "grad_norm": 4.711007118225098, + "learning_rate": 6.962962962962964e-06, + "loss": 1.0723, + "step": 4480 + }, + { + "epoch": 0.880047040376323, + "grad_norm": 4.8820881843566895, + "learning_rate": 6.953703703703704e-06, + "loss": 1.0562, + "step": 4490 + }, + { + "epoch": 0.8820070560564485, + "grad_norm": 5.674580097198486, + "learning_rate": 6.944444444444445e-06, + "loss": 1.0657, + "step": 4500 + }, + { + "epoch": 0.8839670717365739, + "grad_norm": 4.517936706542969, + "learning_rate": 6.935185185185186e-06, + "loss": 1.0604, + "step": 4510 + }, + { + "epoch": 0.8859270874166993, + "grad_norm": 5.491835594177246, + "learning_rate": 6.9259259259259256e-06, + "loss": 1.0635, + "step": 4520 + }, + { + "epoch": 0.8878871030968247, + "grad_norm": 5.377597808837891, + "learning_rate": 6.916666666666667e-06, + "loss": 1.0567, + "step": 4530 + }, + { + "epoch": 0.8898471187769502, + "grad_norm": 5.251382350921631, + "learning_rate": 6.907407407407408e-06, + "loss": 1.0564, + "step": 4540 + }, + { + "epoch": 0.8918071344570757, + "grad_norm": 4.623981952667236, + "learning_rate": 6.898148148148148e-06, + "loss": 1.0663, + "step": 4550 + }, + { + "epoch": 0.8937671501372011, + "grad_norm": 4.460480690002441, + "learning_rate": 6.88888888888889e-06, + "loss": 1.0773, + "step": 4560 + }, + { + "epoch": 0.8957271658173266, + "grad_norm": 4.909152984619141, + "learning_rate": 6.879629629629631e-06, + "loss": 1.0462, + "step": 4570 + }, + { + "epoch": 0.897687181497452, + "grad_norm": 4.5406951904296875, + "learning_rate": 6.8703703703703704e-06, + "loss": 1.0619, + "step": 4580 + }, + { + "epoch": 0.8996471971775775, + "grad_norm": 4.8268890380859375, + "learning_rate": 6.861111111111112e-06, + "loss": 1.0749, + "step": 4590 + }, + { + "epoch": 0.9016072128577028, + "grad_norm": 5.024204254150391, + "learning_rate": 6.851851851851853e-06, + "loss": 1.082, + "step": 4600 + }, + { + "epoch": 0.9035672285378283, + "grad_norm": 4.6898908615112305, + "learning_rate": 6.842592592592593e-06, + "loss": 1.0527, + "step": 4610 + }, + { + "epoch": 0.9055272442179537, + "grad_norm": 5.445582866668701, + "learning_rate": 6.833333333333334e-06, + "loss": 1.05, + "step": 4620 + }, + { + "epoch": 0.9074872598980792, + "grad_norm": 4.625185966491699, + "learning_rate": 6.824074074074075e-06, + "loss": 1.0623, + "step": 4630 + }, + { + "epoch": 0.9094472755782046, + "grad_norm": 4.991087913513184, + "learning_rate": 6.814814814814815e-06, + "loss": 1.048, + "step": 4640 + }, + { + "epoch": 0.9114072912583301, + "grad_norm": 5.890523433685303, + "learning_rate": 6.8055555555555566e-06, + "loss": 1.0539, + "step": 4650 + }, + { + "epoch": 0.9133673069384555, + "grad_norm": 4.583831787109375, + "learning_rate": 6.796296296296296e-06, + "loss": 1.0616, + "step": 4660 + }, + { + "epoch": 0.915327322618581, + "grad_norm": 5.098804473876953, + "learning_rate": 6.787037037037037e-06, + "loss": 1.0555, + "step": 4670 + }, + { + "epoch": 0.9172873382987063, + "grad_norm": 4.51339864730835, + "learning_rate": 6.777777777777779e-06, + "loss": 1.0563, + "step": 4680 + }, + { + "epoch": 0.9192473539788318, + "grad_norm": 5.437107086181641, + "learning_rate": 6.768518518518519e-06, + "loss": 1.0639, + "step": 4690 + }, + { + "epoch": 0.9212073696589572, + "grad_norm": 5.220668315887451, + "learning_rate": 6.75925925925926e-06, + "loss": 1.0335, + "step": 4700 + }, + { + "epoch": 0.9231673853390827, + "grad_norm": 5.214885711669922, + "learning_rate": 6.750000000000001e-06, + "loss": 1.0616, + "step": 4710 + }, + { + "epoch": 0.9251274010192082, + "grad_norm": 4.807358264923096, + "learning_rate": 6.740740740740741e-06, + "loss": 1.0556, + "step": 4720 + }, + { + "epoch": 0.9270874166993336, + "grad_norm": 4.536919116973877, + "learning_rate": 6.731481481481482e-06, + "loss": 1.0427, + "step": 4730 + }, + { + "epoch": 0.9290474323794591, + "grad_norm": 4.960407733917236, + "learning_rate": 6.7222222222222235e-06, + "loss": 1.0573, + "step": 4740 + }, + { + "epoch": 0.9310074480595845, + "grad_norm": 5.058611869812012, + "learning_rate": 6.712962962962963e-06, + "loss": 1.0529, + "step": 4750 + }, + { + "epoch": 0.93296746373971, + "grad_norm": 4.611677169799805, + "learning_rate": 6.703703703703704e-06, + "loss": 1.0609, + "step": 4760 + }, + { + "epoch": 0.9349274794198353, + "grad_norm": 4.662081241607666, + "learning_rate": 6.694444444444445e-06, + "loss": 1.0585, + "step": 4770 + }, + { + "epoch": 0.9368874950999608, + "grad_norm": 4.845633029937744, + "learning_rate": 6.685185185185186e-06, + "loss": 1.0591, + "step": 4780 + }, + { + "epoch": 0.9388475107800862, + "grad_norm": 4.63094425201416, + "learning_rate": 6.675925925925927e-06, + "loss": 1.051, + "step": 4790 + }, + { + "epoch": 0.9408075264602117, + "grad_norm": 5.272851467132568, + "learning_rate": 6.666666666666667e-06, + "loss": 1.0643, + "step": 4800 + }, + { + "epoch": 0.9427675421403371, + "grad_norm": 5.056580066680908, + "learning_rate": 6.657407407407408e-06, + "loss": 1.0483, + "step": 4810 + }, + { + "epoch": 0.9447275578204626, + "grad_norm": 4.985904693603516, + "learning_rate": 6.648148148148149e-06, + "loss": 1.0526, + "step": 4820 + }, + { + "epoch": 0.946687573500588, + "grad_norm": 5.002072334289551, + "learning_rate": 6.6388888888888895e-06, + "loss": 1.0693, + "step": 4830 + }, + { + "epoch": 0.9486475891807135, + "grad_norm": 5.099184036254883, + "learning_rate": 6.62962962962963e-06, + "loss": 1.0629, + "step": 4840 + }, + { + "epoch": 0.9506076048608388, + "grad_norm": 4.710622310638428, + "learning_rate": 6.620370370370371e-06, + "loss": 1.0458, + "step": 4850 + }, + { + "epoch": 0.9525676205409643, + "grad_norm": 5.143291473388672, + "learning_rate": 6.6111111111111115e-06, + "loss": 1.0552, + "step": 4860 + }, + { + "epoch": 0.9545276362210897, + "grad_norm": 5.28003454208374, + "learning_rate": 6.601851851851853e-06, + "loss": 1.0643, + "step": 4870 + }, + { + "epoch": 0.9564876519012152, + "grad_norm": 5.275862216949463, + "learning_rate": 6.592592592592592e-06, + "loss": 1.0566, + "step": 4880 + }, + { + "epoch": 0.9584476675813407, + "grad_norm": 4.868998050689697, + "learning_rate": 6.5833333333333335e-06, + "loss": 1.0703, + "step": 4890 + }, + { + "epoch": 0.9604076832614661, + "grad_norm": 4.964428424835205, + "learning_rate": 6.574074074074075e-06, + "loss": 1.0564, + "step": 4900 + }, + { + "epoch": 0.9604076832614661, + "eval_loss": 1.0497612953186035, + "eval_runtime": 14.0356, + "eval_samples_per_second": 46.596, + "eval_steps_per_second": 5.842, + "step": 4900 + }, + { + "epoch": 0.9623676989415916, + "grad_norm": 4.545341968536377, + "learning_rate": 6.564814814814815e-06, + "loss": 1.0741, + "step": 4910 + }, + { + "epoch": 0.964327714621717, + "grad_norm": 5.679270267486572, + "learning_rate": 6.555555555555556e-06, + "loss": 1.0528, + "step": 4920 + }, + { + "epoch": 0.9662877303018425, + "grad_norm": 4.5773749351501465, + "learning_rate": 6.546296296296298e-06, + "loss": 1.048, + "step": 4930 + }, + { + "epoch": 0.9682477459819678, + "grad_norm": 5.288699150085449, + "learning_rate": 6.537037037037037e-06, + "loss": 1.0487, + "step": 4940 + }, + { + "epoch": 0.9702077616620933, + "grad_norm": 5.2343363761901855, + "learning_rate": 6.5277777777777784e-06, + "loss": 1.0523, + "step": 4950 + }, + { + "epoch": 0.9721677773422187, + "grad_norm": 4.845068454742432, + "learning_rate": 6.51851851851852e-06, + "loss": 1.0519, + "step": 4960 + }, + { + "epoch": 0.9741277930223442, + "grad_norm": 5.101285457611084, + "learning_rate": 6.509259259259259e-06, + "loss": 1.0618, + "step": 4970 + }, + { + "epoch": 0.9760878087024696, + "grad_norm": 4.684749126434326, + "learning_rate": 6.5000000000000004e-06, + "loss": 1.0506, + "step": 4980 + }, + { + "epoch": 0.9780478243825951, + "grad_norm": 4.986086368560791, + "learning_rate": 6.490740740740741e-06, + "loss": 1.0621, + "step": 4990 + }, + { + "epoch": 0.9800078400627205, + "grad_norm": 4.658474922180176, + "learning_rate": 6.481481481481482e-06, + "loss": 1.0727, + "step": 5000 + }, + { + "epoch": 0.981967855742846, + "grad_norm": 5.1348419189453125, + "learning_rate": 6.472222222222223e-06, + "loss": 1.0542, + "step": 5010 + }, + { + "epoch": 0.9839278714229713, + "grad_norm": 5.368491172790527, + "learning_rate": 6.462962962962963e-06, + "loss": 1.0656, + "step": 5020 + }, + { + "epoch": 0.9858878871030968, + "grad_norm": 5.312760829925537, + "learning_rate": 6.453703703703704e-06, + "loss": 1.0587, + "step": 5030 + }, + { + "epoch": 0.9878479027832223, + "grad_norm": 5.235229969024658, + "learning_rate": 6.444444444444445e-06, + "loss": 1.0588, + "step": 5040 + }, + { + "epoch": 0.9898079184633477, + "grad_norm": 4.761176586151123, + "learning_rate": 6.435185185185186e-06, + "loss": 1.0507, + "step": 5050 + }, + { + "epoch": 0.9917679341434732, + "grad_norm": 5.440311431884766, + "learning_rate": 6.425925925925927e-06, + "loss": 1.0566, + "step": 5060 + }, + { + "epoch": 0.9937279498235986, + "grad_norm": 4.793478488922119, + "learning_rate": 6.416666666666667e-06, + "loss": 1.0607, + "step": 5070 + }, + { + "epoch": 0.9956879655037241, + "grad_norm": 5.079543113708496, + "learning_rate": 6.407407407407408e-06, + "loss": 1.0494, + "step": 5080 + }, + { + "epoch": 0.9976479811838495, + "grad_norm": 4.377906322479248, + "learning_rate": 6.398148148148149e-06, + "loss": 1.0558, + "step": 5090 + }, + { + "epoch": 0.999607996863975, + "grad_norm": 4.611093521118164, + "learning_rate": 6.3888888888888885e-06, + "loss": 1.0693, + "step": 5100 + }, + { + "epoch": 1.0015680125441004, + "grad_norm": 4.401882171630859, + "learning_rate": 6.37962962962963e-06, + "loss": 1.0297, + "step": 5110 + }, + { + "epoch": 1.0035280282242258, + "grad_norm": 4.928489685058594, + "learning_rate": 6.370370370370371e-06, + "loss": 1.0276, + "step": 5120 + }, + { + "epoch": 1.0054880439043512, + "grad_norm": 5.5292487144470215, + "learning_rate": 6.361111111111111e-06, + "loss": 1.0267, + "step": 5130 + }, + { + "epoch": 1.0074480595844766, + "grad_norm": 5.00799036026001, + "learning_rate": 6.351851851851853e-06, + "loss": 1.0173, + "step": 5140 + }, + { + "epoch": 1.0094080752646022, + "grad_norm": 4.399227619171143, + "learning_rate": 6.342592592592594e-06, + "loss": 1.0106, + "step": 5150 + }, + { + "epoch": 1.0113680909447276, + "grad_norm": 5.132420539855957, + "learning_rate": 6.333333333333333e-06, + "loss": 1.0212, + "step": 5160 + }, + { + "epoch": 1.013328106624853, + "grad_norm": 5.298924446105957, + "learning_rate": 6.324074074074075e-06, + "loss": 1.0122, + "step": 5170 + }, + { + "epoch": 1.0152881223049783, + "grad_norm": 5.066485404968262, + "learning_rate": 6.314814814814816e-06, + "loss": 1.0114, + "step": 5180 + }, + { + "epoch": 1.017248137985104, + "grad_norm": 6.66790771484375, + "learning_rate": 6.305555555555556e-06, + "loss": 1.0128, + "step": 5190 + }, + { + "epoch": 1.0192081536652293, + "grad_norm": 4.905239105224609, + "learning_rate": 6.296296296296297e-06, + "loss": 1.0118, + "step": 5200 + }, + { + "epoch": 1.0211681693453547, + "grad_norm": 5.377009868621826, + "learning_rate": 6.287037037037037e-06, + "loss": 1.0229, + "step": 5210 + }, + { + "epoch": 1.0231281850254803, + "grad_norm": 5.1396942138671875, + "learning_rate": 6.277777777777778e-06, + "loss": 1.0221, + "step": 5220 + }, + { + "epoch": 1.0250882007056057, + "grad_norm": 5.001946449279785, + "learning_rate": 6.2685185185185195e-06, + "loss": 1.0461, + "step": 5230 + }, + { + "epoch": 1.027048216385731, + "grad_norm": 6.100106716156006, + "learning_rate": 6.259259259259259e-06, + "loss": 1.0169, + "step": 5240 + }, + { + "epoch": 1.0290082320658565, + "grad_norm": 5.1383209228515625, + "learning_rate": 6.25e-06, + "loss": 1.0282, + "step": 5250 + }, + { + "epoch": 1.030968247745982, + "grad_norm": 5.477851390838623, + "learning_rate": 6.2407407407407415e-06, + "loss": 1.0266, + "step": 5260 + }, + { + "epoch": 1.0329282634261074, + "grad_norm": 5.047489643096924, + "learning_rate": 6.231481481481482e-06, + "loss": 1.0256, + "step": 5270 + }, + { + "epoch": 1.0348882791062328, + "grad_norm": 4.994131565093994, + "learning_rate": 6.222222222222223e-06, + "loss": 1.0224, + "step": 5280 + }, + { + "epoch": 1.0368482947863582, + "grad_norm": 5.558932781219482, + "learning_rate": 6.2129629629629636e-06, + "loss": 1.0186, + "step": 5290 + }, + { + "epoch": 1.0388083104664838, + "grad_norm": 5.365056037902832, + "learning_rate": 6.203703703703704e-06, + "loss": 1.0095, + "step": 5300 + }, + { + "epoch": 1.0407683261466092, + "grad_norm": 5.53687858581543, + "learning_rate": 6.194444444444445e-06, + "loss": 1.0132, + "step": 5310 + }, + { + "epoch": 1.0427283418267346, + "grad_norm": 5.936549663543701, + "learning_rate": 6.1851851851851856e-06, + "loss": 1.0347, + "step": 5320 + }, + { + "epoch": 1.04468835750686, + "grad_norm": 5.228058338165283, + "learning_rate": 6.175925925925926e-06, + "loss": 1.0234, + "step": 5330 + }, + { + "epoch": 1.0466483731869856, + "grad_norm": 4.896886348724365, + "learning_rate": 6.166666666666667e-06, + "loss": 1.0065, + "step": 5340 + }, + { + "epoch": 1.048608388867111, + "grad_norm": 6.313221454620361, + "learning_rate": 6.157407407407408e-06, + "loss": 1.0191, + "step": 5350 + }, + { + "epoch": 1.0505684045472363, + "grad_norm": 5.8848114013671875, + "learning_rate": 6.148148148148149e-06, + "loss": 1.0378, + "step": 5360 + }, + { + "epoch": 1.052528420227362, + "grad_norm": 5.468878746032715, + "learning_rate": 6.13888888888889e-06, + "loss": 1.0056, + "step": 5370 + }, + { + "epoch": 1.0544884359074873, + "grad_norm": 5.208171367645264, + "learning_rate": 6.12962962962963e-06, + "loss": 1.0194, + "step": 5380 + }, + { + "epoch": 1.0564484515876127, + "grad_norm": 5.1294450759887695, + "learning_rate": 6.120370370370371e-06, + "loss": 1.0084, + "step": 5390 + }, + { + "epoch": 1.058408467267738, + "grad_norm": 5.981677532196045, + "learning_rate": 6.111111111111112e-06, + "loss": 1.0148, + "step": 5400 + }, + { + "epoch": 1.0603684829478637, + "grad_norm": 5.662644863128662, + "learning_rate": 6.1018518518518525e-06, + "loss": 1.0163, + "step": 5410 + }, + { + "epoch": 1.062328498627989, + "grad_norm": 5.452698230743408, + "learning_rate": 6.092592592592593e-06, + "loss": 1.0047, + "step": 5420 + }, + { + "epoch": 1.0642885143081144, + "grad_norm": 4.969189167022705, + "learning_rate": 6.083333333333333e-06, + "loss": 1.0031, + "step": 5430 + }, + { + "epoch": 1.0662485299882398, + "grad_norm": 5.396177768707275, + "learning_rate": 6.0740740740740745e-06, + "loss": 1.0109, + "step": 5440 + }, + { + "epoch": 1.0682085456683654, + "grad_norm": 6.180945873260498, + "learning_rate": 6.064814814814816e-06, + "loss": 1.0318, + "step": 5450 + }, + { + "epoch": 1.0701685613484908, + "grad_norm": 4.878726959228516, + "learning_rate": 6.055555555555555e-06, + "loss": 1.0257, + "step": 5460 + }, + { + "epoch": 1.0721285770286162, + "grad_norm": 5.634671688079834, + "learning_rate": 6.0462962962962965e-06, + "loss": 1.0241, + "step": 5470 + }, + { + "epoch": 1.0740885927087416, + "grad_norm": 5.971144199371338, + "learning_rate": 6.037037037037038e-06, + "loss": 1.0247, + "step": 5480 + }, + { + "epoch": 1.0760486083888672, + "grad_norm": 5.1925950050354, + "learning_rate": 6.027777777777778e-06, + "loss": 1.0094, + "step": 5490 + }, + { + "epoch": 1.0780086240689926, + "grad_norm": 4.921753406524658, + "learning_rate": 6.018518518518519e-06, + "loss": 1.0233, + "step": 5500 + }, + { + "epoch": 1.079968639749118, + "grad_norm": 5.969301700592041, + "learning_rate": 6.009259259259261e-06, + "loss": 0.994, + "step": 5510 + }, + { + "epoch": 1.0819286554292433, + "grad_norm": 5.875487804412842, + "learning_rate": 6e-06, + "loss": 1.0171, + "step": 5520 + }, + { + "epoch": 1.083888671109369, + "grad_norm": 5.540329933166504, + "learning_rate": 5.990740740740741e-06, + "loss": 1.0198, + "step": 5530 + }, + { + "epoch": 1.0858486867894943, + "grad_norm": 5.154452323913574, + "learning_rate": 5.981481481481482e-06, + "loss": 1.0174, + "step": 5540 + }, + { + "epoch": 1.0878087024696197, + "grad_norm": 5.469849586486816, + "learning_rate": 5.972222222222222e-06, + "loss": 1.0141, + "step": 5550 + }, + { + "epoch": 1.0897687181497453, + "grad_norm": 6.041581153869629, + "learning_rate": 5.962962962962963e-06, + "loss": 1.0216, + "step": 5560 + }, + { + "epoch": 1.0917287338298707, + "grad_norm": 5.2390522956848145, + "learning_rate": 5.953703703703704e-06, + "loss": 1.0202, + "step": 5570 + }, + { + "epoch": 1.093688749509996, + "grad_norm": 5.63343620300293, + "learning_rate": 5.944444444444445e-06, + "loss": 1.0234, + "step": 5580 + }, + { + "epoch": 1.0956487651901214, + "grad_norm": 14.769401550292969, + "learning_rate": 5.935185185185186e-06, + "loss": 1.019, + "step": 5590 + }, + { + "epoch": 1.097608780870247, + "grad_norm": 5.917840957641602, + "learning_rate": 5.925925925925926e-06, + "loss": 1.0259, + "step": 5600 + }, + { + "epoch": 1.097608780870247, + "eval_loss": 1.0469825267791748, + "eval_runtime": 14.056, + "eval_samples_per_second": 46.528, + "eval_steps_per_second": 5.834, + "step": 5600 + }, + { + "epoch": 1.0995687965503724, + "grad_norm": 4.929704666137695, + "learning_rate": 5.916666666666667e-06, + "loss": 1.0146, + "step": 5610 + }, + { + "epoch": 1.1015288122304978, + "grad_norm": 6.432901382446289, + "learning_rate": 5.907407407407408e-06, + "loss": 1.0268, + "step": 5620 + }, + { + "epoch": 1.1034888279106232, + "grad_norm": 5.020516872406006, + "learning_rate": 5.898148148148149e-06, + "loss": 1.0272, + "step": 5630 + }, + { + "epoch": 1.1054488435907488, + "grad_norm": 5.231268882751465, + "learning_rate": 5.88888888888889e-06, + "loss": 1.0046, + "step": 5640 + }, + { + "epoch": 1.1074088592708742, + "grad_norm": 5.203913688659668, + "learning_rate": 5.8796296296296295e-06, + "loss": 1.015, + "step": 5650 + }, + { + "epoch": 1.1093688749509996, + "grad_norm": 5.593926906585693, + "learning_rate": 5.870370370370371e-06, + "loss": 1.0185, + "step": 5660 + }, + { + "epoch": 1.1113288906311252, + "grad_norm": 5.165170669555664, + "learning_rate": 5.861111111111112e-06, + "loss": 1.0056, + "step": 5670 + }, + { + "epoch": 1.1132889063112505, + "grad_norm": 5.521978855133057, + "learning_rate": 5.8518518518518515e-06, + "loss": 1.022, + "step": 5680 + }, + { + "epoch": 1.115248921991376, + "grad_norm": 5.280319690704346, + "learning_rate": 5.842592592592593e-06, + "loss": 1.0357, + "step": 5690 + }, + { + "epoch": 1.1172089376715013, + "grad_norm": 6.761595726013184, + "learning_rate": 5.833333333333334e-06, + "loss": 1.0229, + "step": 5700 + }, + { + "epoch": 1.119168953351627, + "grad_norm": 5.606836318969727, + "learning_rate": 5.824074074074074e-06, + "loss": 1.019, + "step": 5710 + }, + { + "epoch": 1.1211289690317523, + "grad_norm": 5.351015567779541, + "learning_rate": 5.814814814814816e-06, + "loss": 1.0343, + "step": 5720 + }, + { + "epoch": 1.1230889847118777, + "grad_norm": 4.853587627410889, + "learning_rate": 5.805555555555557e-06, + "loss": 1.0239, + "step": 5730 + }, + { + "epoch": 1.125049000392003, + "grad_norm": 16.355466842651367, + "learning_rate": 5.796296296296296e-06, + "loss": 1.0013, + "step": 5740 + }, + { + "epoch": 1.1270090160721287, + "grad_norm": 5.404659748077393, + "learning_rate": 5.787037037037038e-06, + "loss": 1.0181, + "step": 5750 + }, + { + "epoch": 1.128969031752254, + "grad_norm": 5.103781700134277, + "learning_rate": 5.777777777777778e-06, + "loss": 1.0213, + "step": 5760 + }, + { + "epoch": 1.1309290474323794, + "grad_norm": 5.420756816864014, + "learning_rate": 5.768518518518519e-06, + "loss": 1.0072, + "step": 5770 + }, + { + "epoch": 1.1328890631125048, + "grad_norm": 5.301304817199707, + "learning_rate": 5.75925925925926e-06, + "loss": 0.9901, + "step": 5780 + }, + { + "epoch": 1.1348490787926304, + "grad_norm": 4.864381790161133, + "learning_rate": 5.75e-06, + "loss": 1.0129, + "step": 5790 + }, + { + "epoch": 1.1368090944727558, + "grad_norm": 6.017991542816162, + "learning_rate": 5.740740740740741e-06, + "loss": 1.0113, + "step": 5800 + }, + { + "epoch": 1.1387691101528812, + "grad_norm": 5.6904520988464355, + "learning_rate": 5.7314814814814825e-06, + "loss": 1.0211, + "step": 5810 + }, + { + "epoch": 1.1407291258330066, + "grad_norm": 5.778028964996338, + "learning_rate": 5.722222222222222e-06, + "loss": 1.0247, + "step": 5820 + }, + { + "epoch": 1.1426891415131322, + "grad_norm": 5.682682991027832, + "learning_rate": 5.712962962962963e-06, + "loss": 1.0309, + "step": 5830 + }, + { + "epoch": 1.1446491571932575, + "grad_norm": 5.886664867401123, + "learning_rate": 5.7037037037037045e-06, + "loss": 1.0244, + "step": 5840 + }, + { + "epoch": 1.146609172873383, + "grad_norm": 5.014996528625488, + "learning_rate": 5.694444444444445e-06, + "loss": 1.0207, + "step": 5850 + }, + { + "epoch": 1.1485691885535085, + "grad_norm": 5.563379287719727, + "learning_rate": 5.685185185185186e-06, + "loss": 1.0276, + "step": 5860 + }, + { + "epoch": 1.150529204233634, + "grad_norm": 5.241649627685547, + "learning_rate": 5.675925925925926e-06, + "loss": 1.022, + "step": 5870 + }, + { + "epoch": 1.1524892199137593, + "grad_norm": 5.154331207275391, + "learning_rate": 5.666666666666667e-06, + "loss": 1.0232, + "step": 5880 + }, + { + "epoch": 1.1544492355938847, + "grad_norm": 5.823696136474609, + "learning_rate": 5.657407407407408e-06, + "loss": 1.0364, + "step": 5890 + }, + { + "epoch": 1.1564092512740103, + "grad_norm": 5.451704978942871, + "learning_rate": 5.6481481481481485e-06, + "loss": 1.02, + "step": 5900 + }, + { + "epoch": 1.1583692669541357, + "grad_norm": 5.540503978729248, + "learning_rate": 5.638888888888889e-06, + "loss": 1.0127, + "step": 5910 + }, + { + "epoch": 1.160329282634261, + "grad_norm": 5.463021278381348, + "learning_rate": 5.62962962962963e-06, + "loss": 1.0205, + "step": 5920 + }, + { + "epoch": 1.1622892983143864, + "grad_norm": 5.372107028961182, + "learning_rate": 5.6203703703703705e-06, + "loss": 1.0085, + "step": 5930 + }, + { + "epoch": 1.164249313994512, + "grad_norm": 5.14231538772583, + "learning_rate": 5.611111111111112e-06, + "loss": 1.0189, + "step": 5940 + }, + { + "epoch": 1.1662093296746374, + "grad_norm": 5.416358470916748, + "learning_rate": 5.601851851851853e-06, + "loss": 1.0066, + "step": 5950 + }, + { + "epoch": 1.1681693453547628, + "grad_norm": 5.824409484863281, + "learning_rate": 5.5925925925925926e-06, + "loss": 1.0294, + "step": 5960 + }, + { + "epoch": 1.1701293610348882, + "grad_norm": 5.362888336181641, + "learning_rate": 5.583333333333334e-06, + "loss": 1.011, + "step": 5970 + }, + { + "epoch": 1.1720893767150138, + "grad_norm": 5.583401679992676, + "learning_rate": 5.574074074074075e-06, + "loss": 1.0218, + "step": 5980 + }, + { + "epoch": 1.1740493923951392, + "grad_norm": 5.501253128051758, + "learning_rate": 5.5648148148148154e-06, + "loss": 1.0198, + "step": 5990 + }, + { + "epoch": 1.1760094080752646, + "grad_norm": 5.536640167236328, + "learning_rate": 5.555555555555557e-06, + "loss": 1.0243, + "step": 6000 + }, + { + "epoch": 1.17796942375539, + "grad_norm": 5.525614261627197, + "learning_rate": 5.546296296296296e-06, + "loss": 1.0234, + "step": 6010 + }, + { + "epoch": 1.1799294394355155, + "grad_norm": 5.270368576049805, + "learning_rate": 5.5370370370370374e-06, + "loss": 1.0214, + "step": 6020 + }, + { + "epoch": 1.181889455115641, + "grad_norm": 5.895664691925049, + "learning_rate": 5.527777777777779e-06, + "loss": 1.0203, + "step": 6030 + }, + { + "epoch": 1.1838494707957663, + "grad_norm": 5.530089378356934, + "learning_rate": 5.518518518518518e-06, + "loss": 1.0254, + "step": 6040 + }, + { + "epoch": 1.185809486475892, + "grad_norm": 5.275720119476318, + "learning_rate": 5.5092592592592595e-06, + "loss": 1.015, + "step": 6050 + }, + { + "epoch": 1.1877695021560173, + "grad_norm": 5.464200496673584, + "learning_rate": 5.500000000000001e-06, + "loss": 1.0134, + "step": 6060 + }, + { + "epoch": 1.1897295178361427, + "grad_norm": 5.803837299346924, + "learning_rate": 5.490740740740741e-06, + "loss": 1.0125, + "step": 6070 + }, + { + "epoch": 1.191689533516268, + "grad_norm": 5.1351752281188965, + "learning_rate": 5.481481481481482e-06, + "loss": 1.0061, + "step": 6080 + }, + { + "epoch": 1.1936495491963937, + "grad_norm": 5.232799530029297, + "learning_rate": 5.4722222222222236e-06, + "loss": 1.0213, + "step": 6090 + }, + { + "epoch": 1.195609564876519, + "grad_norm": 5.996954917907715, + "learning_rate": 5.462962962962963e-06, + "loss": 1.0177, + "step": 6100 + }, + { + "epoch": 1.1975695805566444, + "grad_norm": 7.579484939575195, + "learning_rate": 5.453703703703704e-06, + "loss": 1.0133, + "step": 6110 + }, + { + "epoch": 1.1995295962367698, + "grad_norm": 5.351324081420898, + "learning_rate": 5.444444444444445e-06, + "loss": 1.0019, + "step": 6120 + }, + { + "epoch": 1.2014896119168954, + "grad_norm": 5.5653581619262695, + "learning_rate": 5.435185185185186e-06, + "loss": 1.0069, + "step": 6130 + }, + { + "epoch": 1.2034496275970208, + "grad_norm": 5.198008060455322, + "learning_rate": 5.425925925925926e-06, + "loss": 1.0157, + "step": 6140 + }, + { + "epoch": 1.2054096432771462, + "grad_norm": 5.67602014541626, + "learning_rate": 5.416666666666667e-06, + "loss": 1.0327, + "step": 6150 + }, + { + "epoch": 1.2073696589572718, + "grad_norm": 5.185479164123535, + "learning_rate": 5.407407407407408e-06, + "loss": 1.0118, + "step": 6160 + }, + { + "epoch": 1.2093296746373972, + "grad_norm": 5.254889965057373, + "learning_rate": 5.398148148148149e-06, + "loss": 1.0282, + "step": 6170 + }, + { + "epoch": 1.2112896903175225, + "grad_norm": 5.635478496551514, + "learning_rate": 5.388888888888889e-06, + "loss": 1.0211, + "step": 6180 + }, + { + "epoch": 1.213249705997648, + "grad_norm": 5.388778209686279, + "learning_rate": 5.37962962962963e-06, + "loss": 1.0171, + "step": 6190 + }, + { + "epoch": 1.2152097216777733, + "grad_norm": 5.546689987182617, + "learning_rate": 5.370370370370371e-06, + "loss": 1.0113, + "step": 6200 + }, + { + "epoch": 1.217169737357899, + "grad_norm": 5.316011428833008, + "learning_rate": 5.361111111111112e-06, + "loss": 1.0183, + "step": 6210 + }, + { + "epoch": 1.2191297530380243, + "grad_norm": 6.057391166687012, + "learning_rate": 5.351851851851853e-06, + "loss": 1.0149, + "step": 6220 + }, + { + "epoch": 1.2210897687181497, + "grad_norm": 5.451624393463135, + "learning_rate": 5.342592592592592e-06, + "loss": 1.0129, + "step": 6230 + }, + { + "epoch": 1.2230497843982753, + "grad_norm": 5.58920955657959, + "learning_rate": 5.333333333333334e-06, + "loss": 1.0152, + "step": 6240 + }, + { + "epoch": 1.2250098000784007, + "grad_norm": 5.132778167724609, + "learning_rate": 5.324074074074075e-06, + "loss": 1.0059, + "step": 6250 + }, + { + "epoch": 1.226969815758526, + "grad_norm": 4.807291030883789, + "learning_rate": 5.314814814814815e-06, + "loss": 1.0146, + "step": 6260 + }, + { + "epoch": 1.2289298314386514, + "grad_norm": 5.3906474113464355, + "learning_rate": 5.305555555555556e-06, + "loss": 1.0045, + "step": 6270 + }, + { + "epoch": 1.230889847118777, + "grad_norm": 5.229481220245361, + "learning_rate": 5.296296296296297e-06, + "loss": 1.0018, + "step": 6280 + }, + { + "epoch": 1.2328498627989024, + "grad_norm": 5.5774126052856445, + "learning_rate": 5.287037037037037e-06, + "loss": 1.0096, + "step": 6290 + }, + { + "epoch": 1.2348098784790278, + "grad_norm": 5.752622127532959, + "learning_rate": 5.2777777777777785e-06, + "loss": 1.0173, + "step": 6300 + }, + { + "epoch": 1.2348098784790278, + "eval_loss": 1.0454390048980713, + "eval_runtime": 14.0845, + "eval_samples_per_second": 46.434, + "eval_steps_per_second": 5.822, + "step": 6300 + }, + { + "epoch": 1.2367698941591532, + "grad_norm": 4.995856285095215, + "learning_rate": 5.26851851851852e-06, + "loss": 1.0274, + "step": 6310 + }, + { + "epoch": 1.2387299098392788, + "grad_norm": 5.631256103515625, + "learning_rate": 5.259259259259259e-06, + "loss": 1.029, + "step": 6320 + }, + { + "epoch": 1.2406899255194042, + "grad_norm": 5.617944717407227, + "learning_rate": 5.2500000000000006e-06, + "loss": 1.0087, + "step": 6330 + }, + { + "epoch": 1.2426499411995295, + "grad_norm": 5.174662113189697, + "learning_rate": 5.240740740740741e-06, + "loss": 1.0114, + "step": 6340 + }, + { + "epoch": 1.2446099568796551, + "grad_norm": 5.841761112213135, + "learning_rate": 5.231481481481482e-06, + "loss": 1.0113, + "step": 6350 + }, + { + "epoch": 1.2465699725597805, + "grad_norm": 5.008285999298096, + "learning_rate": 5.2222222222222226e-06, + "loss": 1.0058, + "step": 6360 + }, + { + "epoch": 1.248529988239906, + "grad_norm": 5.529172420501709, + "learning_rate": 5.212962962962963e-06, + "loss": 1.0225, + "step": 6370 + }, + { + "epoch": 1.2504900039200313, + "grad_norm": 6.809742450714111, + "learning_rate": 5.203703703703704e-06, + "loss": 1.0194, + "step": 6380 + }, + { + "epoch": 1.2524500196001567, + "grad_norm": 5.778404712677002, + "learning_rate": 5.1944444444444454e-06, + "loss": 1.0312, + "step": 6390 + }, + { + "epoch": 1.2544100352802823, + "grad_norm": 6.095964431762695, + "learning_rate": 5.185185185185185e-06, + "loss": 1.0155, + "step": 6400 + }, + { + "epoch": 1.2563700509604077, + "grad_norm": 5.453831672668457, + "learning_rate": 5.175925925925926e-06, + "loss": 1.0155, + "step": 6410 + }, + { + "epoch": 1.258330066640533, + "grad_norm": 5.425611972808838, + "learning_rate": 5.1666666666666675e-06, + "loss": 1.0353, + "step": 6420 + }, + { + "epoch": 1.2602900823206586, + "grad_norm": 6.835920810699463, + "learning_rate": 5.157407407407408e-06, + "loss": 1.0089, + "step": 6430 + }, + { + "epoch": 1.262250098000784, + "grad_norm": 8.35204029083252, + "learning_rate": 5.148148148148149e-06, + "loss": 1.0235, + "step": 6440 + }, + { + "epoch": 1.2642101136809094, + "grad_norm": 5.1433234214782715, + "learning_rate": 5.138888888888889e-06, + "loss": 1.0226, + "step": 6450 + }, + { + "epoch": 1.266170129361035, + "grad_norm": 5.907522201538086, + "learning_rate": 5.12962962962963e-06, + "loss": 1.0161, + "step": 6460 + }, + { + "epoch": 1.2681301450411604, + "grad_norm": 5.631725788116455, + "learning_rate": 5.120370370370371e-06, + "loss": 1.014, + "step": 6470 + }, + { + "epoch": 1.2700901607212858, + "grad_norm": 5.635436058044434, + "learning_rate": 5.1111111111111115e-06, + "loss": 1.0216, + "step": 6480 + }, + { + "epoch": 1.2720501764014112, + "grad_norm": 6.376223087310791, + "learning_rate": 5.101851851851852e-06, + "loss": 1.0135, + "step": 6490 + }, + { + "epoch": 1.2740101920815365, + "grad_norm": 6.083188533782959, + "learning_rate": 5.092592592592593e-06, + "loss": 1.0117, + "step": 6500 + }, + { + "epoch": 1.2759702077616621, + "grad_norm": 5.436441898345947, + "learning_rate": 5.0833333333333335e-06, + "loss": 1.0079, + "step": 6510 + }, + { + "epoch": 1.2779302234417875, + "grad_norm": 5.303137302398682, + "learning_rate": 5.074074074074075e-06, + "loss": 1.02, + "step": 6520 + }, + { + "epoch": 1.279890239121913, + "grad_norm": 5.048837661743164, + "learning_rate": 5.064814814814816e-06, + "loss": 1.0045, + "step": 6530 + }, + { + "epoch": 1.2818502548020385, + "grad_norm": 6.105396747589111, + "learning_rate": 5.0555555555555555e-06, + "loss": 1.0105, + "step": 6540 + }, + { + "epoch": 1.283810270482164, + "grad_norm": 4.97545051574707, + "learning_rate": 5.046296296296297e-06, + "loss": 1.0128, + "step": 6550 + }, + { + "epoch": 1.2857702861622893, + "grad_norm": 6.359137535095215, + "learning_rate": 5.037037037037037e-06, + "loss": 1.0127, + "step": 6560 + }, + { + "epoch": 1.2877303018424147, + "grad_norm": 6.185931205749512, + "learning_rate": 5.027777777777778e-06, + "loss": 1.0156, + "step": 6570 + }, + { + "epoch": 1.28969031752254, + "grad_norm": 6.333951950073242, + "learning_rate": 5.01851851851852e-06, + "loss": 1.0113, + "step": 6580 + }, + { + "epoch": 1.2916503332026656, + "grad_norm": 5.14535665512085, + "learning_rate": 5.009259259259259e-06, + "loss": 1.0062, + "step": 6590 + }, + { + "epoch": 1.293610348882791, + "grad_norm": 6.086668491363525, + "learning_rate": 5e-06, + "loss": 1.0123, + "step": 6600 + }, + { + "epoch": 1.2955703645629164, + "grad_norm": 5.336708068847656, + "learning_rate": 4.990740740740741e-06, + "loss": 1.0058, + "step": 6610 + }, + { + "epoch": 1.297530380243042, + "grad_norm": 5.65183687210083, + "learning_rate": 4.981481481481482e-06, + "loss": 1.0232, + "step": 6620 + }, + { + "epoch": 1.2994903959231674, + "grad_norm": 5.356630802154541, + "learning_rate": 4.9722222222222224e-06, + "loss": 1.0248, + "step": 6630 + }, + { + "epoch": 1.3014504116032928, + "grad_norm": 5.61469841003418, + "learning_rate": 4.962962962962964e-06, + "loss": 1.0049, + "step": 6640 + }, + { + "epoch": 1.3034104272834184, + "grad_norm": 5.575668811798096, + "learning_rate": 4.953703703703704e-06, + "loss": 1.0046, + "step": 6650 + }, + { + "epoch": 1.3053704429635438, + "grad_norm": 5.938807010650635, + "learning_rate": 4.944444444444445e-06, + "loss": 1.0254, + "step": 6660 + }, + { + "epoch": 1.3073304586436691, + "grad_norm": 5.57589054107666, + "learning_rate": 4.935185185185186e-06, + "loss": 1.016, + "step": 6670 + }, + { + "epoch": 1.3092904743237945, + "grad_norm": 5.684128284454346, + "learning_rate": 4.925925925925926e-06, + "loss": 1.0102, + "step": 6680 + }, + { + "epoch": 1.31125049000392, + "grad_norm": 5.43093729019165, + "learning_rate": 4.9166666666666665e-06, + "loss": 1.0092, + "step": 6690 + }, + { + "epoch": 1.3132105056840455, + "grad_norm": 5.595739841461182, + "learning_rate": 4.907407407407408e-06, + "loss": 1.0103, + "step": 6700 + }, + { + "epoch": 1.315170521364171, + "grad_norm": 6.116888046264648, + "learning_rate": 4.898148148148149e-06, + "loss": 1.0177, + "step": 6710 + }, + { + "epoch": 1.3171305370442963, + "grad_norm": 5.478735446929932, + "learning_rate": 4.888888888888889e-06, + "loss": 1.0147, + "step": 6720 + }, + { + "epoch": 1.3190905527244219, + "grad_norm": 5.878492832183838, + "learning_rate": 4.8796296296296306e-06, + "loss": 1.0189, + "step": 6730 + }, + { + "epoch": 1.3210505684045473, + "grad_norm": 5.890417098999023, + "learning_rate": 4.870370370370371e-06, + "loss": 1.0246, + "step": 6740 + }, + { + "epoch": 1.3230105840846726, + "grad_norm": 5.2328410148620605, + "learning_rate": 4.861111111111111e-06, + "loss": 1.0116, + "step": 6750 + }, + { + "epoch": 1.3249705997647983, + "grad_norm": 6.1324005126953125, + "learning_rate": 4.851851851851852e-06, + "loss": 1.0028, + "step": 6760 + }, + { + "epoch": 1.3269306154449236, + "grad_norm": 5.392694473266602, + "learning_rate": 4.842592592592593e-06, + "loss": 1.0192, + "step": 6770 + }, + { + "epoch": 1.328890631125049, + "grad_norm": 5.902957439422607, + "learning_rate": 4.833333333333333e-06, + "loss": 1.0145, + "step": 6780 + }, + { + "epoch": 1.3308506468051744, + "grad_norm": 5.553819179534912, + "learning_rate": 4.824074074074075e-06, + "loss": 1.0091, + "step": 6790 + }, + { + "epoch": 1.3328106624852998, + "grad_norm": 6.013294219970703, + "learning_rate": 4.814814814814815e-06, + "loss": 1.0198, + "step": 6800 + }, + { + "epoch": 1.3347706781654254, + "grad_norm": 5.731135368347168, + "learning_rate": 4.805555555555556e-06, + "loss": 1.014, + "step": 6810 + }, + { + "epoch": 1.3367306938455508, + "grad_norm": 5.954074382781982, + "learning_rate": 4.796296296296297e-06, + "loss": 1.0128, + "step": 6820 + }, + { + "epoch": 1.3386907095256761, + "grad_norm": 5.550869464874268, + "learning_rate": 4.787037037037037e-06, + "loss": 1.0151, + "step": 6830 + }, + { + "epoch": 1.3406507252058018, + "grad_norm": 6.043259620666504, + "learning_rate": 4.777777777777778e-06, + "loss": 1.0069, + "step": 6840 + }, + { + "epoch": 1.3426107408859271, + "grad_norm": 5.002622127532959, + "learning_rate": 4.768518518518519e-06, + "loss": 1.0219, + "step": 6850 + }, + { + "epoch": 1.3445707565660525, + "grad_norm": 6.016299247741699, + "learning_rate": 4.75925925925926e-06, + "loss": 1.0158, + "step": 6860 + }, + { + "epoch": 1.346530772246178, + "grad_norm": 5.6830220222473145, + "learning_rate": 4.75e-06, + "loss": 1.0092, + "step": 6870 + }, + { + "epoch": 1.3484907879263033, + "grad_norm": 6.092140197753906, + "learning_rate": 4.7407407407407415e-06, + "loss": 1.0241, + "step": 6880 + }, + { + "epoch": 1.3504508036064289, + "grad_norm": 5.837907791137695, + "learning_rate": 4.731481481481482e-06, + "loss": 1.0259, + "step": 6890 + }, + { + "epoch": 1.3524108192865543, + "grad_norm": 5.598001480102539, + "learning_rate": 4.722222222222222e-06, + "loss": 1.0043, + "step": 6900 + }, + { + "epoch": 1.3543708349666796, + "grad_norm": 5.820865631103516, + "learning_rate": 4.712962962962963e-06, + "loss": 1.0325, + "step": 6910 + }, + { + "epoch": 1.3563308506468053, + "grad_norm": 6.162250518798828, + "learning_rate": 4.703703703703704e-06, + "loss": 1.0118, + "step": 6920 + }, + { + "epoch": 1.3582908663269306, + "grad_norm": 5.53606653213501, + "learning_rate": 4.694444444444445e-06, + "loss": 1.0099, + "step": 6930 + }, + { + "epoch": 1.360250882007056, + "grad_norm": 6.468924045562744, + "learning_rate": 4.6851851851851855e-06, + "loss": 1.0172, + "step": 6940 + }, + { + "epoch": 1.3622108976871816, + "grad_norm": 6.867157936096191, + "learning_rate": 4.675925925925927e-06, + "loss": 1.0021, + "step": 6950 + }, + { + "epoch": 1.364170913367307, + "grad_norm": 6.068783760070801, + "learning_rate": 4.666666666666667e-06, + "loss": 1.0083, + "step": 6960 + }, + { + "epoch": 1.3661309290474324, + "grad_norm": 5.502538204193115, + "learning_rate": 4.6574074074074076e-06, + "loss": 1.0173, + "step": 6970 + }, + { + "epoch": 1.3680909447275578, + "grad_norm": 5.83467960357666, + "learning_rate": 4.648148148148148e-06, + "loss": 1.031, + "step": 6980 + }, + { + "epoch": 1.3700509604076831, + "grad_norm": 5.383768558502197, + "learning_rate": 4.638888888888889e-06, + "loss": 1.0187, + "step": 6990 + }, + { + "epoch": 1.3720109760878088, + "grad_norm": 5.172204494476318, + "learning_rate": 4.62962962962963e-06, + "loss": 1.0079, + "step": 7000 + }, + { + "epoch": 1.3720109760878088, + "eval_loss": 1.0383639335632324, + "eval_runtime": 14.0609, + "eval_samples_per_second": 46.512, + "eval_steps_per_second": 5.832, + "step": 7000 + }, + { + "epoch": 1.3739709917679341, + "grad_norm": 5.811285495758057, + "learning_rate": 4.620370370370371e-06, + "loss": 1.0153, + "step": 7010 + }, + { + "epoch": 1.3759310074480595, + "grad_norm": 5.887601375579834, + "learning_rate": 4.611111111111112e-06, + "loss": 1.0104, + "step": 7020 + }, + { + "epoch": 1.3778910231281851, + "grad_norm": 5.245418548583984, + "learning_rate": 4.6018518518518524e-06, + "loss": 1.0063, + "step": 7030 + }, + { + "epoch": 1.3798510388083105, + "grad_norm": 6.191716194152832, + "learning_rate": 4.592592592592593e-06, + "loss": 1.0254, + "step": 7040 + }, + { + "epoch": 1.3818110544884359, + "grad_norm": 6.047815799713135, + "learning_rate": 4.583333333333333e-06, + "loss": 1.0291, + "step": 7050 + }, + { + "epoch": 1.3837710701685613, + "grad_norm": 5.58579158782959, + "learning_rate": 4.5740740740740745e-06, + "loss": 1.0126, + "step": 7060 + }, + { + "epoch": 1.3857310858486869, + "grad_norm": 5.651394844055176, + "learning_rate": 4.564814814814815e-06, + "loss": 1.013, + "step": 7070 + }, + { + "epoch": 1.3876911015288123, + "grad_norm": 5.703254699707031, + "learning_rate": 4.555555555555556e-06, + "loss": 1.0209, + "step": 7080 + }, + { + "epoch": 1.3896511172089376, + "grad_norm": 6.428845405578613, + "learning_rate": 4.5462962962962965e-06, + "loss": 1.0065, + "step": 7090 + }, + { + "epoch": 1.391611132889063, + "grad_norm": 5.854006767272949, + "learning_rate": 4.537037037037038e-06, + "loss": 1.0192, + "step": 7100 + }, + { + "epoch": 1.3935711485691886, + "grad_norm": 5.257528781890869, + "learning_rate": 4.527777777777778e-06, + "loss": 1.0125, + "step": 7110 + }, + { + "epoch": 1.395531164249314, + "grad_norm": 5.832418441772461, + "learning_rate": 4.5185185185185185e-06, + "loss": 1.0077, + "step": 7120 + }, + { + "epoch": 1.3974911799294394, + "grad_norm": 18.90093994140625, + "learning_rate": 4.50925925925926e-06, + "loss": 1.0039, + "step": 7130 + }, + { + "epoch": 1.399451195609565, + "grad_norm": 6.230194568634033, + "learning_rate": 4.5e-06, + "loss": 1.0168, + "step": 7140 + }, + { + "epoch": 1.4014112112896904, + "grad_norm": 5.916494846343994, + "learning_rate": 4.490740740740741e-06, + "loss": 1.0367, + "step": 7150 + }, + { + "epoch": 1.4033712269698158, + "grad_norm": 6.498172760009766, + "learning_rate": 4.481481481481482e-06, + "loss": 1.0226, + "step": 7160 + }, + { + "epoch": 1.4053312426499411, + "grad_norm": 5.681482791900635, + "learning_rate": 4.472222222222223e-06, + "loss": 1.0135, + "step": 7170 + }, + { + "epoch": 1.4072912583300665, + "grad_norm": 5.3359880447387695, + "learning_rate": 4.462962962962963e-06, + "loss": 1.0222, + "step": 7180 + }, + { + "epoch": 1.4092512740101921, + "grad_norm": 6.866293430328369, + "learning_rate": 4.453703703703704e-06, + "loss": 1.015, + "step": 7190 + }, + { + "epoch": 1.4112112896903175, + "grad_norm": 6.149731159210205, + "learning_rate": 4.444444444444444e-06, + "loss": 0.998, + "step": 7200 + }, + { + "epoch": 1.4131713053704429, + "grad_norm": 5.832232475280762, + "learning_rate": 4.435185185185185e-06, + "loss": 1.0097, + "step": 7210 + }, + { + "epoch": 1.4151313210505685, + "grad_norm": 5.8602986335754395, + "learning_rate": 4.425925925925927e-06, + "loss": 1.0133, + "step": 7220 + }, + { + "epoch": 1.4170913367306939, + "grad_norm": 6.716607093811035, + "learning_rate": 4.416666666666667e-06, + "loss": 1.0181, + "step": 7230 + }, + { + "epoch": 1.4190513524108193, + "grad_norm": 5.983255863189697, + "learning_rate": 4.407407407407408e-06, + "loss": 1.0152, + "step": 7240 + }, + { + "epoch": 1.4210113680909449, + "grad_norm": 5.979559898376465, + "learning_rate": 4.398148148148149e-06, + "loss": 1.0187, + "step": 7250 + }, + { + "epoch": 1.4229713837710702, + "grad_norm": 5.463419437408447, + "learning_rate": 4.388888888888889e-06, + "loss": 1.0018, + "step": 7260 + }, + { + "epoch": 1.4249313994511956, + "grad_norm": 5.58013916015625, + "learning_rate": 4.379629629629629e-06, + "loss": 1.0161, + "step": 7270 + }, + { + "epoch": 1.426891415131321, + "grad_norm": 5.896617412567139, + "learning_rate": 4.370370370370371e-06, + "loss": 1.0234, + "step": 7280 + }, + { + "epoch": 1.4288514308114464, + "grad_norm": 6.59088659286499, + "learning_rate": 4.361111111111112e-06, + "loss": 1.0116, + "step": 7290 + }, + { + "epoch": 1.430811446491572, + "grad_norm": 5.4909348487854, + "learning_rate": 4.351851851851852e-06, + "loss": 1.0241, + "step": 7300 + }, + { + "epoch": 1.4327714621716974, + "grad_norm": 5.706267356872559, + "learning_rate": 4.342592592592593e-06, + "loss": 1.0113, + "step": 7310 + }, + { + "epoch": 1.4347314778518228, + "grad_norm": 5.893515586853027, + "learning_rate": 4.333333333333334e-06, + "loss": 1.0074, + "step": 7320 + }, + { + "epoch": 1.4366914935319484, + "grad_norm": 5.337756156921387, + "learning_rate": 4.324074074074074e-06, + "loss": 1.0086, + "step": 7330 + }, + { + "epoch": 1.4386515092120737, + "grad_norm": 6.063902854919434, + "learning_rate": 4.314814814814815e-06, + "loss": 1.0135, + "step": 7340 + }, + { + "epoch": 1.4406115248921991, + "grad_norm": 5.862417221069336, + "learning_rate": 4.305555555555556e-06, + "loss": 1.0218, + "step": 7350 + }, + { + "epoch": 1.4425715405723245, + "grad_norm": 6.047614574432373, + "learning_rate": 4.296296296296296e-06, + "loss": 1.0147, + "step": 7360 + }, + { + "epoch": 1.4445315562524499, + "grad_norm": 7.018215656280518, + "learning_rate": 4.2870370370370376e-06, + "loss": 1.029, + "step": 7370 + }, + { + "epoch": 1.4464915719325755, + "grad_norm": 5.391064167022705, + "learning_rate": 4.277777777777778e-06, + "loss": 1.007, + "step": 7380 + }, + { + "epoch": 1.4484515876127009, + "grad_norm": 5.364874362945557, + "learning_rate": 4.268518518518519e-06, + "loss": 1.0188, + "step": 7390 + }, + { + "epoch": 1.4504116032928263, + "grad_norm": 5.990734577178955, + "learning_rate": 4.2592592592592596e-06, + "loss": 1.0003, + "step": 7400 + }, + { + "epoch": 1.4523716189729519, + "grad_norm": 5.727319717407227, + "learning_rate": 4.25e-06, + "loss": 1.0198, + "step": 7410 + }, + { + "epoch": 1.4543316346530772, + "grad_norm": 5.658199310302734, + "learning_rate": 4.240740740740741e-06, + "loss": 1.0221, + "step": 7420 + }, + { + "epoch": 1.4562916503332026, + "grad_norm": 5.460375785827637, + "learning_rate": 4.231481481481482e-06, + "loss": 1.0127, + "step": 7430 + }, + { + "epoch": 1.4582516660133282, + "grad_norm": 6.161991596221924, + "learning_rate": 4.222222222222223e-06, + "loss": 1.0025, + "step": 7440 + }, + { + "epoch": 1.4602116816934536, + "grad_norm": 5.7905192375183105, + "learning_rate": 4.212962962962963e-06, + "loss": 1.0071, + "step": 7450 + }, + { + "epoch": 1.462171697373579, + "grad_norm": 5.652228832244873, + "learning_rate": 4.2037037037037045e-06, + "loss": 1.0112, + "step": 7460 + }, + { + "epoch": 1.4641317130537044, + "grad_norm": 5.817644119262695, + "learning_rate": 4.194444444444445e-06, + "loss": 1.0191, + "step": 7470 + }, + { + "epoch": 1.4660917287338298, + "grad_norm": 6.155873775482178, + "learning_rate": 4.185185185185185e-06, + "loss": 1.013, + "step": 7480 + }, + { + "epoch": 1.4680517444139554, + "grad_norm": 5.775312423706055, + "learning_rate": 4.175925925925926e-06, + "loss": 1.0107, + "step": 7490 + }, + { + "epoch": 1.4700117600940807, + "grad_norm": 5.765095233917236, + "learning_rate": 4.166666666666667e-06, + "loss": 1.0151, + "step": 7500 + }, + { + "epoch": 1.4719717757742061, + "grad_norm": 5.336995601654053, + "learning_rate": 4.157407407407408e-06, + "loss": 1.0048, + "step": 7510 + }, + { + "epoch": 1.4739317914543317, + "grad_norm": 6.261089324951172, + "learning_rate": 4.1481481481481485e-06, + "loss": 1.0002, + "step": 7520 + }, + { + "epoch": 1.4758918071344571, + "grad_norm": 5.8046746253967285, + "learning_rate": 4.138888888888889e-06, + "loss": 1.0079, + "step": 7530 + }, + { + "epoch": 1.4778518228145825, + "grad_norm": 5.5930867195129395, + "learning_rate": 4.12962962962963e-06, + "loss": 1.0094, + "step": 7540 + }, + { + "epoch": 1.4798118384947079, + "grad_norm": 5.929795742034912, + "learning_rate": 4.1203703703703705e-06, + "loss": 1.0122, + "step": 7550 + }, + { + "epoch": 1.4817718541748335, + "grad_norm": 5.746526718139648, + "learning_rate": 4.111111111111111e-06, + "loss": 1.0186, + "step": 7560 + }, + { + "epoch": 1.4837318698549589, + "grad_norm": 6.316400051116943, + "learning_rate": 4.101851851851852e-06, + "loss": 1.0019, + "step": 7570 + }, + { + "epoch": 1.4856918855350842, + "grad_norm": 5.561999320983887, + "learning_rate": 4.092592592592593e-06, + "loss": 1.0157, + "step": 7580 + }, + { + "epoch": 1.4876519012152096, + "grad_norm": 6.012383937835693, + "learning_rate": 4.083333333333334e-06, + "loss": 1.0108, + "step": 7590 + }, + { + "epoch": 1.4896119168953352, + "grad_norm": 5.594381809234619, + "learning_rate": 4.074074074074074e-06, + "loss": 0.9917, + "step": 7600 + }, + { + "epoch": 1.4915719325754606, + "grad_norm": 6.158176898956299, + "learning_rate": 4.064814814814815e-06, + "loss": 1.0063, + "step": 7610 + }, + { + "epoch": 1.493531948255586, + "grad_norm": 5.4881510734558105, + "learning_rate": 4.055555555555556e-06, + "loss": 1.0122, + "step": 7620 + }, + { + "epoch": 1.4954919639357116, + "grad_norm": 6.7185869216918945, + "learning_rate": 4.046296296296296e-06, + "loss": 0.9927, + "step": 7630 + }, + { + "epoch": 1.497451979615837, + "grad_norm": 6.015646934509277, + "learning_rate": 4.037037037037037e-06, + "loss": 1.0054, + "step": 7640 + }, + { + "epoch": 1.4994119952959624, + "grad_norm": 5.855646133422852, + "learning_rate": 4.027777777777779e-06, + "loss": 0.9988, + "step": 7650 + }, + { + "epoch": 1.501372010976088, + "grad_norm": 6.158780574798584, + "learning_rate": 4.018518518518519e-06, + "loss": 1.027, + "step": 7660 + }, + { + "epoch": 1.5033320266562131, + "grad_norm": 5.943362712860107, + "learning_rate": 4.0092592592592594e-06, + "loss": 1.0285, + "step": 7670 + }, + { + "epoch": 1.5052920423363387, + "grad_norm": 5.775484085083008, + "learning_rate": 4.000000000000001e-06, + "loss": 1.0093, + "step": 7680 + }, + { + "epoch": 1.5072520580164641, + "grad_norm": 6.551724910736084, + "learning_rate": 3.990740740740741e-06, + "loss": 1.0185, + "step": 7690 + }, + { + "epoch": 1.5092120736965895, + "grad_norm": 5.986613750457764, + "learning_rate": 3.9814814814814814e-06, + "loss": 1.0058, + "step": 7700 + }, + { + "epoch": 1.5092120736965895, + "eval_loss": 1.0326706171035767, + "eval_runtime": 13.9953, + "eval_samples_per_second": 46.73, + "eval_steps_per_second": 5.859, + "step": 7700 + }, + { + "epoch": 1.511172089376715, + "grad_norm": 5.968577861785889, + "learning_rate": 3.972222222222223e-06, + "loss": 1.0111, + "step": 7710 + }, + { + "epoch": 1.5131321050568405, + "grad_norm": 5.956306457519531, + "learning_rate": 3.962962962962963e-06, + "loss": 0.9822, + "step": 7720 + }, + { + "epoch": 1.5150921207369659, + "grad_norm": 6.391535758972168, + "learning_rate": 3.953703703703704e-06, + "loss": 1.0073, + "step": 7730 + }, + { + "epoch": 1.5170521364170915, + "grad_norm": 6.031282901763916, + "learning_rate": 3.944444444444445e-06, + "loss": 0.9992, + "step": 7740 + }, + { + "epoch": 1.5190121520972166, + "grad_norm": 6.070298194885254, + "learning_rate": 3.935185185185186e-06, + "loss": 0.9995, + "step": 7750 + }, + { + "epoch": 1.5209721677773422, + "grad_norm": 6.470989227294922, + "learning_rate": 3.925925925925926e-06, + "loss": 1.0041, + "step": 7760 + }, + { + "epoch": 1.5229321834574676, + "grad_norm": 6.339069843292236, + "learning_rate": 3.916666666666667e-06, + "loss": 1.0145, + "step": 7770 + }, + { + "epoch": 1.524892199137593, + "grad_norm": 6.204465866088867, + "learning_rate": 3.907407407407408e-06, + "loss": 1.0202, + "step": 7780 + }, + { + "epoch": 1.5268522148177186, + "grad_norm": 5.39393424987793, + "learning_rate": 3.898148148148148e-06, + "loss": 1.0144, + "step": 7790 + }, + { + "epoch": 1.528812230497844, + "grad_norm": 6.083856105804443, + "learning_rate": 3.88888888888889e-06, + "loss": 1.0087, + "step": 7800 + }, + { + "epoch": 1.5307722461779694, + "grad_norm": 6.141456604003906, + "learning_rate": 3.87962962962963e-06, + "loss": 1.0189, + "step": 7810 + }, + { + "epoch": 1.532732261858095, + "grad_norm": 6.509696960449219, + "learning_rate": 3.87037037037037e-06, + "loss": 1.0136, + "step": 7820 + }, + { + "epoch": 1.5346922775382204, + "grad_norm": 5.372650146484375, + "learning_rate": 3.861111111111112e-06, + "loss": 1.0009, + "step": 7830 + }, + { + "epoch": 1.5366522932183457, + "grad_norm": 5.931579113006592, + "learning_rate": 3.851851851851852e-06, + "loss": 1.0144, + "step": 7840 + }, + { + "epoch": 1.5386123088984713, + "grad_norm": 6.2442626953125, + "learning_rate": 3.842592592592592e-06, + "loss": 1.0132, + "step": 7850 + }, + { + "epoch": 1.5405723245785965, + "grad_norm": 6.475978851318359, + "learning_rate": 3.833333333333334e-06, + "loss": 1.0058, + "step": 7860 + }, + { + "epoch": 1.542532340258722, + "grad_norm": 5.5199103355407715, + "learning_rate": 3.824074074074075e-06, + "loss": 1.0017, + "step": 7870 + }, + { + "epoch": 1.5444923559388475, + "grad_norm": 5.857003211975098, + "learning_rate": 3.814814814814815e-06, + "loss": 1.0108, + "step": 7880 + }, + { + "epoch": 1.5464523716189729, + "grad_norm": 5.81916618347168, + "learning_rate": 3.8055555555555556e-06, + "loss": 1.0185, + "step": 7890 + }, + { + "epoch": 1.5484123872990985, + "grad_norm": 6.099959850311279, + "learning_rate": 3.796296296296297e-06, + "loss": 1.0109, + "step": 7900 + }, + { + "epoch": 1.5503724029792239, + "grad_norm": 5.837829113006592, + "learning_rate": 3.7870370370370373e-06, + "loss": 1.0004, + "step": 7910 + }, + { + "epoch": 1.5523324186593492, + "grad_norm": 6.274796485900879, + "learning_rate": 3.777777777777778e-06, + "loss": 1.0165, + "step": 7920 + }, + { + "epoch": 1.5542924343394748, + "grad_norm": 5.926800727844238, + "learning_rate": 3.7685185185185185e-06, + "loss": 1.0067, + "step": 7930 + }, + { + "epoch": 1.5562524500196, + "grad_norm": 6.2746381759643555, + "learning_rate": 3.7592592592592597e-06, + "loss": 1.0078, + "step": 7940 + }, + { + "epoch": 1.5582124656997256, + "grad_norm": 6.315319538116455, + "learning_rate": 3.7500000000000005e-06, + "loss": 1.0085, + "step": 7950 + }, + { + "epoch": 1.560172481379851, + "grad_norm": 6.36542272567749, + "learning_rate": 3.740740740740741e-06, + "loss": 1.0073, + "step": 7960 + }, + { + "epoch": 1.5621324970599764, + "grad_norm": 6.047231197357178, + "learning_rate": 3.731481481481482e-06, + "loss": 1.0019, + "step": 7970 + }, + { + "epoch": 1.564092512740102, + "grad_norm": 6.49619722366333, + "learning_rate": 3.7222222222222225e-06, + "loss": 1.0126, + "step": 7980 + }, + { + "epoch": 1.5660525284202274, + "grad_norm": 6.945578098297119, + "learning_rate": 3.7129629629629633e-06, + "loss": 1.0141, + "step": 7990 + }, + { + "epoch": 1.5680125441003527, + "grad_norm": 5.595407485961914, + "learning_rate": 3.7037037037037037e-06, + "loss": 1.0077, + "step": 8000 + }, + { + "epoch": 1.5699725597804783, + "grad_norm": 6.3146867752075195, + "learning_rate": 3.694444444444445e-06, + "loss": 1.0178, + "step": 8010 + }, + { + "epoch": 1.5719325754606037, + "grad_norm": 8.867284774780273, + "learning_rate": 3.6851851851851854e-06, + "loss": 0.9974, + "step": 8020 + }, + { + "epoch": 1.573892591140729, + "grad_norm": 5.997375011444092, + "learning_rate": 3.675925925925926e-06, + "loss": 1.0169, + "step": 8030 + }, + { + "epoch": 1.5758526068208547, + "grad_norm": 6.150747776031494, + "learning_rate": 3.6666666666666666e-06, + "loss": 1.0087, + "step": 8040 + }, + { + "epoch": 1.5778126225009799, + "grad_norm": 6.344914436340332, + "learning_rate": 3.657407407407408e-06, + "loss": 1.008, + "step": 8050 + }, + { + "epoch": 1.5797726381811055, + "grad_norm": 5.457338333129883, + "learning_rate": 3.6481481481481486e-06, + "loss": 1.0169, + "step": 8060 + }, + { + "epoch": 1.5817326538612309, + "grad_norm": 6.128787994384766, + "learning_rate": 3.638888888888889e-06, + "loss": 1.0104, + "step": 8070 + }, + { + "epoch": 1.5836926695413562, + "grad_norm": 6.624429702758789, + "learning_rate": 3.6296296296296302e-06, + "loss": 0.9982, + "step": 8080 + }, + { + "epoch": 1.5856526852214818, + "grad_norm": 5.785494327545166, + "learning_rate": 3.6203703703703706e-06, + "loss": 0.9968, + "step": 8090 + }, + { + "epoch": 1.5876127009016072, + "grad_norm": 6.378090858459473, + "learning_rate": 3.6111111111111115e-06, + "loss": 0.9972, + "step": 8100 + }, + { + "epoch": 1.5895727165817326, + "grad_norm": 7.523343086242676, + "learning_rate": 3.601851851851852e-06, + "loss": 1.0077, + "step": 8110 + }, + { + "epoch": 1.5915327322618582, + "grad_norm": 5.578686714172363, + "learning_rate": 3.592592592592593e-06, + "loss": 1.0103, + "step": 8120 + }, + { + "epoch": 1.5934927479419836, + "grad_norm": 6.031825065612793, + "learning_rate": 3.5833333333333335e-06, + "loss": 1.015, + "step": 8130 + }, + { + "epoch": 1.595452763622109, + "grad_norm": 6.009352207183838, + "learning_rate": 3.5740740740740743e-06, + "loss": 0.9997, + "step": 8140 + }, + { + "epoch": 1.5974127793022346, + "grad_norm": 6.1863694190979, + "learning_rate": 3.5648148148148147e-06, + "loss": 0.9902, + "step": 8150 + }, + { + "epoch": 1.5993727949823597, + "grad_norm": 5.477410316467285, + "learning_rate": 3.555555555555556e-06, + "loss": 1.0077, + "step": 8160 + }, + { + "epoch": 1.6013328106624853, + "grad_norm": 6.446470260620117, + "learning_rate": 3.5462962962962967e-06, + "loss": 1.0123, + "step": 8170 + }, + { + "epoch": 1.6032928263426107, + "grad_norm": 5.979601860046387, + "learning_rate": 3.537037037037037e-06, + "loss": 1.0106, + "step": 8180 + }, + { + "epoch": 1.605252842022736, + "grad_norm": 5.748382091522217, + "learning_rate": 3.5277777777777784e-06, + "loss": 1.0083, + "step": 8190 + }, + { + "epoch": 1.6072128577028617, + "grad_norm": 6.122110843658447, + "learning_rate": 3.5185185185185187e-06, + "loss": 1.0124, + "step": 8200 + }, + { + "epoch": 1.609172873382987, + "grad_norm": 6.104898452758789, + "learning_rate": 3.5092592592592596e-06, + "loss": 1.0118, + "step": 8210 + }, + { + "epoch": 1.6111328890631125, + "grad_norm": 6.527801513671875, + "learning_rate": 3.5e-06, + "loss": 1.0084, + "step": 8220 + }, + { + "epoch": 1.613092904743238, + "grad_norm": 6.3016676902771, + "learning_rate": 3.490740740740741e-06, + "loss": 1.0104, + "step": 8230 + }, + { + "epoch": 1.6150529204233632, + "grad_norm": 6.44483757019043, + "learning_rate": 3.481481481481482e-06, + "loss": 1.0102, + "step": 8240 + }, + { + "epoch": 1.6170129361034888, + "grad_norm": 6.052326679229736, + "learning_rate": 3.4722222222222224e-06, + "loss": 1.0147, + "step": 8250 + }, + { + "epoch": 1.6189729517836142, + "grad_norm": 5.94256067276001, + "learning_rate": 3.4629629629629628e-06, + "loss": 1.0108, + "step": 8260 + }, + { + "epoch": 1.6209329674637396, + "grad_norm": 6.634081840515137, + "learning_rate": 3.453703703703704e-06, + "loss": 1.0139, + "step": 8270 + }, + { + "epoch": 1.6228929831438652, + "grad_norm": 6.239657878875732, + "learning_rate": 3.444444444444445e-06, + "loss": 0.9949, + "step": 8280 + }, + { + "epoch": 1.6248529988239906, + "grad_norm": 6.047982692718506, + "learning_rate": 3.4351851851851852e-06, + "loss": 1.0223, + "step": 8290 + }, + { + "epoch": 1.626813014504116, + "grad_norm": 5.618074417114258, + "learning_rate": 3.4259259259259265e-06, + "loss": 1.022, + "step": 8300 + }, + { + "epoch": 1.6287730301842416, + "grad_norm": 5.856754779815674, + "learning_rate": 3.416666666666667e-06, + "loss": 1.0053, + "step": 8310 + }, + { + "epoch": 1.630733045864367, + "grad_norm": 6.435812473297119, + "learning_rate": 3.4074074074074077e-06, + "loss": 0.9989, + "step": 8320 + }, + { + "epoch": 1.6326930615444923, + "grad_norm": 5.980440139770508, + "learning_rate": 3.398148148148148e-06, + "loss": 1.0108, + "step": 8330 + }, + { + "epoch": 1.634653077224618, + "grad_norm": 6.19628381729126, + "learning_rate": 3.3888888888888893e-06, + "loss": 0.995, + "step": 8340 + }, + { + "epoch": 1.636613092904743, + "grad_norm": 5.549736976623535, + "learning_rate": 3.37962962962963e-06, + "loss": 1.0181, + "step": 8350 + }, + { + "epoch": 1.6385731085848687, + "grad_norm": 6.485236644744873, + "learning_rate": 3.3703703703703705e-06, + "loss": 1.0092, + "step": 8360 + }, + { + "epoch": 1.640533124264994, + "grad_norm": 6.0221848487854, + "learning_rate": 3.3611111111111117e-06, + "loss": 1.0038, + "step": 8370 + }, + { + "epoch": 1.6424931399451195, + "grad_norm": 5.699035167694092, + "learning_rate": 3.351851851851852e-06, + "loss": 1.0092, + "step": 8380 + }, + { + "epoch": 1.644453155625245, + "grad_norm": 6.401322364807129, + "learning_rate": 3.342592592592593e-06, + "loss": 1.0013, + "step": 8390 + }, + { + "epoch": 1.6464131713053705, + "grad_norm": 6.202515125274658, + "learning_rate": 3.3333333333333333e-06, + "loss": 1.0165, + "step": 8400 + }, + { + "epoch": 1.6464131713053705, + "eval_loss": 1.0264891386032104, + "eval_runtime": 14.05, + "eval_samples_per_second": 46.548, + "eval_steps_per_second": 5.836, + "step": 8400 + }, + { + "epoch": 1.6483731869854958, + "grad_norm": 6.184670448303223, + "learning_rate": 3.3240740740740746e-06, + "loss": 1.0055, + "step": 8410 + }, + { + "epoch": 1.6503332026656214, + "grad_norm": 6.040178298950195, + "learning_rate": 3.314814814814815e-06, + "loss": 1.007, + "step": 8420 + }, + { + "epoch": 1.6522932183457466, + "grad_norm": 6.925061225891113, + "learning_rate": 3.3055555555555558e-06, + "loss": 1.0101, + "step": 8430 + }, + { + "epoch": 1.6542532340258722, + "grad_norm": 6.050986289978027, + "learning_rate": 3.296296296296296e-06, + "loss": 0.9969, + "step": 8440 + }, + { + "epoch": 1.6562132497059978, + "grad_norm": 5.321831703186035, + "learning_rate": 3.2870370370370374e-06, + "loss": 1.0092, + "step": 8450 + }, + { + "epoch": 1.658173265386123, + "grad_norm": 5.440800666809082, + "learning_rate": 3.277777777777778e-06, + "loss": 1.0115, + "step": 8460 + }, + { + "epoch": 1.6601332810662486, + "grad_norm": 8.311331748962402, + "learning_rate": 3.2685185185185186e-06, + "loss": 1.0014, + "step": 8470 + }, + { + "epoch": 1.662093296746374, + "grad_norm": 6.014708518981934, + "learning_rate": 3.25925925925926e-06, + "loss": 1.0135, + "step": 8480 + }, + { + "epoch": 1.6640533124264993, + "grad_norm": 6.3742475509643555, + "learning_rate": 3.2500000000000002e-06, + "loss": 1.0024, + "step": 8490 + }, + { + "epoch": 1.666013328106625, + "grad_norm": 6.1811041831970215, + "learning_rate": 3.240740740740741e-06, + "loss": 0.9983, + "step": 8500 + }, + { + "epoch": 1.6679733437867503, + "grad_norm": 6.055778503417969, + "learning_rate": 3.2314814814814814e-06, + "loss": 0.9986, + "step": 8510 + }, + { + "epoch": 1.6699333594668757, + "grad_norm": 6.654355525970459, + "learning_rate": 3.2222222222222227e-06, + "loss": 1.0111, + "step": 8520 + }, + { + "epoch": 1.6718933751470013, + "grad_norm": 6.156193733215332, + "learning_rate": 3.2129629629629635e-06, + "loss": 0.9879, + "step": 8530 + }, + { + "epoch": 1.6738533908271265, + "grad_norm": 6.491064548492432, + "learning_rate": 3.203703703703704e-06, + "loss": 1.0197, + "step": 8540 + }, + { + "epoch": 1.675813406507252, + "grad_norm": 5.876689434051514, + "learning_rate": 3.1944444444444443e-06, + "loss": 0.9934, + "step": 8550 + }, + { + "epoch": 1.6777734221873775, + "grad_norm": 6.113574504852295, + "learning_rate": 3.1851851851851855e-06, + "loss": 0.9992, + "step": 8560 + }, + { + "epoch": 1.6797334378675028, + "grad_norm": 6.2524518966674805, + "learning_rate": 3.1759259259259263e-06, + "loss": 1.0101, + "step": 8570 + }, + { + "epoch": 1.6816934535476284, + "grad_norm": 6.034154891967773, + "learning_rate": 3.1666666666666667e-06, + "loss": 1.0009, + "step": 8580 + }, + { + "epoch": 1.6836534692277538, + "grad_norm": 5.920796871185303, + "learning_rate": 3.157407407407408e-06, + "loss": 1.0088, + "step": 8590 + }, + { + "epoch": 1.6856134849078792, + "grad_norm": 5.578927993774414, + "learning_rate": 3.1481481481481483e-06, + "loss": 1.0004, + "step": 8600 + }, + { + "epoch": 1.6875735005880048, + "grad_norm": 6.1217803955078125, + "learning_rate": 3.138888888888889e-06, + "loss": 1.0034, + "step": 8610 + }, + { + "epoch": 1.6895335162681302, + "grad_norm": 6.209445953369141, + "learning_rate": 3.1296296296296295e-06, + "loss": 0.9912, + "step": 8620 + }, + { + "epoch": 1.6914935319482556, + "grad_norm": 6.419200897216797, + "learning_rate": 3.1203703703703708e-06, + "loss": 1.0048, + "step": 8630 + }, + { + "epoch": 1.6934535476283812, + "grad_norm": 6.344346046447754, + "learning_rate": 3.1111111111111116e-06, + "loss": 0.9986, + "step": 8640 + }, + { + "epoch": 1.6954135633085063, + "grad_norm": 6.1742963790893555, + "learning_rate": 3.101851851851852e-06, + "loss": 1.0085, + "step": 8650 + }, + { + "epoch": 1.697373578988632, + "grad_norm": 5.269043445587158, + "learning_rate": 3.0925925925925928e-06, + "loss": 1.0045, + "step": 8660 + }, + { + "epoch": 1.6993335946687573, + "grad_norm": 6.707599639892578, + "learning_rate": 3.0833333333333336e-06, + "loss": 1.0074, + "step": 8670 + }, + { + "epoch": 1.7012936103488827, + "grad_norm": 6.476319313049316, + "learning_rate": 3.0740740740740744e-06, + "loss": 1.0066, + "step": 8680 + }, + { + "epoch": 1.7032536260290083, + "grad_norm": 5.808709621429443, + "learning_rate": 3.064814814814815e-06, + "loss": 0.9969, + "step": 8690 + }, + { + "epoch": 1.7052136417091337, + "grad_norm": 10.500265121459961, + "learning_rate": 3.055555555555556e-06, + "loss": 1.0062, + "step": 8700 + }, + { + "epoch": 1.707173657389259, + "grad_norm": 5.992859840393066, + "learning_rate": 3.0462962962962964e-06, + "loss": 1.0133, + "step": 8710 + }, + { + "epoch": 1.7091336730693847, + "grad_norm": 6.0215325355529785, + "learning_rate": 3.0370370370370372e-06, + "loss": 1.0175, + "step": 8720 + }, + { + "epoch": 1.7110936887495098, + "grad_norm": 6.671189785003662, + "learning_rate": 3.0277777777777776e-06, + "loss": 1.0113, + "step": 8730 + }, + { + "epoch": 1.7130537044296354, + "grad_norm": 6.012123107910156, + "learning_rate": 3.018518518518519e-06, + "loss": 0.9931, + "step": 8740 + }, + { + "epoch": 1.7150137201097608, + "grad_norm": 5.328975200653076, + "learning_rate": 3.0092592592592597e-06, + "loss": 1.0052, + "step": 8750 + }, + { + "epoch": 1.7169737357898862, + "grad_norm": 6.184416770935059, + "learning_rate": 3e-06, + "loss": 1.0125, + "step": 8760 + }, + { + "epoch": 1.7189337514700118, + "grad_norm": 6.222989559173584, + "learning_rate": 2.990740740740741e-06, + "loss": 1.0166, + "step": 8770 + }, + { + "epoch": 1.7208937671501372, + "grad_norm": 6.706179141998291, + "learning_rate": 2.9814814814814817e-06, + "loss": 1.0095, + "step": 8780 + }, + { + "epoch": 1.7228537828302626, + "grad_norm": 6.823479652404785, + "learning_rate": 2.9722222222222225e-06, + "loss": 1.0147, + "step": 8790 + }, + { + "epoch": 1.7248137985103882, + "grad_norm": 7.233177185058594, + "learning_rate": 2.962962962962963e-06, + "loss": 0.9963, + "step": 8800 + }, + { + "epoch": 1.7267738141905136, + "grad_norm": 5.568224906921387, + "learning_rate": 2.953703703703704e-06, + "loss": 1.0052, + "step": 8810 + }, + { + "epoch": 1.728733829870639, + "grad_norm": 6.352148532867432, + "learning_rate": 2.944444444444445e-06, + "loss": 0.9951, + "step": 8820 + }, + { + "epoch": 1.7306938455507646, + "grad_norm": 6.312885284423828, + "learning_rate": 2.9351851851851853e-06, + "loss": 1.0126, + "step": 8830 + }, + { + "epoch": 1.7326538612308897, + "grad_norm": 6.648190975189209, + "learning_rate": 2.9259259259259257e-06, + "loss": 1.0095, + "step": 8840 + }, + { + "epoch": 1.7346138769110153, + "grad_norm": 7.448049545288086, + "learning_rate": 2.916666666666667e-06, + "loss": 1.0058, + "step": 8850 + }, + { + "epoch": 1.7365738925911407, + "grad_norm": 6.0499653816223145, + "learning_rate": 2.907407407407408e-06, + "loss": 1.002, + "step": 8860 + }, + { + "epoch": 1.738533908271266, + "grad_norm": 6.453138828277588, + "learning_rate": 2.898148148148148e-06, + "loss": 1.0107, + "step": 8870 + }, + { + "epoch": 1.7404939239513917, + "grad_norm": 6.5974555015563965, + "learning_rate": 2.888888888888889e-06, + "loss": 1.0119, + "step": 8880 + }, + { + "epoch": 1.742453939631517, + "grad_norm": 6.174962043762207, + "learning_rate": 2.87962962962963e-06, + "loss": 1.0067, + "step": 8890 + }, + { + "epoch": 1.7444139553116425, + "grad_norm": 6.343291282653809, + "learning_rate": 2.8703703703703706e-06, + "loss": 0.9903, + "step": 8900 + }, + { + "epoch": 1.746373970991768, + "grad_norm": 6.71571159362793, + "learning_rate": 2.861111111111111e-06, + "loss": 1.018, + "step": 8910 + }, + { + "epoch": 1.7483339866718932, + "grad_norm": 6.948620319366455, + "learning_rate": 2.8518518518518522e-06, + "loss": 1.0138, + "step": 8920 + }, + { + "epoch": 1.7502940023520188, + "grad_norm": 6.206993579864502, + "learning_rate": 2.842592592592593e-06, + "loss": 1.0024, + "step": 8930 + }, + { + "epoch": 1.7522540180321444, + "grad_norm": 6.701568603515625, + "learning_rate": 2.8333333333333335e-06, + "loss": 1.0117, + "step": 8940 + }, + { + "epoch": 1.7542140337122696, + "grad_norm": 6.476100921630859, + "learning_rate": 2.8240740740740743e-06, + "loss": 1.0, + "step": 8950 + }, + { + "epoch": 1.7561740493923952, + "grad_norm": 7.1540703773498535, + "learning_rate": 2.814814814814815e-06, + "loss": 0.9933, + "step": 8960 + }, + { + "epoch": 1.7581340650725206, + "grad_norm": 6.764838695526123, + "learning_rate": 2.805555555555556e-06, + "loss": 1.0051, + "step": 8970 + }, + { + "epoch": 1.760094080752646, + "grad_norm": 5.835758209228516, + "learning_rate": 2.7962962962962963e-06, + "loss": 1.0058, + "step": 8980 + }, + { + "epoch": 1.7620540964327716, + "grad_norm": 6.640206813812256, + "learning_rate": 2.7870370370370375e-06, + "loss": 1.0012, + "step": 8990 + }, + { + "epoch": 1.764014112112897, + "grad_norm": 7.032010555267334, + "learning_rate": 2.7777777777777783e-06, + "loss": 0.9988, + "step": 9000 + }, + { + "epoch": 1.7659741277930223, + "grad_norm": 6.398713111877441, + "learning_rate": 2.7685185185185187e-06, + "loss": 1.0179, + "step": 9010 + }, + { + "epoch": 1.767934143473148, + "grad_norm": 6.625948905944824, + "learning_rate": 2.759259259259259e-06, + "loss": 1.0089, + "step": 9020 + }, + { + "epoch": 1.769894159153273, + "grad_norm": 5.858684062957764, + "learning_rate": 2.7500000000000004e-06, + "loss": 1.0057, + "step": 9030 + }, + { + "epoch": 1.7718541748333987, + "grad_norm": 7.03538703918457, + "learning_rate": 2.740740740740741e-06, + "loss": 1.0223, + "step": 9040 + }, + { + "epoch": 1.773814190513524, + "grad_norm": 6.332048416137695, + "learning_rate": 2.7314814814814816e-06, + "loss": 1.0027, + "step": 9050 + }, + { + "epoch": 1.7757742061936495, + "grad_norm": 6.304274082183838, + "learning_rate": 2.7222222222222224e-06, + "loss": 0.994, + "step": 9060 + }, + { + "epoch": 1.777734221873775, + "grad_norm": 5.588042736053467, + "learning_rate": 2.712962962962963e-06, + "loss": 1.0044, + "step": 9070 + }, + { + "epoch": 1.7796942375539004, + "grad_norm": 6.531040191650391, + "learning_rate": 2.703703703703704e-06, + "loss": 1.0059, + "step": 9080 + }, + { + "epoch": 1.7816542532340258, + "grad_norm": 6.511843681335449, + "learning_rate": 2.6944444444444444e-06, + "loss": 1.0046, + "step": 9090 + }, + { + "epoch": 1.7836142689141514, + "grad_norm": 6.718422889709473, + "learning_rate": 2.6851851851851856e-06, + "loss": 1.0084, + "step": 9100 + }, + { + "epoch": 1.7836142689141514, + "eval_loss": 1.0242658853530884, + "eval_runtime": 14.0168, + "eval_samples_per_second": 46.658, + "eval_steps_per_second": 5.85, + "step": 9100 + }, + { + "epoch": 1.7855742845942768, + "grad_norm": 5.511220455169678, + "learning_rate": 2.6759259259259264e-06, + "loss": 1.0028, + "step": 9110 + }, + { + "epoch": 1.7875343002744022, + "grad_norm": 6.125180244445801, + "learning_rate": 2.666666666666667e-06, + "loss": 0.9937, + "step": 9120 + }, + { + "epoch": 1.7894943159545278, + "grad_norm": 6.439998626708984, + "learning_rate": 2.6574074074074076e-06, + "loss": 0.9946, + "step": 9130 + }, + { + "epoch": 1.791454331634653, + "grad_norm": 6.341091156005859, + "learning_rate": 2.6481481481481485e-06, + "loss": 1.0131, + "step": 9140 + }, + { + "epoch": 1.7934143473147786, + "grad_norm": 6.909265041351318, + "learning_rate": 2.6388888888888893e-06, + "loss": 1.019, + "step": 9150 + }, + { + "epoch": 1.795374362994904, + "grad_norm": 5.798973083496094, + "learning_rate": 2.6296296296296297e-06, + "loss": 1.0058, + "step": 9160 + }, + { + "epoch": 1.7973343786750293, + "grad_norm": 6.564090728759766, + "learning_rate": 2.6203703703703705e-06, + "loss": 1.0012, + "step": 9170 + }, + { + "epoch": 1.799294394355155, + "grad_norm": 6.725678443908691, + "learning_rate": 2.6111111111111113e-06, + "loss": 0.9987, + "step": 9180 + }, + { + "epoch": 1.8012544100352803, + "grad_norm": 6.482659339904785, + "learning_rate": 2.601851851851852e-06, + "loss": 0.9986, + "step": 9190 + }, + { + "epoch": 1.8032144257154057, + "grad_norm": 10.73702621459961, + "learning_rate": 2.5925925925925925e-06, + "loss": 0.9968, + "step": 9200 + }, + { + "epoch": 1.8051744413955313, + "grad_norm": 6.893209457397461, + "learning_rate": 2.5833333333333337e-06, + "loss": 1.0111, + "step": 9210 + }, + { + "epoch": 1.8071344570756565, + "grad_norm": 6.212709426879883, + "learning_rate": 2.5740740740740745e-06, + "loss": 1.0036, + "step": 9220 + }, + { + "epoch": 1.809094472755782, + "grad_norm": 7.134426116943359, + "learning_rate": 2.564814814814815e-06, + "loss": 1.0008, + "step": 9230 + }, + { + "epoch": 1.8110544884359074, + "grad_norm": 12.403793334960938, + "learning_rate": 2.5555555555555557e-06, + "loss": 0.9933, + "step": 9240 + }, + { + "epoch": 1.8130145041160328, + "grad_norm": 6.31683349609375, + "learning_rate": 2.5462962962962966e-06, + "loss": 1.0083, + "step": 9250 + }, + { + "epoch": 1.8149745197961584, + "grad_norm": 5.857621669769287, + "learning_rate": 2.5370370370370374e-06, + "loss": 0.993, + "step": 9260 + }, + { + "epoch": 1.8169345354762838, + "grad_norm": 5.9981584548950195, + "learning_rate": 2.5277777777777778e-06, + "loss": 1.008, + "step": 9270 + }, + { + "epoch": 1.8188945511564092, + "grad_norm": 6.6797404289245605, + "learning_rate": 2.5185185185185186e-06, + "loss": 0.9915, + "step": 9280 + }, + { + "epoch": 1.8208545668365348, + "grad_norm": 6.050951957702637, + "learning_rate": 2.50925925925926e-06, + "loss": 1.0248, + "step": 9290 + }, + { + "epoch": 1.8228145825166602, + "grad_norm": 6.692429542541504, + "learning_rate": 2.5e-06, + "loss": 1.0103, + "step": 9300 + }, + { + "epoch": 1.8247745981967856, + "grad_norm": 6.396121978759766, + "learning_rate": 2.490740740740741e-06, + "loss": 1.0026, + "step": 9310 + }, + { + "epoch": 1.8267346138769112, + "grad_norm": 6.65963077545166, + "learning_rate": 2.481481481481482e-06, + "loss": 0.9912, + "step": 9320 + }, + { + "epoch": 1.8286946295570363, + "grad_norm": 7.833820343017578, + "learning_rate": 2.4722222222222226e-06, + "loss": 0.994, + "step": 9330 + }, + { + "epoch": 1.830654645237162, + "grad_norm": 5.663256645202637, + "learning_rate": 2.462962962962963e-06, + "loss": 1.0014, + "step": 9340 + }, + { + "epoch": 1.8326146609172873, + "grad_norm": 6.487533092498779, + "learning_rate": 2.453703703703704e-06, + "loss": 1.0022, + "step": 9350 + }, + { + "epoch": 1.8345746765974127, + "grad_norm": 6.143520832061768, + "learning_rate": 2.4444444444444447e-06, + "loss": 0.9852, + "step": 9360 + }, + { + "epoch": 1.8365346922775383, + "grad_norm": 6.246576309204102, + "learning_rate": 2.4351851851851855e-06, + "loss": 0.9995, + "step": 9370 + }, + { + "epoch": 1.8384947079576637, + "grad_norm": 6.709537982940674, + "learning_rate": 2.425925925925926e-06, + "loss": 1.0039, + "step": 9380 + }, + { + "epoch": 1.840454723637789, + "grad_norm": 6.031428813934326, + "learning_rate": 2.4166666666666667e-06, + "loss": 1.0103, + "step": 9390 + }, + { + "epoch": 1.8424147393179147, + "grad_norm": 6.703726291656494, + "learning_rate": 2.4074074074074075e-06, + "loss": 0.9999, + "step": 9400 + }, + { + "epoch": 1.8443747549980398, + "grad_norm": 6.8644819259643555, + "learning_rate": 2.3981481481481483e-06, + "loss": 0.9994, + "step": 9410 + }, + { + "epoch": 1.8463347706781654, + "grad_norm": 7.306818962097168, + "learning_rate": 2.388888888888889e-06, + "loss": 0.9995, + "step": 9420 + }, + { + "epoch": 1.848294786358291, + "grad_norm": 5.757288932800293, + "learning_rate": 2.37962962962963e-06, + "loss": 0.9883, + "step": 9430 + }, + { + "epoch": 1.8502548020384162, + "grad_norm": 6.339224815368652, + "learning_rate": 2.3703703703703707e-06, + "loss": 0.9953, + "step": 9440 + }, + { + "epoch": 1.8522148177185418, + "grad_norm": 8.186257362365723, + "learning_rate": 2.361111111111111e-06, + "loss": 0.9999, + "step": 9450 + }, + { + "epoch": 1.8541748333986672, + "grad_norm": 6.2635111808776855, + "learning_rate": 2.351851851851852e-06, + "loss": 1.0047, + "step": 9460 + }, + { + "epoch": 1.8561348490787926, + "grad_norm": 6.483547210693359, + "learning_rate": 2.3425925925925928e-06, + "loss": 1.0003, + "step": 9470 + }, + { + "epoch": 1.8580948647589182, + "grad_norm": 6.6517333984375, + "learning_rate": 2.3333333333333336e-06, + "loss": 1.0152, + "step": 9480 + }, + { + "epoch": 1.8600548804390435, + "grad_norm": 6.463938236236572, + "learning_rate": 2.324074074074074e-06, + "loss": 0.9917, + "step": 9490 + }, + { + "epoch": 1.862014896119169, + "grad_norm": 6.184695243835449, + "learning_rate": 2.314814814814815e-06, + "loss": 0.9889, + "step": 9500 + }, + { + "epoch": 1.8639749117992945, + "grad_norm": 6.435100078582764, + "learning_rate": 2.305555555555556e-06, + "loss": 1.002, + "step": 9510 + }, + { + "epoch": 1.8659349274794197, + "grad_norm": 5.961505889892578, + "learning_rate": 2.2962962962962964e-06, + "loss": 0.9931, + "step": 9520 + }, + { + "epoch": 1.8678949431595453, + "grad_norm": 6.590498924255371, + "learning_rate": 2.2870370370370372e-06, + "loss": 0.9967, + "step": 9530 + }, + { + "epoch": 1.8698549588396707, + "grad_norm": 6.019999027252197, + "learning_rate": 2.277777777777778e-06, + "loss": 0.9887, + "step": 9540 + }, + { + "epoch": 1.871814974519796, + "grad_norm": 6.332427978515625, + "learning_rate": 2.268518518518519e-06, + "loss": 1.0014, + "step": 9550 + }, + { + "epoch": 1.8737749901999217, + "grad_norm": 6.301807880401611, + "learning_rate": 2.2592592592592592e-06, + "loss": 0.9969, + "step": 9560 + }, + { + "epoch": 1.875735005880047, + "grad_norm": 6.279837131500244, + "learning_rate": 2.25e-06, + "loss": 0.9944, + "step": 9570 + }, + { + "epoch": 1.8776950215601724, + "grad_norm": 7.422451496124268, + "learning_rate": 2.240740740740741e-06, + "loss": 1.0007, + "step": 9580 + }, + { + "epoch": 1.879655037240298, + "grad_norm": 6.722255229949951, + "learning_rate": 2.2314814814814817e-06, + "loss": 0.9985, + "step": 9590 + }, + { + "epoch": 1.8816150529204234, + "grad_norm": 6.471696853637695, + "learning_rate": 2.222222222222222e-06, + "loss": 0.9984, + "step": 9600 + }, + { + "epoch": 1.8835750686005488, + "grad_norm": 6.347057819366455, + "learning_rate": 2.2129629629629633e-06, + "loss": 1.0092, + "step": 9610 + }, + { + "epoch": 1.8855350842806744, + "grad_norm": 6.340747833251953, + "learning_rate": 2.203703703703704e-06, + "loss": 0.9978, + "step": 9620 + }, + { + "epoch": 1.8874950999607996, + "grad_norm": 5.975046157836914, + "learning_rate": 2.1944444444444445e-06, + "loss": 1.0006, + "step": 9630 + }, + { + "epoch": 1.8894551156409252, + "grad_norm": 6.538125038146973, + "learning_rate": 2.1851851851851853e-06, + "loss": 1.006, + "step": 9640 + }, + { + "epoch": 1.8914151313210505, + "grad_norm": 6.345036506652832, + "learning_rate": 2.175925925925926e-06, + "loss": 1.0106, + "step": 9650 + }, + { + "epoch": 1.893375147001176, + "grad_norm": 6.604660987854004, + "learning_rate": 2.166666666666667e-06, + "loss": 0.9964, + "step": 9660 + }, + { + "epoch": 1.8953351626813015, + "grad_norm": 6.32635498046875, + "learning_rate": 2.1574074074074073e-06, + "loss": 1.0108, + "step": 9670 + }, + { + "epoch": 1.897295178361427, + "grad_norm": 6.804265975952148, + "learning_rate": 2.148148148148148e-06, + "loss": 1.0053, + "step": 9680 + }, + { + "epoch": 1.8992551940415523, + "grad_norm": 6.880146503448486, + "learning_rate": 2.138888888888889e-06, + "loss": 1.0073, + "step": 9690 + }, + { + "epoch": 1.901215209721678, + "grad_norm": 7.2419538497924805, + "learning_rate": 2.1296296296296298e-06, + "loss": 1.004, + "step": 9700 + }, + { + "epoch": 1.903175225401803, + "grad_norm": 6.319647789001465, + "learning_rate": 2.1203703703703706e-06, + "loss": 1.0049, + "step": 9710 + }, + { + "epoch": 1.9051352410819287, + "grad_norm": 6.6896491050720215, + "learning_rate": 2.1111111111111114e-06, + "loss": 1.0007, + "step": 9720 + }, + { + "epoch": 1.907095256762054, + "grad_norm": 5.643420696258545, + "learning_rate": 2.1018518518518522e-06, + "loss": 0.9898, + "step": 9730 + }, + { + "epoch": 1.9090552724421794, + "grad_norm": 6.660647869110107, + "learning_rate": 2.0925925925925926e-06, + "loss": 1.0039, + "step": 9740 + }, + { + "epoch": 1.911015288122305, + "grad_norm": 7.0795464515686035, + "learning_rate": 2.0833333333333334e-06, + "loss": 0.986, + "step": 9750 + }, + { + "epoch": 1.9129753038024304, + "grad_norm": 6.72257137298584, + "learning_rate": 2.0740740740740742e-06, + "loss": 1.0045, + "step": 9760 + }, + { + "epoch": 1.9149353194825558, + "grad_norm": 7.132964134216309, + "learning_rate": 2.064814814814815e-06, + "loss": 1.0043, + "step": 9770 + }, + { + "epoch": 1.9168953351626814, + "grad_norm": 6.032354354858398, + "learning_rate": 2.0555555555555555e-06, + "loss": 1.0159, + "step": 9780 + }, + { + "epoch": 1.9188553508428068, + "grad_norm": 6.651278495788574, + "learning_rate": 2.0462962962962967e-06, + "loss": 1.0009, + "step": 9790 + }, + { + "epoch": 1.9208153665229322, + "grad_norm": 6.432239532470703, + "learning_rate": 2.037037037037037e-06, + "loss": 1.0035, + "step": 9800 + }, + { + "epoch": 1.9208153665229322, + "eval_loss": 1.0201009511947632, + "eval_runtime": 14.0626, + "eval_samples_per_second": 46.506, + "eval_steps_per_second": 5.831, + "step": 9800 + }, + { + "epoch": 1.9227753822030578, + "grad_norm": 6.820202827453613, + "learning_rate": 2.027777777777778e-06, + "loss": 0.9879, + "step": 9810 + }, + { + "epoch": 1.924735397883183, + "grad_norm": 6.324273586273193, + "learning_rate": 2.0185185185185187e-06, + "loss": 1.0022, + "step": 9820 + }, + { + "epoch": 1.9266954135633085, + "grad_norm": 5.975357532501221, + "learning_rate": 2.0092592592592595e-06, + "loss": 1.0168, + "step": 9830 + }, + { + "epoch": 1.928655429243434, + "grad_norm": 7.312973976135254, + "learning_rate": 2.0000000000000003e-06, + "loss": 1.0017, + "step": 9840 + }, + { + "epoch": 1.9306154449235593, + "grad_norm": 7.051765441894531, + "learning_rate": 1.9907407407407407e-06, + "loss": 0.992, + "step": 9850 + }, + { + "epoch": 1.932575460603685, + "grad_norm": 6.269126892089844, + "learning_rate": 1.9814814814814815e-06, + "loss": 1.0018, + "step": 9860 + }, + { + "epoch": 1.9345354762838103, + "grad_norm": 6.204473972320557, + "learning_rate": 1.9722222222222224e-06, + "loss": 0.999, + "step": 9870 + }, + { + "epoch": 1.9364954919639357, + "grad_norm": 6.758467197418213, + "learning_rate": 1.962962962962963e-06, + "loss": 1.0023, + "step": 9880 + }, + { + "epoch": 1.9384555076440613, + "grad_norm": 6.538867950439453, + "learning_rate": 1.953703703703704e-06, + "loss": 0.9907, + "step": 9890 + }, + { + "epoch": 1.9404155233241864, + "grad_norm": 7.003431797027588, + "learning_rate": 1.944444444444445e-06, + "loss": 1.0047, + "step": 9900 + }, + { + "epoch": 1.942375539004312, + "grad_norm": 6.028417587280273, + "learning_rate": 1.935185185185185e-06, + "loss": 1.0027, + "step": 9910 + }, + { + "epoch": 1.9443355546844376, + "grad_norm": 6.356867790222168, + "learning_rate": 1.925925925925926e-06, + "loss": 0.9902, + "step": 9920 + }, + { + "epoch": 1.9462955703645628, + "grad_norm": 7.202489376068115, + "learning_rate": 1.916666666666667e-06, + "loss": 1.0106, + "step": 9930 + }, + { + "epoch": 1.9482555860446884, + "grad_norm": 6.344156265258789, + "learning_rate": 1.9074074074074076e-06, + "loss": 0.9916, + "step": 9940 + }, + { + "epoch": 1.9502156017248138, + "grad_norm": 6.817245006561279, + "learning_rate": 1.8981481481481484e-06, + "loss": 0.9829, + "step": 9950 + }, + { + "epoch": 1.9521756174049392, + "grad_norm": 6.988001823425293, + "learning_rate": 1.888888888888889e-06, + "loss": 0.9997, + "step": 9960 + }, + { + "epoch": 1.9541356330850648, + "grad_norm": 6.65484619140625, + "learning_rate": 1.8796296296296299e-06, + "loss": 1.0026, + "step": 9970 + }, + { + "epoch": 1.9560956487651902, + "grad_norm": 6.717133045196533, + "learning_rate": 1.8703703703703705e-06, + "loss": 1.0071, + "step": 9980 + }, + { + "epoch": 1.9580556644453155, + "grad_norm": 6.037536144256592, + "learning_rate": 1.8611111111111113e-06, + "loss": 0.993, + "step": 9990 + }, + { + "epoch": 1.9600156801254411, + "grad_norm": 6.965274810791016, + "learning_rate": 1.8518518518518519e-06, + "loss": 0.9918, + "step": 10000 + }, + { + "epoch": 1.9619756958055663, + "grad_norm": 6.768052101135254, + "learning_rate": 1.8425925925925927e-06, + "loss": 0.9873, + "step": 10010 + }, + { + "epoch": 1.963935711485692, + "grad_norm": 6.111681938171387, + "learning_rate": 1.8333333333333333e-06, + "loss": 1.0111, + "step": 10020 + }, + { + "epoch": 1.9658957271658173, + "grad_norm": 6.606689453125, + "learning_rate": 1.8240740740740743e-06, + "loss": 0.9997, + "step": 10030 + }, + { + "epoch": 1.9678557428459427, + "grad_norm": 6.256832599639893, + "learning_rate": 1.8148148148148151e-06, + "loss": 0.9977, + "step": 10040 + }, + { + "epoch": 1.9698157585260683, + "grad_norm": 6.401496410369873, + "learning_rate": 1.8055555555555557e-06, + "loss": 1.0069, + "step": 10050 + }, + { + "epoch": 1.9717757742061937, + "grad_norm": 6.166873455047607, + "learning_rate": 1.7962962962962965e-06, + "loss": 0.9998, + "step": 10060 + }, + { + "epoch": 1.973735789886319, + "grad_norm": 5.77340030670166, + "learning_rate": 1.7870370370370371e-06, + "loss": 1.0008, + "step": 10070 + }, + { + "epoch": 1.9756958055664446, + "grad_norm": 7.6040120124816895, + "learning_rate": 1.777777777777778e-06, + "loss": 0.9991, + "step": 10080 + }, + { + "epoch": 1.97765582124657, + "grad_norm": 6.114837169647217, + "learning_rate": 1.7685185185185186e-06, + "loss": 0.9949, + "step": 10090 + }, + { + "epoch": 1.9796158369266954, + "grad_norm": 6.565464496612549, + "learning_rate": 1.7592592592592594e-06, + "loss": 1.0141, + "step": 10100 + }, + { + "epoch": 1.981575852606821, + "grad_norm": 6.436923980712891, + "learning_rate": 1.75e-06, + "loss": 0.9975, + "step": 10110 + }, + { + "epoch": 1.9835358682869462, + "grad_norm": 7.330633640289307, + "learning_rate": 1.740740740740741e-06, + "loss": 1.0014, + "step": 10120 + }, + { + "epoch": 1.9854958839670718, + "grad_norm": 7.172529697418213, + "learning_rate": 1.7314814814814814e-06, + "loss": 0.9811, + "step": 10130 + }, + { + "epoch": 1.9874558996471972, + "grad_norm": 6.164400577545166, + "learning_rate": 1.7222222222222224e-06, + "loss": 0.9972, + "step": 10140 + }, + { + "epoch": 1.9894159153273225, + "grad_norm": 6.392111778259277, + "learning_rate": 1.7129629629629632e-06, + "loss": 1.0001, + "step": 10150 + }, + { + "epoch": 1.9913759310074481, + "grad_norm": 8.478386878967285, + "learning_rate": 1.7037037037037038e-06, + "loss": 1.0015, + "step": 10160 + }, + { + "epoch": 1.9933359466875735, + "grad_norm": 6.89587926864624, + "learning_rate": 1.6944444444444446e-06, + "loss": 1.0047, + "step": 10170 + }, + { + "epoch": 1.995295962367699, + "grad_norm": 6.686724662780762, + "learning_rate": 1.6851851851851852e-06, + "loss": 0.9953, + "step": 10180 + }, + { + "epoch": 1.9972559780478245, + "grad_norm": 6.857761859893799, + "learning_rate": 1.675925925925926e-06, + "loss": 0.9836, + "step": 10190 + }, + { + "epoch": 1.9992159937279497, + "grad_norm": 6.607189178466797, + "learning_rate": 1.6666666666666667e-06, + "loss": 0.9979, + "step": 10200 + }, + { + "epoch": 2.0011760094080753, + "grad_norm": 6.168262481689453, + "learning_rate": 1.6574074074074075e-06, + "loss": 0.9719, + "step": 10210 + }, + { + "epoch": 2.003136025088201, + "grad_norm": 6.525495529174805, + "learning_rate": 1.648148148148148e-06, + "loss": 0.9789, + "step": 10220 + }, + { + "epoch": 2.005096040768326, + "grad_norm": 5.562708854675293, + "learning_rate": 1.638888888888889e-06, + "loss": 0.98, + "step": 10230 + }, + { + "epoch": 2.0070560564484516, + "grad_norm": 7.023379802703857, + "learning_rate": 1.62962962962963e-06, + "loss": 0.985, + "step": 10240 + }, + { + "epoch": 2.0090160721285772, + "grad_norm": 7.55203104019165, + "learning_rate": 1.6203703703703705e-06, + "loss": 0.9758, + "step": 10250 + }, + { + "epoch": 2.0109760878087024, + "grad_norm": 13.453210830688477, + "learning_rate": 1.6111111111111113e-06, + "loss": 0.9656, + "step": 10260 + }, + { + "epoch": 2.012936103488828, + "grad_norm": 6.530963897705078, + "learning_rate": 1.601851851851852e-06, + "loss": 0.9615, + "step": 10270 + }, + { + "epoch": 2.014896119168953, + "grad_norm": 6.9765520095825195, + "learning_rate": 1.5925925925925927e-06, + "loss": 0.9786, + "step": 10280 + }, + { + "epoch": 2.0168561348490788, + "grad_norm": 7.124659538269043, + "learning_rate": 1.5833333333333333e-06, + "loss": 0.9795, + "step": 10290 + }, + { + "epoch": 2.0188161505292044, + "grad_norm": 7.119600772857666, + "learning_rate": 1.5740740740740742e-06, + "loss": 0.9672, + "step": 10300 + }, + { + "epoch": 2.0207761662093295, + "grad_norm": 6.711264133453369, + "learning_rate": 1.5648148148148148e-06, + "loss": 0.965, + "step": 10310 + }, + { + "epoch": 2.022736181889455, + "grad_norm": 7.027464389801025, + "learning_rate": 1.5555555555555558e-06, + "loss": 0.97, + "step": 10320 + }, + { + "epoch": 2.0246961975695807, + "grad_norm": 7.815002918243408, + "learning_rate": 1.5462962962962964e-06, + "loss": 0.9706, + "step": 10330 + }, + { + "epoch": 2.026656213249706, + "grad_norm": 6.989712238311768, + "learning_rate": 1.5370370370370372e-06, + "loss": 0.9741, + "step": 10340 + }, + { + "epoch": 2.0286162289298315, + "grad_norm": 6.903376579284668, + "learning_rate": 1.527777777777778e-06, + "loss": 0.9707, + "step": 10350 + }, + { + "epoch": 2.0305762446099567, + "grad_norm": 5.8991570472717285, + "learning_rate": 1.5185185185185186e-06, + "loss": 0.9731, + "step": 10360 + }, + { + "epoch": 2.0325362602900823, + "grad_norm": 7.208601474761963, + "learning_rate": 1.5092592592592594e-06, + "loss": 0.9845, + "step": 10370 + }, + { + "epoch": 2.034496275970208, + "grad_norm": 6.827620506286621, + "learning_rate": 1.5e-06, + "loss": 0.968, + "step": 10380 + }, + { + "epoch": 2.036456291650333, + "grad_norm": 6.6345295906066895, + "learning_rate": 1.4907407407407409e-06, + "loss": 0.9626, + "step": 10390 + }, + { + "epoch": 2.0384163073304586, + "grad_norm": 7.7344865798950195, + "learning_rate": 1.4814814814814815e-06, + "loss": 0.9717, + "step": 10400 + }, + { + "epoch": 2.0403763230105842, + "grad_norm": 7.092048168182373, + "learning_rate": 1.4722222222222225e-06, + "loss": 0.9721, + "step": 10410 + }, + { + "epoch": 2.0423363386907094, + "grad_norm": 6.746728897094727, + "learning_rate": 1.4629629629629629e-06, + "loss": 0.9755, + "step": 10420 + }, + { + "epoch": 2.044296354370835, + "grad_norm": 7.295462131500244, + "learning_rate": 1.453703703703704e-06, + "loss": 0.9633, + "step": 10430 + }, + { + "epoch": 2.0462563700509606, + "grad_norm": 6.882177829742432, + "learning_rate": 1.4444444444444445e-06, + "loss": 0.9668, + "step": 10440 + }, + { + "epoch": 2.0482163857310858, + "grad_norm": 6.1587300300598145, + "learning_rate": 1.4351851851851853e-06, + "loss": 0.9773, + "step": 10450 + }, + { + "epoch": 2.0501764014112114, + "grad_norm": 7.219092845916748, + "learning_rate": 1.4259259259259261e-06, + "loss": 0.9724, + "step": 10460 + }, + { + "epoch": 2.0521364170913365, + "grad_norm": 6.330183029174805, + "learning_rate": 1.4166666666666667e-06, + "loss": 0.965, + "step": 10470 + }, + { + "epoch": 2.054096432771462, + "grad_norm": 7.83421516418457, + "learning_rate": 1.4074074074074075e-06, + "loss": 0.9715, + "step": 10480 + }, + { + "epoch": 2.0560564484515877, + "grad_norm": 6.530287265777588, + "learning_rate": 1.3981481481481481e-06, + "loss": 0.9859, + "step": 10490 + }, + { + "epoch": 2.058016464131713, + "grad_norm": 7.4246625900268555, + "learning_rate": 1.3888888888888892e-06, + "loss": 0.9567, + "step": 10500 + }, + { + "epoch": 2.058016464131713, + "eval_loss": 1.0192580223083496, + "eval_runtime": 14.0156, + "eval_samples_per_second": 46.662, + "eval_steps_per_second": 5.851, + "step": 10500 + }, + { + "epoch": 2.0599764798118385, + "grad_norm": 7.052817344665527, + "learning_rate": 1.3796296296296296e-06, + "loss": 0.9657, + "step": 10510 + }, + { + "epoch": 2.061936495491964, + "grad_norm": 7.35888147354126, + "learning_rate": 1.3703703703703706e-06, + "loss": 0.9702, + "step": 10520 + }, + { + "epoch": 2.0638965111720893, + "grad_norm": 6.373018264770508, + "learning_rate": 1.3611111111111112e-06, + "loss": 0.9753, + "step": 10530 + }, + { + "epoch": 2.065856526852215, + "grad_norm": 6.720044136047363, + "learning_rate": 1.351851851851852e-06, + "loss": 0.9708, + "step": 10540 + }, + { + "epoch": 2.0678165425323405, + "grad_norm": 6.653097152709961, + "learning_rate": 1.3425925925925928e-06, + "loss": 0.9702, + "step": 10550 + }, + { + "epoch": 2.0697765582124656, + "grad_norm": 7.028474807739258, + "learning_rate": 1.3333333333333334e-06, + "loss": 0.9787, + "step": 10560 + }, + { + "epoch": 2.0717365738925912, + "grad_norm": 7.457616329193115, + "learning_rate": 1.3240740740740742e-06, + "loss": 0.9585, + "step": 10570 + }, + { + "epoch": 2.0736965895727164, + "grad_norm": 6.340860843658447, + "learning_rate": 1.3148148148148148e-06, + "loss": 0.9592, + "step": 10580 + }, + { + "epoch": 2.075656605252842, + "grad_norm": 7.086592674255371, + "learning_rate": 1.3055555555555556e-06, + "loss": 0.9701, + "step": 10590 + }, + { + "epoch": 2.0776166209329676, + "grad_norm": 7.2209320068359375, + "learning_rate": 1.2962962962962962e-06, + "loss": 0.9655, + "step": 10600 + }, + { + "epoch": 2.0795766366130928, + "grad_norm": 6.678873062133789, + "learning_rate": 1.2870370370370373e-06, + "loss": 0.9674, + "step": 10610 + }, + { + "epoch": 2.0815366522932184, + "grad_norm": 7.573465347290039, + "learning_rate": 1.2777777777777779e-06, + "loss": 0.9754, + "step": 10620 + }, + { + "epoch": 2.083496667973344, + "grad_norm": 7.289187908172607, + "learning_rate": 1.2685185185185187e-06, + "loss": 0.9769, + "step": 10630 + }, + { + "epoch": 2.085456683653469, + "grad_norm": 6.930860996246338, + "learning_rate": 1.2592592592592593e-06, + "loss": 0.9697, + "step": 10640 + }, + { + "epoch": 2.0874166993335947, + "grad_norm": 6.5301618576049805, + "learning_rate": 1.25e-06, + "loss": 0.9766, + "step": 10650 + }, + { + "epoch": 2.08937671501372, + "grad_norm": 8.65379524230957, + "learning_rate": 1.240740740740741e-06, + "loss": 0.9545, + "step": 10660 + }, + { + "epoch": 2.0913367306938455, + "grad_norm": 6.923322677612305, + "learning_rate": 1.2314814814814815e-06, + "loss": 0.9643, + "step": 10670 + }, + { + "epoch": 2.093296746373971, + "grad_norm": 7.2696614265441895, + "learning_rate": 1.2222222222222223e-06, + "loss": 0.9677, + "step": 10680 + }, + { + "epoch": 2.0952567620540963, + "grad_norm": 7.1026835441589355, + "learning_rate": 1.212962962962963e-06, + "loss": 0.9791, + "step": 10690 + }, + { + "epoch": 2.097216777734222, + "grad_norm": 6.935880661010742, + "learning_rate": 1.2037037037037037e-06, + "loss": 0.9639, + "step": 10700 + }, + { + "epoch": 2.0991767934143475, + "grad_norm": 6.7199201583862305, + "learning_rate": 1.1944444444444446e-06, + "loss": 0.9795, + "step": 10710 + }, + { + "epoch": 2.1011368090944726, + "grad_norm": 7.126111030578613, + "learning_rate": 1.1851851851851854e-06, + "loss": 0.9604, + "step": 10720 + }, + { + "epoch": 2.1030968247745983, + "grad_norm": 7.274319171905518, + "learning_rate": 1.175925925925926e-06, + "loss": 0.9696, + "step": 10730 + }, + { + "epoch": 2.105056840454724, + "grad_norm": 7.07975959777832, + "learning_rate": 1.1666666666666668e-06, + "loss": 0.9708, + "step": 10740 + }, + { + "epoch": 2.107016856134849, + "grad_norm": 24.05130386352539, + "learning_rate": 1.1574074074074076e-06, + "loss": 0.9765, + "step": 10750 + }, + { + "epoch": 2.1089768718149746, + "grad_norm": 7.428647518157959, + "learning_rate": 1.1481481481481482e-06, + "loss": 0.9665, + "step": 10760 + }, + { + "epoch": 2.1109368874951, + "grad_norm": 6.800421714782715, + "learning_rate": 1.138888888888889e-06, + "loss": 0.9687, + "step": 10770 + }, + { + "epoch": 2.1128969031752254, + "grad_norm": 8.014981269836426, + "learning_rate": 1.1296296296296296e-06, + "loss": 0.9773, + "step": 10780 + }, + { + "epoch": 2.114856918855351, + "grad_norm": 7.821423053741455, + "learning_rate": 1.1203703703703704e-06, + "loss": 0.9694, + "step": 10790 + }, + { + "epoch": 2.116816934535476, + "grad_norm": 8.069396018981934, + "learning_rate": 1.111111111111111e-06, + "loss": 0.9741, + "step": 10800 + }, + { + "epoch": 2.1187769502156018, + "grad_norm": 7.268844127655029, + "learning_rate": 1.101851851851852e-06, + "loss": 0.9685, + "step": 10810 + }, + { + "epoch": 2.1207369658957274, + "grad_norm": 7.122548580169678, + "learning_rate": 1.0925925925925927e-06, + "loss": 0.9491, + "step": 10820 + }, + { + "epoch": 2.1226969815758525, + "grad_norm": 6.834280014038086, + "learning_rate": 1.0833333333333335e-06, + "loss": 0.9836, + "step": 10830 + }, + { + "epoch": 2.124656997255978, + "grad_norm": 6.231177806854248, + "learning_rate": 1.074074074074074e-06, + "loss": 0.9729, + "step": 10840 + }, + { + "epoch": 2.1266170129361033, + "grad_norm": 7.387545585632324, + "learning_rate": 1.0648148148148149e-06, + "loss": 0.9658, + "step": 10850 + }, + { + "epoch": 2.128577028616229, + "grad_norm": 8.538785934448242, + "learning_rate": 1.0555555555555557e-06, + "loss": 0.9521, + "step": 10860 + }, + { + "epoch": 2.1305370442963545, + "grad_norm": 7.329588890075684, + "learning_rate": 1.0462962962962963e-06, + "loss": 0.9804, + "step": 10870 + }, + { + "epoch": 2.1324970599764796, + "grad_norm": 6.391878128051758, + "learning_rate": 1.0370370370370371e-06, + "loss": 0.9684, + "step": 10880 + }, + { + "epoch": 2.1344570756566053, + "grad_norm": 7.090411186218262, + "learning_rate": 1.0277777777777777e-06, + "loss": 0.9622, + "step": 10890 + }, + { + "epoch": 2.136417091336731, + "grad_norm": 6.1176958084106445, + "learning_rate": 1.0185185185185185e-06, + "loss": 0.9696, + "step": 10900 + }, + { + "epoch": 2.138377107016856, + "grad_norm": 10.32075023651123, + "learning_rate": 1.0092592592592594e-06, + "loss": 0.9686, + "step": 10910 + }, + { + "epoch": 2.1403371226969816, + "grad_norm": 7.174540996551514, + "learning_rate": 1.0000000000000002e-06, + "loss": 0.9647, + "step": 10920 + }, + { + "epoch": 2.1422971383771072, + "grad_norm": 7.739523410797119, + "learning_rate": 9.907407407407408e-07, + "loss": 0.9646, + "step": 10930 + }, + { + "epoch": 2.1442571540572324, + "grad_norm": 7.084393501281738, + "learning_rate": 9.814814814814816e-07, + "loss": 0.9695, + "step": 10940 + }, + { + "epoch": 2.146217169737358, + "grad_norm": 7.240608215332031, + "learning_rate": 9.722222222222224e-07, + "loss": 0.9655, + "step": 10950 + }, + { + "epoch": 2.148177185417483, + "grad_norm": 7.608102798461914, + "learning_rate": 9.62962962962963e-07, + "loss": 0.9598, + "step": 10960 + }, + { + "epoch": 2.1501372010976088, + "grad_norm": 7.160534858703613, + "learning_rate": 9.537037037037038e-07, + "loss": 0.9649, + "step": 10970 + }, + { + "epoch": 2.1520972167777344, + "grad_norm": 7.145265102386475, + "learning_rate": 9.444444444444445e-07, + "loss": 0.9646, + "step": 10980 + }, + { + "epoch": 2.1540572324578595, + "grad_norm": 25.257863998413086, + "learning_rate": 9.351851851851852e-07, + "loss": 0.9529, + "step": 10990 + }, + { + "epoch": 2.156017248137985, + "grad_norm": 7.226001262664795, + "learning_rate": 9.259259259259259e-07, + "loss": 0.9556, + "step": 11000 + }, + { + "epoch": 2.1579772638181107, + "grad_norm": 8.324951171875, + "learning_rate": 9.166666666666666e-07, + "loss": 0.9666, + "step": 11010 + }, + { + "epoch": 2.159937279498236, + "grad_norm": 7.676420211791992, + "learning_rate": 9.074074074074076e-07, + "loss": 0.9822, + "step": 11020 + }, + { + "epoch": 2.1618972951783615, + "grad_norm": 8.159540176391602, + "learning_rate": 8.981481481481483e-07, + "loss": 0.9608, + "step": 11030 + }, + { + "epoch": 2.1638573108584866, + "grad_norm": 7.420976161956787, + "learning_rate": 8.88888888888889e-07, + "loss": 0.974, + "step": 11040 + }, + { + "epoch": 2.1658173265386123, + "grad_norm": 7.221758842468262, + "learning_rate": 8.796296296296297e-07, + "loss": 0.9679, + "step": 11050 + }, + { + "epoch": 2.167777342218738, + "grad_norm": 7.399634838104248, + "learning_rate": 8.703703703703705e-07, + "loss": 0.9699, + "step": 11060 + }, + { + "epoch": 2.169737357898863, + "grad_norm": 6.9496026039123535, + "learning_rate": 8.611111111111112e-07, + "loss": 0.963, + "step": 11070 + }, + { + "epoch": 2.1716973735789886, + "grad_norm": 7.964886665344238, + "learning_rate": 8.518518518518519e-07, + "loss": 0.9649, + "step": 11080 + }, + { + "epoch": 2.1736573892591142, + "grad_norm": 6.945036888122559, + "learning_rate": 8.425925925925926e-07, + "loss": 0.9658, + "step": 11090 + }, + { + "epoch": 2.1756174049392394, + "grad_norm": 7.315770149230957, + "learning_rate": 8.333333333333333e-07, + "loss": 0.9715, + "step": 11100 + }, + { + "epoch": 2.177577420619365, + "grad_norm": 7.41797399520874, + "learning_rate": 8.24074074074074e-07, + "loss": 0.9579, + "step": 11110 + }, + { + "epoch": 2.1795374362994906, + "grad_norm": 6.9420366287231445, + "learning_rate": 8.14814814814815e-07, + "loss": 0.9785, + "step": 11120 + }, + { + "epoch": 2.1814974519796158, + "grad_norm": 6.257632255554199, + "learning_rate": 8.055555555555557e-07, + "loss": 0.9673, + "step": 11130 + }, + { + "epoch": 2.1834574676597414, + "grad_norm": 7.180861949920654, + "learning_rate": 7.962962962962964e-07, + "loss": 0.9691, + "step": 11140 + }, + { + "epoch": 2.1854174833398665, + "grad_norm": 7.286172389984131, + "learning_rate": 7.870370370370371e-07, + "loss": 0.9583, + "step": 11150 + }, + { + "epoch": 2.187377499019992, + "grad_norm": 7.305385589599609, + "learning_rate": 7.777777777777779e-07, + "loss": 0.9632, + "step": 11160 + }, + { + "epoch": 2.1893375147001177, + "grad_norm": 7.4768290519714355, + "learning_rate": 7.685185185185186e-07, + "loss": 0.9777, + "step": 11170 + }, + { + "epoch": 2.191297530380243, + "grad_norm": 7.44431209564209, + "learning_rate": 7.592592592592593e-07, + "loss": 0.9748, + "step": 11180 + }, + { + "epoch": 2.1932575460603685, + "grad_norm": 7.362459182739258, + "learning_rate": 7.5e-07, + "loss": 0.97, + "step": 11190 + }, + { + "epoch": 2.195217561740494, + "grad_norm": 6.118969917297363, + "learning_rate": 7.407407407407407e-07, + "loss": 0.967, + "step": 11200 + }, + { + "epoch": 2.195217561740494, + "eval_loss": 1.015982985496521, + "eval_runtime": 14.0402, + "eval_samples_per_second": 46.58, + "eval_steps_per_second": 5.84, + "step": 11200 + }, + { + "epoch": 2.1971775774206193, + "grad_norm": 7.092718601226807, + "learning_rate": 7.314814814814814e-07, + "loss": 0.9661, + "step": 11210 + }, + { + "epoch": 2.199137593100745, + "grad_norm": 7.384366512298584, + "learning_rate": 7.222222222222222e-07, + "loss": 0.9669, + "step": 11220 + }, + { + "epoch": 2.20109760878087, + "grad_norm": 7.769095420837402, + "learning_rate": 7.129629629629631e-07, + "loss": 0.9674, + "step": 11230 + }, + { + "epoch": 2.2030576244609956, + "grad_norm": 6.9204864501953125, + "learning_rate": 7.037037037037038e-07, + "loss": 0.9661, + "step": 11240 + }, + { + "epoch": 2.2050176401411212, + "grad_norm": 6.869389057159424, + "learning_rate": 6.944444444444446e-07, + "loss": 0.9619, + "step": 11250 + }, + { + "epoch": 2.2069776558212464, + "grad_norm": 7.39153528213501, + "learning_rate": 6.851851851851853e-07, + "loss": 0.9605, + "step": 11260 + }, + { + "epoch": 2.208937671501372, + "grad_norm": 6.61447811126709, + "learning_rate": 6.75925925925926e-07, + "loss": 0.9796, + "step": 11270 + }, + { + "epoch": 2.2108976871814976, + "grad_norm": 6.7030863761901855, + "learning_rate": 6.666666666666667e-07, + "loss": 0.96, + "step": 11280 + }, + { + "epoch": 2.2128577028616228, + "grad_norm": 7.7893829345703125, + "learning_rate": 6.574074074074074e-07, + "loss": 0.9571, + "step": 11290 + }, + { + "epoch": 2.2148177185417484, + "grad_norm": 7.597008228302002, + "learning_rate": 6.481481481481481e-07, + "loss": 0.9672, + "step": 11300 + }, + { + "epoch": 2.216777734221874, + "grad_norm": 6.373151779174805, + "learning_rate": 6.388888888888889e-07, + "loss": 0.9713, + "step": 11310 + }, + { + "epoch": 2.218737749901999, + "grad_norm": 7.9771575927734375, + "learning_rate": 6.296296296296296e-07, + "loss": 0.9674, + "step": 11320 + }, + { + "epoch": 2.2206977655821247, + "grad_norm": 7.23372745513916, + "learning_rate": 6.203703703703705e-07, + "loss": 0.9787, + "step": 11330 + }, + { + "epoch": 2.2226577812622503, + "grad_norm": 7.515969276428223, + "learning_rate": 6.111111111111112e-07, + "loss": 0.9702, + "step": 11340 + }, + { + "epoch": 2.2246177969423755, + "grad_norm": 6.468687534332275, + "learning_rate": 6.018518518518519e-07, + "loss": 0.981, + "step": 11350 + }, + { + "epoch": 2.226577812622501, + "grad_norm": 6.765408992767334, + "learning_rate": 5.925925925925927e-07, + "loss": 0.9757, + "step": 11360 + }, + { + "epoch": 2.2285378283026263, + "grad_norm": 8.031702041625977, + "learning_rate": 5.833333333333334e-07, + "loss": 0.9643, + "step": 11370 + }, + { + "epoch": 2.230497843982752, + "grad_norm": 7.857892990112305, + "learning_rate": 5.740740740740741e-07, + "loss": 0.9712, + "step": 11380 + }, + { + "epoch": 2.2324578596628775, + "grad_norm": 6.461490154266357, + "learning_rate": 5.648148148148148e-07, + "loss": 0.9664, + "step": 11390 + }, + { + "epoch": 2.2344178753430026, + "grad_norm": 7.290163993835449, + "learning_rate": 5.555555555555555e-07, + "loss": 0.9554, + "step": 11400 + }, + { + "epoch": 2.2363778910231282, + "grad_norm": 7.196146011352539, + "learning_rate": 5.462962962962963e-07, + "loss": 0.9755, + "step": 11410 + }, + { + "epoch": 2.238337906703254, + "grad_norm": 7.482301712036133, + "learning_rate": 5.37037037037037e-07, + "loss": 0.9577, + "step": 11420 + }, + { + "epoch": 2.240297922383379, + "grad_norm": 7.794521808624268, + "learning_rate": 5.277777777777779e-07, + "loss": 0.9679, + "step": 11430 + }, + { + "epoch": 2.2422579380635046, + "grad_norm": 6.857132434844971, + "learning_rate": 5.185185185185186e-07, + "loss": 0.9561, + "step": 11440 + }, + { + "epoch": 2.2442179537436298, + "grad_norm": 6.887381553649902, + "learning_rate": 5.092592592592593e-07, + "loss": 0.9712, + "step": 11450 + }, + { + "epoch": 2.2461779694237554, + "grad_norm": 7.537138938903809, + "learning_rate": 5.000000000000001e-07, + "loss": 0.9755, + "step": 11460 + }, + { + "epoch": 2.248137985103881, + "grad_norm": 7.604095935821533, + "learning_rate": 4.907407407407408e-07, + "loss": 0.968, + "step": 11470 + }, + { + "epoch": 2.250098000784006, + "grad_norm": 6.840078830718994, + "learning_rate": 4.814814814814815e-07, + "loss": 0.9677, + "step": 11480 + }, + { + "epoch": 2.2520580164641317, + "grad_norm": 7.0317301750183105, + "learning_rate": 4.7222222222222226e-07, + "loss": 0.9635, + "step": 11490 + }, + { + "epoch": 2.2540180321442573, + "grad_norm": 7.006657123565674, + "learning_rate": 4.6296296296296297e-07, + "loss": 0.96, + "step": 11500 + }, + { + "epoch": 2.2559780478243825, + "grad_norm": 7.701179504394531, + "learning_rate": 4.537037037037038e-07, + "loss": 0.9747, + "step": 11510 + }, + { + "epoch": 2.257938063504508, + "grad_norm": 7.04123592376709, + "learning_rate": 4.444444444444445e-07, + "loss": 0.9632, + "step": 11520 + }, + { + "epoch": 2.2598980791846337, + "grad_norm": 7.217432022094727, + "learning_rate": 4.3518518518518525e-07, + "loss": 0.9675, + "step": 11530 + }, + { + "epoch": 2.261858094864759, + "grad_norm": 7.661186695098877, + "learning_rate": 4.2592592592592596e-07, + "loss": 0.984, + "step": 11540 + }, + { + "epoch": 2.2638181105448845, + "grad_norm": 11.790616989135742, + "learning_rate": 4.1666666666666667e-07, + "loss": 0.9724, + "step": 11550 + }, + { + "epoch": 2.2657781262250096, + "grad_norm": 7.301620960235596, + "learning_rate": 4.074074074074075e-07, + "loss": 0.9601, + "step": 11560 + }, + { + "epoch": 2.2677381419051352, + "grad_norm": 8.462876319885254, + "learning_rate": 3.981481481481482e-07, + "loss": 0.9752, + "step": 11570 + }, + { + "epoch": 2.269698157585261, + "grad_norm": 6.7517547607421875, + "learning_rate": 3.8888888888888895e-07, + "loss": 0.9762, + "step": 11580 + }, + { + "epoch": 2.271658173265386, + "grad_norm": 7.299182891845703, + "learning_rate": 3.7962962962962966e-07, + "loss": 0.972, + "step": 11590 + }, + { + "epoch": 2.2736181889455116, + "grad_norm": 7.082560062408447, + "learning_rate": 3.7037037037037036e-07, + "loss": 0.9679, + "step": 11600 + }, + { + "epoch": 2.2755782046256368, + "grad_norm": 7.469284534454346, + "learning_rate": 3.611111111111111e-07, + "loss": 0.9546, + "step": 11610 + }, + { + "epoch": 2.2775382203057624, + "grad_norm": 7.5840535163879395, + "learning_rate": 3.518518518518519e-07, + "loss": 0.9761, + "step": 11620 + }, + { + "epoch": 2.279498235985888, + "grad_norm": 8.262914657592773, + "learning_rate": 3.4259259259259265e-07, + "loss": 0.9646, + "step": 11630 + }, + { + "epoch": 2.281458251666013, + "grad_norm": 8.672895431518555, + "learning_rate": 3.3333333333333335e-07, + "loss": 0.9732, + "step": 11640 + }, + { + "epoch": 2.2834182673461387, + "grad_norm": 7.6948347091674805, + "learning_rate": 3.2407407407407406e-07, + "loss": 0.9726, + "step": 11650 + }, + { + "epoch": 2.2853782830262643, + "grad_norm": 6.536133289337158, + "learning_rate": 3.148148148148148e-07, + "loss": 0.9694, + "step": 11660 + }, + { + "epoch": 2.2873382987063895, + "grad_norm": 8.90619945526123, + "learning_rate": 3.055555555555556e-07, + "loss": 0.9793, + "step": 11670 + }, + { + "epoch": 2.289298314386515, + "grad_norm": 7.702218055725098, + "learning_rate": 2.9629629629629634e-07, + "loss": 0.9642, + "step": 11680 + }, + { + "epoch": 2.2912583300666407, + "grad_norm": 8.119073867797852, + "learning_rate": 2.8703703703703705e-07, + "loss": 0.9641, + "step": 11690 + }, + { + "epoch": 2.293218345746766, + "grad_norm": 7.159200191497803, + "learning_rate": 2.7777777777777776e-07, + "loss": 0.9636, + "step": 11700 + }, + { + "epoch": 2.2951783614268915, + "grad_norm": 6.998493671417236, + "learning_rate": 2.685185185185185e-07, + "loss": 0.9664, + "step": 11710 + }, + { + "epoch": 2.297138377107017, + "grad_norm": 7.50088357925415, + "learning_rate": 2.592592592592593e-07, + "loss": 0.9787, + "step": 11720 + }, + { + "epoch": 2.2990983927871422, + "grad_norm": 7.634678363800049, + "learning_rate": 2.5000000000000004e-07, + "loss": 0.9712, + "step": 11730 + }, + { + "epoch": 2.301058408467268, + "grad_norm": 7.635111331939697, + "learning_rate": 2.4074074074074075e-07, + "loss": 0.9694, + "step": 11740 + }, + { + "epoch": 2.303018424147393, + "grad_norm": 7.645463466644287, + "learning_rate": 2.3148148148148148e-07, + "loss": 0.9717, + "step": 11750 + }, + { + "epoch": 2.3049784398275186, + "grad_norm": 6.8600568771362305, + "learning_rate": 2.2222222222222224e-07, + "loss": 0.9495, + "step": 11760 + }, + { + "epoch": 2.306938455507644, + "grad_norm": 7.019852638244629, + "learning_rate": 2.1296296296296298e-07, + "loss": 0.9527, + "step": 11770 + }, + { + "epoch": 2.3088984711877694, + "grad_norm": 8.227087020874023, + "learning_rate": 2.0370370370370374e-07, + "loss": 0.9608, + "step": 11780 + }, + { + "epoch": 2.310858486867895, + "grad_norm": 7.640336036682129, + "learning_rate": 1.9444444444444447e-07, + "loss": 0.9684, + "step": 11790 + }, + { + "epoch": 2.3128185025480206, + "grad_norm": 7.304652690887451, + "learning_rate": 1.8518518518518518e-07, + "loss": 0.9698, + "step": 11800 + }, + { + "epoch": 2.3147785182281457, + "grad_norm": 7.173428535461426, + "learning_rate": 1.7592592592592594e-07, + "loss": 0.9635, + "step": 11810 + }, + { + "epoch": 2.3167385339082713, + "grad_norm": 7.876226902008057, + "learning_rate": 1.6666666666666668e-07, + "loss": 0.9606, + "step": 11820 + }, + { + "epoch": 2.3186985495883965, + "grad_norm": 6.757920265197754, + "learning_rate": 1.574074074074074e-07, + "loss": 0.9692, + "step": 11830 + }, + { + "epoch": 2.320658565268522, + "grad_norm": 8.267070770263672, + "learning_rate": 1.4814814814814817e-07, + "loss": 0.9565, + "step": 11840 + }, + { + "epoch": 2.3226185809486477, + "grad_norm": 7.382258892059326, + "learning_rate": 1.3888888888888888e-07, + "loss": 0.9672, + "step": 11850 + }, + { + "epoch": 2.324578596628773, + "grad_norm": 7.452052116394043, + "learning_rate": 1.2962962962962964e-07, + "loss": 0.9649, + "step": 11860 + }, + { + "epoch": 2.3265386123088985, + "grad_norm": 9.273903846740723, + "learning_rate": 1.2037037037037037e-07, + "loss": 0.9652, + "step": 11870 + }, + { + "epoch": 2.328498627989024, + "grad_norm": 6.987192153930664, + "learning_rate": 1.1111111111111112e-07, + "loss": 0.9571, + "step": 11880 + }, + { + "epoch": 2.3304586436691492, + "grad_norm": 6.799464702606201, + "learning_rate": 1.0185185185185187e-07, + "loss": 0.9707, + "step": 11890 + }, + { + "epoch": 2.332418659349275, + "grad_norm": 7.16199254989624, + "learning_rate": 9.259259259259259e-08, + "loss": 0.9734, + "step": 11900 + }, + { + "epoch": 2.332418659349275, + "eval_loss": 1.0144330263137817, + "eval_runtime": 14.0924, + "eval_samples_per_second": 46.408, + "eval_steps_per_second": 5.819, + "step": 11900 + }, + { + "epoch": 2.3343786750294004, + "grad_norm": 8.338574409484863, + "learning_rate": 8.333333333333334e-08, + "loss": 0.9714, + "step": 11910 + }, + { + "epoch": 2.3363386907095256, + "grad_norm": 6.403331279754639, + "learning_rate": 7.407407407407409e-08, + "loss": 0.959, + "step": 11920 + }, + { + "epoch": 2.338298706389651, + "grad_norm": 8.215003967285156, + "learning_rate": 6.481481481481482e-08, + "loss": 0.964, + "step": 11930 + }, + { + "epoch": 2.3402587220697764, + "grad_norm": 7.930845737457275, + "learning_rate": 5.555555555555556e-08, + "loss": 0.9575, + "step": 11940 + }, + { + "epoch": 2.342218737749902, + "grad_norm": 7.033510684967041, + "learning_rate": 4.6296296296296295e-08, + "loss": 0.9626, + "step": 11950 + }, + { + "epoch": 2.3441787534300276, + "grad_norm": 7.8353800773620605, + "learning_rate": 3.703703703703704e-08, + "loss": 0.9673, + "step": 11960 + }, + { + "epoch": 2.3461387691101527, + "grad_norm": 7.304898738861084, + "learning_rate": 2.777777777777778e-08, + "loss": 0.9618, + "step": 11970 + }, + { + "epoch": 2.3480987847902783, + "grad_norm": 7.420264720916748, + "learning_rate": 1.851851851851852e-08, + "loss": 0.9544, + "step": 11980 + }, + { + "epoch": 2.350058800470404, + "grad_norm": 7.217324256896973, + "learning_rate": 9.25925925925926e-09, + "loss": 0.9752, + "step": 11990 + }, + { + "epoch": 2.352018816150529, + "grad_norm": 7.595900058746338, + "learning_rate": 0.0, + "loss": 0.9548, + "step": 12000 + }, + { + "epoch": 2.352018816150529, + "eval_loss": 1.0147355794906616, + "eval_runtime": 14.1258, + "eval_samples_per_second": 46.298, + "eval_steps_per_second": 5.805, + "step": 12000 + } + ], + "logging_steps": 10, + "max_steps": 12000, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 1, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 6.2834751593089335e+19, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}