| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.9981298423724285, | |
| "eval_steps": 100, | |
| "global_step": 467, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0021373230029388193, | |
| "grad_norm": 5.562169562786842, | |
| "learning_rate": 1.4893617021276595e-08, | |
| "logits/chosen": -0.9639219045639038, | |
| "logits/rejected": -0.9681637287139893, | |
| "logps/chosen": -254.6912078857422, | |
| "logps/rejected": -263.5589904785156, | |
| "loss": 0.6931, | |
| "rewards/accuracies": 0.0, | |
| "rewards/chosen": 0.0, | |
| "rewards/margins": 0.0, | |
| "rewards/rejected": 0.0, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.004274646005877639, | |
| "grad_norm": 5.450588274868951, | |
| "learning_rate": 2.978723404255319e-08, | |
| "logits/chosen": -0.9656609296798706, | |
| "logits/rejected": -0.9244194030761719, | |
| "logps/chosen": -326.766845703125, | |
| "logps/rejected": -278.7445068359375, | |
| "loss": 0.6931, | |
| "rewards/accuracies": 0.0, | |
| "rewards/chosen": 0.0, | |
| "rewards/margins": 0.0, | |
| "rewards/rejected": 0.0, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.006411969008816457, | |
| "grad_norm": 5.947914793856331, | |
| "learning_rate": 4.468085106382978e-08, | |
| "logits/chosen": -0.8438385725021362, | |
| "logits/rejected": -0.8096433877944946, | |
| "logps/chosen": -235.20404052734375, | |
| "logps/rejected": -236.10145568847656, | |
| "loss": 0.6932, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": 0.0013289405032992363, | |
| "rewards/margins": -0.00010445114457979798, | |
| "rewards/rejected": 0.0014333915896713734, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.008549292011755277, | |
| "grad_norm": 5.296514371602818, | |
| "learning_rate": 5.957446808510638e-08, | |
| "logits/chosen": -0.9574068784713745, | |
| "logits/rejected": -0.9784606099128723, | |
| "logps/chosen": -205.1729736328125, | |
| "logps/rejected": -211.75196838378906, | |
| "loss": 0.6935, | |
| "rewards/accuracies": 0.5, | |
| "rewards/chosen": -0.0005042314296588302, | |
| "rewards/margins": -0.00139321805909276, | |
| "rewards/rejected": 0.0008889864548109472, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.010686615014694095, | |
| "grad_norm": 5.901392519236585, | |
| "learning_rate": 7.446808510638298e-08, | |
| "logits/chosen": -1.0202012062072754, | |
| "logits/rejected": -1.05290687084198, | |
| "logps/chosen": -223.4305419921875, | |
| "logps/rejected": -223.0294647216797, | |
| "loss": 0.6933, | |
| "rewards/accuracies": 0.5, | |
| "rewards/chosen": 0.0012667321134358644, | |
| "rewards/margins": 0.00010157108772546053, | |
| "rewards/rejected": 0.0011651611421257257, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.012823938017632914, | |
| "grad_norm": 6.106968331942851, | |
| "learning_rate": 8.936170212765956e-08, | |
| "logits/chosen": -1.01454496383667, | |
| "logits/rejected": -0.9522863626480103, | |
| "logps/chosen": -310.51141357421875, | |
| "logps/rejected": -288.7611083984375, | |
| "loss": 0.6928, | |
| "rewards/accuracies": 0.3125, | |
| "rewards/chosen": -0.0014815283939242363, | |
| "rewards/margins": -0.0021853544749319553, | |
| "rewards/rejected": 0.0007038259063847363, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.014961261020571734, | |
| "grad_norm": 4.989592008812262, | |
| "learning_rate": 1.0425531914893615e-07, | |
| "logits/chosen": -1.084510087966919, | |
| "logits/rejected": -1.029686689376831, | |
| "logps/chosen": -236.44857788085938, | |
| "logps/rejected": -227.99609375, | |
| "loss": 0.6936, | |
| "rewards/accuracies": 0.5625, | |
| "rewards/chosen": -0.0002356480690650642, | |
| "rewards/margins": 0.001065850374288857, | |
| "rewards/rejected": -0.0013014983851462603, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.017098584023510555, | |
| "grad_norm": 5.864652813608174, | |
| "learning_rate": 1.1914893617021276e-07, | |
| "logits/chosen": -1.0324242115020752, | |
| "logits/rejected": -0.9804142713546753, | |
| "logps/chosen": -248.21893310546875, | |
| "logps/rejected": -262.17919921875, | |
| "loss": 0.6934, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -0.0010573435574769974, | |
| "rewards/margins": 7.791988173266873e-05, | |
| "rewards/rejected": -0.001135263592004776, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.01923590702644937, | |
| "grad_norm": 5.012836085935026, | |
| "learning_rate": 1.3404255319148934e-07, | |
| "logits/chosen": -0.9063066244125366, | |
| "logits/rejected": -0.888552188873291, | |
| "logps/chosen": -281.49481201171875, | |
| "logps/rejected": -285.80462646484375, | |
| "loss": 0.6932, | |
| "rewards/accuracies": 0.4375, | |
| "rewards/chosen": 0.0012888526543974876, | |
| "rewards/margins": -0.00021493433450814337, | |
| "rewards/rejected": 0.0015037870034575462, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.02137323002938819, | |
| "grad_norm": 5.441686883546068, | |
| "learning_rate": 1.4893617021276595e-07, | |
| "logits/chosen": -1.1913858652114868, | |
| "logits/rejected": -1.1670382022857666, | |
| "logps/chosen": -264.4725646972656, | |
| "logps/rejected": -285.6479797363281, | |
| "loss": 0.6932, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": 0.0027979514561593533, | |
| "rewards/margins": 0.0013645172584801912, | |
| "rewards/rejected": 0.0014334343140944839, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.02351055303232701, | |
| "grad_norm": 5.606555817511537, | |
| "learning_rate": 1.6382978723404256e-07, | |
| "logits/chosen": -1.0321263074874878, | |
| "logits/rejected": -1.0099542140960693, | |
| "logps/chosen": -266.173828125, | |
| "logps/rejected": -278.39117431640625, | |
| "loss": 0.6932, | |
| "rewards/accuracies": 0.5, | |
| "rewards/chosen": -5.334848538041115e-05, | |
| "rewards/margins": 0.0002821112284436822, | |
| "rewards/rejected": -0.00033545983023941517, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.02564787603526583, | |
| "grad_norm": 4.68531837452694, | |
| "learning_rate": 1.7872340425531912e-07, | |
| "logits/chosen": -0.9241148233413696, | |
| "logits/rejected": -0.9032956957817078, | |
| "logps/chosen": -232.3054962158203, | |
| "logps/rejected": -213.7734375, | |
| "loss": 0.6926, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": 0.0020639896392822266, | |
| "rewards/margins": 0.00174471375066787, | |
| "rewards/rejected": 0.000319275917718187, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.027785199038204648, | |
| "grad_norm": 6.413087644872293, | |
| "learning_rate": 1.9361702127659575e-07, | |
| "logits/chosen": -0.8520964980125427, | |
| "logits/rejected": -0.7619133591651917, | |
| "logps/chosen": -257.8199462890625, | |
| "logps/rejected": -286.0062255859375, | |
| "loss": 0.6932, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -0.0032129527535289526, | |
| "rewards/margins": 0.004196567460894585, | |
| "rewards/rejected": -0.007409519515931606, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.029922522041143467, | |
| "grad_norm": 6.109488078686547, | |
| "learning_rate": 2.085106382978723e-07, | |
| "logits/chosen": -1.0061469078063965, | |
| "logits/rejected": -1.0426419973373413, | |
| "logps/chosen": -300.9575500488281, | |
| "logps/rejected": -284.1053466796875, | |
| "loss": 0.6933, | |
| "rewards/accuracies": 0.4375, | |
| "rewards/chosen": -0.0033643674105405807, | |
| "rewards/margins": -0.002198734087869525, | |
| "rewards/rejected": -0.001165633206255734, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.03205984504408229, | |
| "grad_norm": 5.1906516925117225, | |
| "learning_rate": 2.2340425531914894e-07, | |
| "logits/chosen": -1.1642401218414307, | |
| "logits/rejected": -1.1202101707458496, | |
| "logps/chosen": -264.85552978515625, | |
| "logps/rejected": -270.9376525878906, | |
| "loss": 0.6927, | |
| "rewards/accuracies": 0.5, | |
| "rewards/chosen": -0.0047036223113536835, | |
| "rewards/margins": 0.0003113269922323525, | |
| "rewards/rejected": -0.005014949012547731, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.03419716804702111, | |
| "grad_norm": 4.9412589689200095, | |
| "learning_rate": 2.3829787234042553e-07, | |
| "logits/chosen": -1.0979671478271484, | |
| "logits/rejected": -1.067629098892212, | |
| "logps/chosen": -271.14385986328125, | |
| "logps/rejected": -280.5599670410156, | |
| "loss": 0.6922, | |
| "rewards/accuracies": 0.5, | |
| "rewards/chosen": -9.246054105460644e-06, | |
| "rewards/margins": 0.0020150854252278805, | |
| "rewards/rejected": -0.0020243311300873756, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.03633449104995993, | |
| "grad_norm": 5.374113235339209, | |
| "learning_rate": 2.5319148936170213e-07, | |
| "logits/chosen": -1.1378012895584106, | |
| "logits/rejected": -1.087912678718567, | |
| "logps/chosen": -237.7337188720703, | |
| "logps/rejected": -237.0882110595703, | |
| "loss": 0.6931, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -0.003059606533497572, | |
| "rewards/margins": -4.628184251487255e-05, | |
| "rewards/rejected": -0.003013324923813343, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.03847181405289874, | |
| "grad_norm": 6.092267360675494, | |
| "learning_rate": 2.680851063829787e-07, | |
| "logits/chosen": -1.0425890684127808, | |
| "logits/rejected": -1.0599486827850342, | |
| "logps/chosen": -242.11541748046875, | |
| "logps/rejected": -256.46417236328125, | |
| "loss": 0.6928, | |
| "rewards/accuracies": 0.5625, | |
| "rewards/chosen": -0.004783968441188335, | |
| "rewards/margins": -9.258274803869426e-05, | |
| "rewards/rejected": -0.004691386129707098, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.04060913705583756, | |
| "grad_norm": 5.1497648303931784, | |
| "learning_rate": 2.829787234042553e-07, | |
| "logits/chosen": -1.1328908205032349, | |
| "logits/rejected": -1.0197917222976685, | |
| "logps/chosen": -253.92127990722656, | |
| "logps/rejected": -278.0308837890625, | |
| "loss": 0.6923, | |
| "rewards/accuracies": 0.5, | |
| "rewards/chosen": -0.0021464251913130283, | |
| "rewards/margins": 0.00046774852671660483, | |
| "rewards/rejected": -0.0026141738053411245, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.04274646005877638, | |
| "grad_norm": 5.9119652568805225, | |
| "learning_rate": 2.978723404255319e-07, | |
| "logits/chosen": -1.1393866539001465, | |
| "logits/rejected": -1.0607531070709229, | |
| "logps/chosen": -303.4728088378906, | |
| "logps/rejected": -295.6599426269531, | |
| "loss": 0.6916, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -0.004760389216244221, | |
| "rewards/margins": 0.003504714695736766, | |
| "rewards/rejected": -0.00826510414481163, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.0448837830617152, | |
| "grad_norm": 5.752883808298607, | |
| "learning_rate": 3.1276595744680846e-07, | |
| "logits/chosen": -1.1522903442382812, | |
| "logits/rejected": -1.1488419771194458, | |
| "logps/chosen": -192.28042602539062, | |
| "logps/rejected": -200.01608276367188, | |
| "loss": 0.6911, | |
| "rewards/accuracies": 0.5, | |
| "rewards/chosen": -0.004406332969665527, | |
| "rewards/margins": -0.0005414390470832586, | |
| "rewards/rejected": -0.0038648939225822687, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.04702110606465402, | |
| "grad_norm": 4.856538956353405, | |
| "learning_rate": 3.276595744680851e-07, | |
| "logits/chosen": -1.0699973106384277, | |
| "logits/rejected": -1.0846004486083984, | |
| "logps/chosen": -205.7268829345703, | |
| "logps/rejected": -201.65975952148438, | |
| "loss": 0.6915, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -0.005930948071181774, | |
| "rewards/margins": 0.0032684658654034138, | |
| "rewards/rejected": -0.009199414402246475, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.04915842906759284, | |
| "grad_norm": 4.906648390477566, | |
| "learning_rate": 3.425531914893617e-07, | |
| "logits/chosen": -0.9317066669464111, | |
| "logits/rejected": -0.8009700179100037, | |
| "logps/chosen": -246.80438232421875, | |
| "logps/rejected": -272.9093322753906, | |
| "loss": 0.6911, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -0.010996412485837936, | |
| "rewards/margins": 0.0014129301998764277, | |
| "rewards/rejected": -0.012409343384206295, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.05129575207053166, | |
| "grad_norm": 5.875935612586806, | |
| "learning_rate": 3.5744680851063824e-07, | |
| "logits/chosen": -1.069819688796997, | |
| "logits/rejected": -0.9709153175354004, | |
| "logps/chosen": -195.9522247314453, | |
| "logps/rejected": -229.61566162109375, | |
| "loss": 0.6909, | |
| "rewards/accuracies": 0.5, | |
| "rewards/chosen": -0.004574608989059925, | |
| "rewards/margins": -0.0003843879676423967, | |
| "rewards/rejected": -0.00419022049754858, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.053433075073470476, | |
| "grad_norm": 4.985096243829474, | |
| "learning_rate": 3.7234042553191484e-07, | |
| "logits/chosen": -0.8998661637306213, | |
| "logits/rejected": -0.8598207831382751, | |
| "logps/chosen": -282.3105773925781, | |
| "logps/rejected": -272.4259338378906, | |
| "loss": 0.6905, | |
| "rewards/accuracies": 0.5, | |
| "rewards/chosen": -0.010662160813808441, | |
| "rewards/margins": 0.005043415818363428, | |
| "rewards/rejected": -0.015705574303865433, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.055570398076409296, | |
| "grad_norm": 5.218346832933627, | |
| "learning_rate": 3.872340425531915e-07, | |
| "logits/chosen": -0.9051691889762878, | |
| "logits/rejected": -0.8448677659034729, | |
| "logps/chosen": -241.62689208984375, | |
| "logps/rejected": -257.25244140625, | |
| "loss": 0.6901, | |
| "rewards/accuracies": 0.4375, | |
| "rewards/chosen": -0.02060263603925705, | |
| "rewards/margins": -0.0019798323046416044, | |
| "rewards/rejected": -0.018622804433107376, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.057707721079348115, | |
| "grad_norm": 5.189530040363191, | |
| "learning_rate": 4.021276595744681e-07, | |
| "logits/chosen": -1.0560821294784546, | |
| "logits/rejected": -1.0800153017044067, | |
| "logps/chosen": -252.28952026367188, | |
| "logps/rejected": -245.88502502441406, | |
| "loss": 0.6898, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -0.015861667692661285, | |
| "rewards/margins": 0.007801623083651066, | |
| "rewards/rejected": -0.023663289844989777, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.059845044082286934, | |
| "grad_norm": 5.8226698693397, | |
| "learning_rate": 4.170212765957446e-07, | |
| "logits/chosen": -1.0804771184921265, | |
| "logits/rejected": -1.099622130393982, | |
| "logps/chosen": -227.35537719726562, | |
| "logps/rejected": -225.36538696289062, | |
| "loss": 0.6887, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -0.017352018505334854, | |
| "rewards/margins": 0.01319974847137928, | |
| "rewards/rejected": -0.030551765114068985, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.061982367085225754, | |
| "grad_norm": 5.460690205724815, | |
| "learning_rate": 4.319148936170213e-07, | |
| "logits/chosen": -1.0153366327285767, | |
| "logits/rejected": -1.0103431940078735, | |
| "logps/chosen": -263.8896789550781, | |
| "logps/rejected": -267.3290100097656, | |
| "loss": 0.6878, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -0.02432805672287941, | |
| "rewards/margins": 0.008317284286022186, | |
| "rewards/rejected": -0.032645341008901596, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.06411969008816458, | |
| "grad_norm": 5.737017895955914, | |
| "learning_rate": 4.468085106382979e-07, | |
| "logits/chosen": -1.0464987754821777, | |
| "logits/rejected": -1.1331242322921753, | |
| "logps/chosen": -206.78468322753906, | |
| "logps/rejected": -192.26502990722656, | |
| "loss": 0.6883, | |
| "rewards/accuracies": 0.5625, | |
| "rewards/chosen": -0.010598372668027878, | |
| "rewards/margins": 0.002255907515063882, | |
| "rewards/rejected": -0.012854279950261116, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.06625701309110339, | |
| "grad_norm": 5.795133035841147, | |
| "learning_rate": 4.617021276595744e-07, | |
| "logits/chosen": -0.9297857880592346, | |
| "logits/rejected": -0.8592264652252197, | |
| "logps/chosen": -227.1981964111328, | |
| "logps/rejected": -247.9413299560547, | |
| "loss": 0.6863, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -0.03003655932843685, | |
| "rewards/margins": 0.009244289249181747, | |
| "rewards/rejected": -0.03928084671497345, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.06839433609404222, | |
| "grad_norm": 6.814768956718574, | |
| "learning_rate": 4.7659574468085105e-07, | |
| "logits/chosen": -1.05179762840271, | |
| "logits/rejected": -1.0814404487609863, | |
| "logps/chosen": -290.69525146484375, | |
| "logps/rejected": -274.92047119140625, | |
| "loss": 0.6818, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -0.05112060159444809, | |
| "rewards/margins": 0.015534894540905952, | |
| "rewards/rejected": -0.06665549427270889, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.07053165909698103, | |
| "grad_norm": 6.074153638196805, | |
| "learning_rate": 4.914893617021277e-07, | |
| "logits/chosen": -0.8849226236343384, | |
| "logits/rejected": -0.9219367504119873, | |
| "logps/chosen": -317.4134826660156, | |
| "logps/rejected": -325.8724365234375, | |
| "loss": 0.6832, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -0.04802538454532623, | |
| "rewards/margins": 0.024860400706529617, | |
| "rewards/rejected": -0.07288578897714615, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.07266898209991986, | |
| "grad_norm": 6.0900485696331215, | |
| "learning_rate": 5.063829787234043e-07, | |
| "logits/chosen": -1.0383641719818115, | |
| "logits/rejected": -1.0028424263000488, | |
| "logps/chosen": -277.57965087890625, | |
| "logps/rejected": -268.2386474609375, | |
| "loss": 0.6848, | |
| "rewards/accuracies": 0.5, | |
| "rewards/chosen": -0.060051530599594116, | |
| "rewards/margins": 0.0026317029260098934, | |
| "rewards/rejected": -0.06268322467803955, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.07480630510285867, | |
| "grad_norm": 6.827504173132289, | |
| "learning_rate": 5.212765957446808e-07, | |
| "logits/chosen": -1.0755884647369385, | |
| "logits/rejected": -1.0121068954467773, | |
| "logps/chosen": -311.99755859375, | |
| "logps/rejected": -350.29052734375, | |
| "loss": 0.682, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -0.061084531247615814, | |
| "rewards/margins": 0.04312053322792053, | |
| "rewards/rejected": -0.10420507192611694, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.07694362810579748, | |
| "grad_norm": 6.012937319519413, | |
| "learning_rate": 5.361702127659574e-07, | |
| "logits/chosen": -1.1560208797454834, | |
| "logits/rejected": -1.0447015762329102, | |
| "logps/chosen": -216.53421020507812, | |
| "logps/rejected": -234.53128051757812, | |
| "loss": 0.6788, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -0.04387297108769417, | |
| "rewards/margins": 0.016591297462582588, | |
| "rewards/rejected": -0.06046426668763161, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.07908095110873631, | |
| "grad_norm": 5.599688064843719, | |
| "learning_rate": 5.51063829787234e-07, | |
| "logits/chosen": -1.2253515720367432, | |
| "logits/rejected": -1.2332267761230469, | |
| "logps/chosen": -237.71002197265625, | |
| "logps/rejected": -236.66990661621094, | |
| "loss": 0.6799, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -0.08036091923713684, | |
| "rewards/margins": 0.03013552539050579, | |
| "rewards/rejected": -0.11049643903970718, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.08121827411167512, | |
| "grad_norm": 6.7746928876503105, | |
| "learning_rate": 5.659574468085106e-07, | |
| "logits/chosen": -1.0830399990081787, | |
| "logits/rejected": -1.0622460842132568, | |
| "logps/chosen": -284.8079833984375, | |
| "logps/rejected": -272.3097839355469, | |
| "loss": 0.6755, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -0.07364249229431152, | |
| "rewards/margins": 0.03328356519341469, | |
| "rewards/rejected": -0.10692604631185532, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.08335559711461395, | |
| "grad_norm": 6.77614986840528, | |
| "learning_rate": 5.808510638297872e-07, | |
| "logits/chosen": -1.0905030965805054, | |
| "logits/rejected": -1.0605907440185547, | |
| "logps/chosen": -210.57684326171875, | |
| "logps/rejected": -239.24871826171875, | |
| "loss": 0.6734, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -0.1003848984837532, | |
| "rewards/margins": 0.024850603193044662, | |
| "rewards/rejected": -0.12523549795150757, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.08549292011755276, | |
| "grad_norm": 6.424490447178767, | |
| "learning_rate": 5.957446808510638e-07, | |
| "logits/chosen": -1.1801645755767822, | |
| "logits/rejected": -1.1392816305160522, | |
| "logps/chosen": -263.45037841796875, | |
| "logps/rejected": -273.5709533691406, | |
| "loss": 0.6736, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -0.11433250457048416, | |
| "rewards/margins": 0.028288790956139565, | |
| "rewards/rejected": -0.14262130856513977, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.08763024312049159, | |
| "grad_norm": 6.4654988266248425, | |
| "learning_rate": 6.106382978723404e-07, | |
| "logits/chosen": -1.0815637111663818, | |
| "logits/rejected": -1.013136625289917, | |
| "logps/chosen": -206.23907470703125, | |
| "logps/rejected": -220.0244140625, | |
| "loss": 0.6722, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -0.11655673384666443, | |
| "rewards/margins": 0.04582422599196434, | |
| "rewards/rejected": -0.16238096356391907, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.0897675661234304, | |
| "grad_norm": 6.539055183308233, | |
| "learning_rate": 6.255319148936169e-07, | |
| "logits/chosen": -1.077747106552124, | |
| "logits/rejected": -1.0630099773406982, | |
| "logps/chosen": -268.242431640625, | |
| "logps/rejected": -304.5572204589844, | |
| "loss": 0.6716, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -0.1858682483434677, | |
| "rewards/margins": 0.05748748779296875, | |
| "rewards/rejected": -0.24335573613643646, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.09190488912636922, | |
| "grad_norm": 6.095364046949072, | |
| "learning_rate": 6.404255319148935e-07, | |
| "logits/chosen": -1.0136995315551758, | |
| "logits/rejected": -0.9615681171417236, | |
| "logps/chosen": -228.80368041992188, | |
| "logps/rejected": -235.71481323242188, | |
| "loss": 0.6689, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -0.1476023644208908, | |
| "rewards/margins": 0.0479072704911232, | |
| "rewards/rejected": -0.1955096423625946, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.09404221212930804, | |
| "grad_norm": 6.7344523666741765, | |
| "learning_rate": 6.553191489361702e-07, | |
| "logits/chosen": -0.8863506317138672, | |
| "logits/rejected": -0.9662103652954102, | |
| "logps/chosen": -259.41595458984375, | |
| "logps/rejected": -272.3363037109375, | |
| "loss": 0.6688, | |
| "rewards/accuracies": 0.4375, | |
| "rewards/chosen": -0.19808562099933624, | |
| "rewards/margins": 0.013540804386138916, | |
| "rewards/rejected": -0.21162641048431396, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.09617953513224686, | |
| "grad_norm": 6.29136903272874, | |
| "learning_rate": 6.702127659574469e-07, | |
| "logits/chosen": -1.0862454175949097, | |
| "logits/rejected": -1.03424870967865, | |
| "logps/chosen": -343.55059814453125, | |
| "logps/rejected": -352.5019836425781, | |
| "loss": 0.6619, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -0.25226354598999023, | |
| "rewards/margins": 0.05790935456752777, | |
| "rewards/rejected": -0.3101729154586792, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.09831685813518568, | |
| "grad_norm": 5.905042821290849, | |
| "learning_rate": 6.851063829787234e-07, | |
| "logits/chosen": -1.086181402206421, | |
| "logits/rejected": -1.0607441663742065, | |
| "logps/chosen": -281.6853942871094, | |
| "logps/rejected": -306.71136474609375, | |
| "loss": 0.6646, | |
| "rewards/accuracies": 0.5, | |
| "rewards/chosen": -0.30046984553337097, | |
| "rewards/margins": 0.11004211008548737, | |
| "rewards/rejected": -0.41051194071769714, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.1004541811381245, | |
| "grad_norm": 6.664836667296562, | |
| "learning_rate": 7e-07, | |
| "logits/chosen": -1.1094589233398438, | |
| "logits/rejected": -1.0737080574035645, | |
| "logps/chosen": -276.6365966796875, | |
| "logps/rejected": -290.7496032714844, | |
| "loss": 0.6509, | |
| "rewards/accuracies": 0.5, | |
| "rewards/chosen": -0.2919345498085022, | |
| "rewards/margins": 0.12451538443565369, | |
| "rewards/rejected": -0.41644996404647827, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.10259150414106331, | |
| "grad_norm": 6.452030342745469, | |
| "learning_rate": 6.999902087714444e-07, | |
| "logits/chosen": -1.0068883895874023, | |
| "logits/rejected": -0.955917239189148, | |
| "logps/chosen": -222.2923583984375, | |
| "logps/rejected": -239.01463317871094, | |
| "loss": 0.6577, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -0.1943490207195282, | |
| "rewards/margins": 0.08093887567520142, | |
| "rewards/rejected": -0.275287926197052, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.10472882714400214, | |
| "grad_norm": 7.754778118919807, | |
| "learning_rate": 6.999608356335957e-07, | |
| "logits/chosen": -1.0677528381347656, | |
| "logits/rejected": -1.0721509456634521, | |
| "logps/chosen": -227.729248046875, | |
| "logps/rejected": -239.76138305664062, | |
| "loss": 0.645, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -0.2937619090080261, | |
| "rewards/margins": 0.048121094703674316, | |
| "rewards/rejected": -0.3418830335140228, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.10686615014694095, | |
| "grad_norm": 6.437697145095338, | |
| "learning_rate": 6.999118822298773e-07, | |
| "logits/chosen": -1.1454989910125732, | |
| "logits/rejected": -1.0754855871200562, | |
| "logps/chosen": -227.07974243164062, | |
| "logps/rejected": -228.60333251953125, | |
| "loss": 0.6547, | |
| "rewards/accuracies": 0.4375, | |
| "rewards/chosen": -0.2852881848812103, | |
| "rewards/margins": 0.0602235309779644, | |
| "rewards/rejected": -0.3455117344856262, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.10900347314987978, | |
| "grad_norm": 7.527535438459104, | |
| "learning_rate": 6.998433512992261e-07, | |
| "logits/chosen": -1.1637341976165771, | |
| "logits/rejected": -1.0740972757339478, | |
| "logps/chosen": -292.79742431640625, | |
| "logps/rejected": -302.4303283691406, | |
| "loss": 0.6493, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -0.5110954642295837, | |
| "rewards/margins": 0.09770005941390991, | |
| "rewards/rejected": -0.6087955236434937, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.11114079615281859, | |
| "grad_norm": 7.575162846371126, | |
| "learning_rate": 6.997552466759395e-07, | |
| "logits/chosen": -1.0683425664901733, | |
| "logits/rejected": -1.1101981401443481, | |
| "logps/chosen": -297.38702392578125, | |
| "logps/rejected": -309.9881286621094, | |
| "loss": 0.6288, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -0.4417276978492737, | |
| "rewards/margins": 0.09606233239173889, | |
| "rewards/rejected": -0.537790060043335, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.11327811915575742, | |
| "grad_norm": 7.584356018239797, | |
| "learning_rate": 6.996475732894601e-07, | |
| "logits/chosen": -1.0953909158706665, | |
| "logits/rejected": -1.0685151815414429, | |
| "logps/chosen": -336.0571594238281, | |
| "logps/rejected": -369.7088623046875, | |
| "loss": 0.6369, | |
| "rewards/accuracies": 0.5625, | |
| "rewards/chosen": -0.6481105089187622, | |
| "rewards/margins": 0.13110791146755219, | |
| "rewards/rejected": -0.779218316078186, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.11541544215869623, | |
| "grad_norm": 6.778514980526427, | |
| "learning_rate": 6.995203371641008e-07, | |
| "logits/chosen": -1.1196974515914917, | |
| "logits/rejected": -1.0348316431045532, | |
| "logps/chosen": -287.28729248046875, | |
| "logps/rejected": -302.94317626953125, | |
| "loss": 0.6444, | |
| "rewards/accuracies": 0.5, | |
| "rewards/chosen": -0.5020595788955688, | |
| "rewards/margins": 0.09727220237255096, | |
| "rewards/rejected": -0.5993317365646362, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.11755276516163506, | |
| "grad_norm": 8.211287073219191, | |
| "learning_rate": 6.993735454187072e-07, | |
| "logits/chosen": -1.0214238166809082, | |
| "logits/rejected": -0.9623696804046631, | |
| "logps/chosen": -265.59417724609375, | |
| "logps/rejected": -298.86444091796875, | |
| "loss": 0.6016, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -0.5154373645782471, | |
| "rewards/margins": 0.24744802713394165, | |
| "rewards/rejected": -0.7628854513168335, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.11969008816457387, | |
| "grad_norm": 7.2415969346538756, | |
| "learning_rate": 6.992072062662596e-07, | |
| "logits/chosen": -0.9752060174942017, | |
| "logits/rejected": -0.9751684069633484, | |
| "logps/chosen": -349.1664123535156, | |
| "logps/rejected": -362.6107177734375, | |
| "loss": 0.6343, | |
| "rewards/accuracies": 0.5625, | |
| "rewards/chosen": -0.7859246134757996, | |
| "rewards/margins": 0.13539882004261017, | |
| "rewards/rejected": -0.9213234186172485, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.1218274111675127, | |
| "grad_norm": 6.635074198180161, | |
| "learning_rate": 6.99021329013413e-07, | |
| "logits/chosen": -1.1370452642440796, | |
| "logits/rejected": -1.0629245042800903, | |
| "logps/chosen": -288.10430908203125, | |
| "logps/rejected": -307.46484375, | |
| "loss": 0.61, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -0.6563497185707092, | |
| "rewards/margins": 0.21564731001853943, | |
| "rewards/rejected": -0.8719969987869263, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.12396473417045151, | |
| "grad_norm": 7.22705417609884, | |
| "learning_rate": 6.988159240599771e-07, | |
| "logits/chosen": -1.1842763423919678, | |
| "logits/rejected": -1.1428574323654175, | |
| "logps/chosen": -288.5982666015625, | |
| "logps/rejected": -291.8437194824219, | |
| "loss": 0.6186, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -0.7895088195800781, | |
| "rewards/margins": 0.07229967415332794, | |
| "rewards/rejected": -0.8618084788322449, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.12610205717339032, | |
| "grad_norm": 7.191134813555289, | |
| "learning_rate": 6.985910028983336e-07, | |
| "logits/chosen": -1.0885567665100098, | |
| "logits/rejected": -1.0549125671386719, | |
| "logps/chosen": -484.6271667480469, | |
| "logps/rejected": -517.6636352539062, | |
| "loss": 0.6022, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -1.1380163431167603, | |
| "rewards/margins": 0.33240166306495667, | |
| "rewards/rejected": -1.4704179763793945, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.12823938017632916, | |
| "grad_norm": 6.856697802256405, | |
| "learning_rate": 6.983465781127942e-07, | |
| "logits/chosen": -1.0434743165969849, | |
| "logits/rejected": -1.0580846071243286, | |
| "logps/chosen": -315.3133239746094, | |
| "logps/rejected": -314.5843811035156, | |
| "loss": 0.6142, | |
| "rewards/accuracies": 0.5, | |
| "rewards/chosen": -0.7601010799407959, | |
| "rewards/margins": 0.006043531000614166, | |
| "rewards/rejected": -0.7661447525024414, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.13037670317926797, | |
| "grad_norm": 6.8390931208538985, | |
| "learning_rate": 6.980826633788957e-07, | |
| "logits/chosen": -1.0378106832504272, | |
| "logits/rejected": -0.9883963465690613, | |
| "logps/chosen": -406.2007751464844, | |
| "logps/rejected": -471.77471923828125, | |
| "loss": 0.611, | |
| "rewards/accuracies": 0.5625, | |
| "rewards/chosen": -1.0759634971618652, | |
| "rewards/margins": 0.3877870440483093, | |
| "rewards/rejected": -1.4637506008148193, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.13251402618220678, | |
| "grad_norm": 8.256162255166101, | |
| "learning_rate": 6.977992734626349e-07, | |
| "logits/chosen": -1.2170382738113403, | |
| "logits/rejected": -1.1635799407958984, | |
| "logps/chosen": -409.862548828125, | |
| "logps/rejected": -459.107421875, | |
| "loss": 0.5815, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -1.3327953815460205, | |
| "rewards/margins": 0.5586321353912354, | |
| "rewards/rejected": -1.8914276361465454, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.1346513491851456, | |
| "grad_norm": 6.7905205812341185, | |
| "learning_rate": 6.97496424219643e-07, | |
| "logits/chosen": -0.9324991106987, | |
| "logits/rejected": -0.9242455959320068, | |
| "logps/chosen": -305.8074035644531, | |
| "logps/rejected": -359.38275146484375, | |
| "loss": 0.5897, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -0.7552638649940491, | |
| "rewards/margins": 0.3947301506996155, | |
| "rewards/rejected": -1.1499940156936646, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.13678867218808444, | |
| "grad_norm": 7.929159551201501, | |
| "learning_rate": 6.971741325942982e-07, | |
| "logits/chosen": -1.0476861000061035, | |
| "logits/rejected": -1.0574254989624023, | |
| "logps/chosen": -397.0540466308594, | |
| "logps/rejected": -445.9757385253906, | |
| "loss": 0.5989, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -1.1025577783584595, | |
| "rewards/margins": 0.5505799055099487, | |
| "rewards/rejected": -1.6531375646591187, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.13892599519102325, | |
| "grad_norm": 6.60300962363944, | |
| "learning_rate": 6.968324166187772e-07, | |
| "logits/chosen": -1.0173978805541992, | |
| "logits/rejected": -1.0060231685638428, | |
| "logps/chosen": -391.6329345703125, | |
| "logps/rejected": -482.0704040527344, | |
| "loss": 0.5949, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -1.2589941024780273, | |
| "rewards/margins": 0.7740709781646729, | |
| "rewards/rejected": -2.0330650806427, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.14106331819396206, | |
| "grad_norm": 7.722534638937169, | |
| "learning_rate": 6.96471295412047e-07, | |
| "logits/chosen": -1.210124135017395, | |
| "logits/rejected": -1.1669820547103882, | |
| "logps/chosen": -436.8033447265625, | |
| "logps/rejected": -447.0545349121094, | |
| "loss": 0.6197, | |
| "rewards/accuracies": 0.5625, | |
| "rewards/chosen": -1.7195076942443848, | |
| "rewards/margins": 0.12996703386306763, | |
| "rewards/rejected": -1.8494747877120972, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.14320064119690087, | |
| "grad_norm": 8.147380246018754, | |
| "learning_rate": 6.960907891787949e-07, | |
| "logits/chosen": -1.165956974029541, | |
| "logits/rejected": -1.1428234577178955, | |
| "logps/chosen": -519.3298950195312, | |
| "logps/rejected": -559.141357421875, | |
| "loss": 0.547, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -2.083847999572754, | |
| "rewards/margins": 0.4430221617221832, | |
| "rewards/rejected": -2.5268702507019043, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.14533796419983971, | |
| "grad_norm": 7.8944938583131945, | |
| "learning_rate": 6.956909192082982e-07, | |
| "logits/chosen": -0.9343075752258301, | |
| "logits/rejected": -0.8972913026809692, | |
| "logps/chosen": -292.84552001953125, | |
| "logps/rejected": -341.25872802734375, | |
| "loss": 0.5741, | |
| "rewards/accuracies": 0.5625, | |
| "rewards/chosen": -0.985328197479248, | |
| "rewards/margins": 0.39189204573631287, | |
| "rewards/rejected": -1.3772202730178833, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.14747528720277853, | |
| "grad_norm": 7.019821529750172, | |
| "learning_rate": 6.952717078732324e-07, | |
| "logits/chosen": -1.0702323913574219, | |
| "logits/rejected": -0.9579718112945557, | |
| "logps/chosen": -378.2999572753906, | |
| "logps/rejected": -452.3642578125, | |
| "loss": 0.6007, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -1.3012688159942627, | |
| "rewards/margins": 0.5469968318939209, | |
| "rewards/rejected": -1.8482656478881836, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.14961261020571734, | |
| "grad_norm": 7.623144170728904, | |
| "learning_rate": 6.948331786284207e-07, | |
| "logits/chosen": -1.0306965112686157, | |
| "logits/rejected": -1.0405011177062988, | |
| "logps/chosen": -489.2743835449219, | |
| "logps/rejected": -580.343994140625, | |
| "loss": 0.5627, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -2.306447982788086, | |
| "rewards/margins": 0.7942487001419067, | |
| "rewards/rejected": -3.100696563720703, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.15174993320865615, | |
| "grad_norm": 7.340408773148345, | |
| "learning_rate": 6.943753560095204e-07, | |
| "logits/chosen": -1.1813827753067017, | |
| "logits/rejected": -1.1872248649597168, | |
| "logps/chosen": -516.26953125, | |
| "logps/rejected": -557.9083251953125, | |
| "loss": 0.563, | |
| "rewards/accuracies": 0.5625, | |
| "rewards/chosen": -2.189997434616089, | |
| "rewards/margins": 0.49108394980430603, | |
| "rewards/rejected": -2.681081533432007, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.15388725621159496, | |
| "grad_norm": 9.09254480811679, | |
| "learning_rate": 6.938982656316509e-07, | |
| "logits/chosen": -1.108089566230774, | |
| "logits/rejected": -1.1164076328277588, | |
| "logps/chosen": -483.02099609375, | |
| "logps/rejected": -499.59503173828125, | |
| "loss": 0.5837, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -1.988714575767517, | |
| "rewards/margins": 0.3150157928466797, | |
| "rewards/rejected": -2.3037304878234863, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.1560245792145338, | |
| "grad_norm": 10.656479693855966, | |
| "learning_rate": 6.934019341879603e-07, | |
| "logits/chosen": -0.9998635649681091, | |
| "logits/rejected": -0.9844257831573486, | |
| "logps/chosen": -449.13592529296875, | |
| "logps/rejected": -543.6675415039062, | |
| "loss": 0.6031, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -1.9461640119552612, | |
| "rewards/margins": 0.7130802869796753, | |
| "rewards/rejected": -2.6592442989349365, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.15816190221747262, | |
| "grad_norm": 10.135891828278513, | |
| "learning_rate": 6.928863894481325e-07, | |
| "logits/chosen": -1.057843804359436, | |
| "logits/rejected": -1.0019515752792358, | |
| "logps/chosen": -499.46343994140625, | |
| "logps/rejected": -607.9090576171875, | |
| "loss": 0.6289, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -1.925887942314148, | |
| "rewards/margins": 0.7596069574356079, | |
| "rewards/rejected": -2.685494899749756, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.16029922522041143, | |
| "grad_norm": 11.196785279167383, | |
| "learning_rate": 6.923516602568319e-07, | |
| "logits/chosen": -1.10731840133667, | |
| "logits/rejected": -1.1392724514007568, | |
| "logps/chosen": -572.486083984375, | |
| "logps/rejected": -581.3994140625, | |
| "loss": 0.6307, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -2.843285322189331, | |
| "rewards/margins": 0.3052109479904175, | |
| "rewards/rejected": -3.148496150970459, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.16243654822335024, | |
| "grad_norm": 8.037071469258775, | |
| "learning_rate": 6.917977765320918e-07, | |
| "logits/chosen": -1.0746040344238281, | |
| "logits/rejected": -1.0597764253616333, | |
| "logps/chosen": -504.6460266113281, | |
| "logps/rejected": -606.507080078125, | |
| "loss": 0.5222, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.3728506565093994, | |
| "rewards/margins": 0.8532505035400391, | |
| "rewards/rejected": -3.2261009216308594, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.16457387122628908, | |
| "grad_norm": 8.046205826147432, | |
| "learning_rate": 6.912247692636382e-07, | |
| "logits/chosen": -0.9819946885108948, | |
| "logits/rejected": -0.9038519859313965, | |
| "logps/chosen": -402.05523681640625, | |
| "logps/rejected": -466.16973876953125, | |
| "loss": 0.5378, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -1.7473161220550537, | |
| "rewards/margins": 0.5481730699539185, | |
| "rewards/rejected": -2.2954893112182617, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.1667111942292279, | |
| "grad_norm": 8.442110303754333, | |
| "learning_rate": 6.90632670511158e-07, | |
| "logits/chosen": -1.0492621660232544, | |
| "logits/rejected": -1.0028660297393799, | |
| "logps/chosen": -452.12823486328125, | |
| "logps/rejected": -512.2908935546875, | |
| "loss": 0.5626, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.2684264183044434, | |
| "rewards/margins": 0.5885436534881592, | |
| "rewards/rejected": -2.8569703102111816, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.1688485172321667, | |
| "grad_norm": 8.834706933620803, | |
| "learning_rate": 6.900215134025036e-07, | |
| "logits/chosen": -0.9072580933570862, | |
| "logits/rejected": -0.896135687828064, | |
| "logps/chosen": -446.088134765625, | |
| "logps/rejected": -527.0875854492188, | |
| "loss": 0.5476, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.0337717533111572, | |
| "rewards/margins": 0.5865947604179382, | |
| "rewards/rejected": -2.6203665733337402, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.17098584023510552, | |
| "grad_norm": 8.951583471345572, | |
| "learning_rate": 6.893913321318404e-07, | |
| "logits/chosen": -0.9730253219604492, | |
| "logits/rejected": -0.988685131072998, | |
| "logps/chosen": -460.3742370605469, | |
| "logps/rejected": -495.2950134277344, | |
| "loss": 0.5833, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.2925660610198975, | |
| "rewards/margins": 0.45126643776893616, | |
| "rewards/rejected": -2.743832588195801, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.17312316323804436, | |
| "grad_norm": 10.31156332894422, | |
| "learning_rate": 6.887421619577333e-07, | |
| "logits/chosen": -0.9978621602058411, | |
| "logits/rejected": -0.914361298084259, | |
| "logps/chosen": -577.339111328125, | |
| "logps/rejected": -620.412353515625, | |
| "loss": 0.5333, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.5646235942840576, | |
| "rewards/margins": 0.5104097127914429, | |
| "rewards/rejected": -3.075033187866211, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.17526048624098317, | |
| "grad_norm": 8.242407615983574, | |
| "learning_rate": 6.880740392011738e-07, | |
| "logits/chosen": -1.0842255353927612, | |
| "logits/rejected": -1.0887690782546997, | |
| "logps/chosen": -484.07733154296875, | |
| "logps/rejected": -524.2720947265625, | |
| "loss": 0.5808, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -2.4766807556152344, | |
| "rewards/margins": 0.48678654432296753, | |
| "rewards/rejected": -2.963467597961426, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.17739780924392198, | |
| "grad_norm": 9.401026195070148, | |
| "learning_rate": 6.873870012435486e-07, | |
| "logits/chosen": -1.0653420686721802, | |
| "logits/rejected": -1.0636709928512573, | |
| "logps/chosen": -476.5955810546875, | |
| "logps/rejected": -574.3685302734375, | |
| "loss": 0.5926, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -2.0014445781707764, | |
| "rewards/margins": 1.0705955028533936, | |
| "rewards/rejected": -3.072040319442749, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.1795351322468608, | |
| "grad_norm": 9.603090833181401, | |
| "learning_rate": 6.866810865245471e-07, | |
| "logits/chosen": -1.1408371925354004, | |
| "logits/rejected": -1.2083356380462646, | |
| "logps/chosen": -481.21270751953125, | |
| "logps/rejected": -499.5158386230469, | |
| "loss": 0.5634, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.340351104736328, | |
| "rewards/margins": 0.2937456965446472, | |
| "rewards/rejected": -2.634096622467041, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.18167245524979964, | |
| "grad_norm": 8.394856753842978, | |
| "learning_rate": 6.859563345400111e-07, | |
| "logits/chosen": -1.0628814697265625, | |
| "logits/rejected": -0.9136342406272888, | |
| "logps/chosen": -399.5640869140625, | |
| "logps/rejected": -533.1271362304688, | |
| "loss": 0.5817, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -1.9623302221298218, | |
| "rewards/margins": 1.1472116708755493, | |
| "rewards/rejected": -3.109541893005371, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.18380977825273845, | |
| "grad_norm": 8.518782849174428, | |
| "learning_rate": 6.852127858397256e-07, | |
| "logits/chosen": -1.066598892211914, | |
| "logits/rejected": -1.0352439880371094, | |
| "logps/chosen": -561.7534790039062, | |
| "logps/rejected": -711.9557495117188, | |
| "loss": 0.521, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -2.871816396713257, | |
| "rewards/margins": 1.3357863426208496, | |
| "rewards/rejected": -4.207602500915527, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.18594710125567726, | |
| "grad_norm": 9.531847125052328, | |
| "learning_rate": 6.844504820251493e-07, | |
| "logits/chosen": -0.8711273670196533, | |
| "logits/rejected": -0.7599645853042603, | |
| "logps/chosen": -630.6957397460938, | |
| "logps/rejected": -768.0391845703125, | |
| "loss": 0.5164, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.074097156524658, | |
| "rewards/margins": 1.3850178718566895, | |
| "rewards/rejected": -4.459115028381348, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.18808442425861607, | |
| "grad_norm": 11.193362121063247, | |
| "learning_rate": 6.836694657470869e-07, | |
| "logits/chosen": -1.0319950580596924, | |
| "logits/rejected": -0.9569318294525146, | |
| "logps/chosen": -283.17431640625, | |
| "logps/rejected": -333.98797607421875, | |
| "loss": 0.5972, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -1.3249003887176514, | |
| "rewards/margins": 0.47719448804855347, | |
| "rewards/rejected": -1.8020949363708496, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.1902217472615549, | |
| "grad_norm": 12.5393990232581, | |
| "learning_rate": 6.828697807033038e-07, | |
| "logits/chosen": -1.016239881515503, | |
| "logits/rejected": -0.9772812724113464, | |
| "logps/chosen": -387.52056884765625, | |
| "logps/rejected": -477.847412109375, | |
| "loss": 0.5869, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -1.845961332321167, | |
| "rewards/margins": 0.6933265328407288, | |
| "rewards/rejected": -2.539287805557251, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.19235907026449373, | |
| "grad_norm": 11.082808377296098, | |
| "learning_rate": 6.820514716360799e-07, | |
| "logits/chosen": -0.9989445209503174, | |
| "logits/rejected": -1.0149664878845215, | |
| "logps/chosen": -550.3873291015625, | |
| "logps/rejected": -627.486328125, | |
| "loss": 0.5262, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -2.6533279418945312, | |
| "rewards/margins": 0.7726477980613708, | |
| "rewards/rejected": -3.425976037979126, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.19449639326743254, | |
| "grad_norm": 10.136450815566235, | |
| "learning_rate": 6.812145843297075e-07, | |
| "logits/chosen": -0.9787087440490723, | |
| "logits/rejected": -0.9385868310928345, | |
| "logps/chosen": -541.9707641601562, | |
| "logps/rejected": -582.03759765625, | |
| "loss": 0.537, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -2.860710382461548, | |
| "rewards/margins": 0.4204621911048889, | |
| "rewards/rejected": -3.281172513961792, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.19663371627037135, | |
| "grad_norm": 9.876929179683522, | |
| "learning_rate": 6.803591656079286e-07, | |
| "logits/chosen": -1.1552644968032837, | |
| "logits/rejected": -1.1622107028961182, | |
| "logps/chosen": -502.1416015625, | |
| "logps/rejected": -571.0797729492188, | |
| "loss": 0.5405, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.547774076461792, | |
| "rewards/margins": 0.6767189502716064, | |
| "rewards/rejected": -3.2244930267333984, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.1987710392733102, | |
| "grad_norm": 8.350168280012312, | |
| "learning_rate": 6.794852633313161e-07, | |
| "logits/chosen": -0.8592644333839417, | |
| "logits/rejected": -0.8664505481719971, | |
| "logps/chosen": -385.8682556152344, | |
| "logps/rejected": -391.97564697265625, | |
| "loss": 0.5449, | |
| "rewards/accuracies": 0.5, | |
| "rewards/chosen": -1.7356358766555786, | |
| "rewards/margins": 0.2472929060459137, | |
| "rewards/rejected": -1.9829288721084595, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.200908362276249, | |
| "grad_norm": 10.381869975048419, | |
| "learning_rate": 6.785929263945951e-07, | |
| "logits/chosen": -0.9588985443115234, | |
| "logits/rejected": -0.982033371925354, | |
| "logps/chosen": -562.7421875, | |
| "logps/rejected": -626.1223754882812, | |
| "loss": 0.5568, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.8523409366607666, | |
| "rewards/margins": 0.6437615752220154, | |
| "rewards/rejected": -3.4961023330688477, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.20304568527918782, | |
| "grad_norm": 8.994516081218443, | |
| "learning_rate": 6.776822047239079e-07, | |
| "logits/chosen": -1.0020171403884888, | |
| "logits/rejected": -0.9751790761947632, | |
| "logps/chosen": -417.7941589355469, | |
| "logps/rejected": -482.18511962890625, | |
| "loss": 0.5187, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.281278610229492, | |
| "rewards/margins": 0.5854904651641846, | |
| "rewards/rejected": -2.866769313812256, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.20518300828212663, | |
| "grad_norm": 9.433844383462462, | |
| "learning_rate": 6.767531492740206e-07, | |
| "logits/chosen": -1.1744908094406128, | |
| "logits/rejected": -1.152523159980774, | |
| "logps/chosen": -517.0827026367188, | |
| "logps/rejected": -538.0679931640625, | |
| "loss": 0.5032, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -1.9695494174957275, | |
| "rewards/margins": 0.5735199451446533, | |
| "rewards/rejected": -2.543069362640381, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.20732033128506547, | |
| "grad_norm": 12.358671729964394, | |
| "learning_rate": 6.758058120254714e-07, | |
| "logits/chosen": -1.0569499731063843, | |
| "logits/rejected": -0.9990131855010986, | |
| "logps/chosen": -499.7625427246094, | |
| "logps/rejected": -545.9909057617188, | |
| "loss": 0.5433, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -2.38741397857666, | |
| "rewards/margins": 0.40309280157089233, | |
| "rewards/rejected": -2.7905068397521973, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.20945765428800428, | |
| "grad_norm": 13.359908029714509, | |
| "learning_rate": 6.748402459816636e-07, | |
| "logits/chosen": -0.9123142957687378, | |
| "logits/rejected": -0.9710780382156372, | |
| "logps/chosen": -639.8603515625, | |
| "logps/rejected": -638.2657470703125, | |
| "loss": 0.5369, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -3.3181374073028564, | |
| "rewards/margins": 0.10219268500804901, | |
| "rewards/rejected": -3.420330286026001, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.2115949772909431, | |
| "grad_norm": 9.360941415209796, | |
| "learning_rate": 6.738565051658988e-07, | |
| "logits/chosen": -0.9268811941146851, | |
| "logits/rejected": -1.0143418312072754, | |
| "logps/chosen": -504.1163330078125, | |
| "logps/rejected": -515.728271484375, | |
| "loss": 0.5485, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -2.5163679122924805, | |
| "rewards/margins": 0.32084643840789795, | |
| "rewards/rejected": -2.837214469909668, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.2137323002938819, | |
| "grad_norm": 10.094138733652892, | |
| "learning_rate": 6.728546446183552e-07, | |
| "logits/chosen": -0.6902378797531128, | |
| "logits/rejected": -0.7625846266746521, | |
| "logps/chosen": -370.24859619140625, | |
| "logps/rejected": -434.644287109375, | |
| "loss": 0.5813, | |
| "rewards/accuracies": 0.5625, | |
| "rewards/chosen": -1.854517936706543, | |
| "rewards/margins": 0.39516037702560425, | |
| "rewards/rejected": -2.249678373336792, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.2137323002938819, | |
| "eval_logits/chosen": -1.1082046031951904, | |
| "eval_logits/rejected": -1.1243932247161865, | |
| "eval_logps/chosen": -541.2933349609375, | |
| "eval_logps/rejected": -612.4234008789062, | |
| "eval_loss": 0.5237933397293091, | |
| "eval_rewards/accuracies": 0.7298387289047241, | |
| "eval_rewards/chosen": -2.681633710861206, | |
| "eval_rewards/margins": 0.7722625732421875, | |
| "eval_rewards/rejected": -3.4538962841033936, | |
| "eval_runtime": 141.8383, | |
| "eval_samples_per_second": 13.826, | |
| "eval_steps_per_second": 0.437, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.21586962329682075, | |
| "grad_norm": 13.595024692594054, | |
| "learning_rate": 6.718347203930077e-07, | |
| "logits/chosen": -1.0973420143127441, | |
| "logits/rejected": -1.1126675605773926, | |
| "logps/chosen": -547.3043823242188, | |
| "logps/rejected": -647.8222045898438, | |
| "loss": 0.6014, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.594785690307617, | |
| "rewards/margins": 1.0374321937561035, | |
| "rewards/rejected": -3.6322176456451416, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.21800694629975956, | |
| "grad_norm": 10.896734874060659, | |
| "learning_rate": 6.707967895544916e-07, | |
| "logits/chosen": -1.0957213640213013, | |
| "logits/rejected": -0.958062469959259, | |
| "logps/chosen": -466.9610595703125, | |
| "logps/rejected": -579.3587036132812, | |
| "loss": 0.4572, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -2.131477117538452, | |
| "rewards/margins": 0.9194568395614624, | |
| "rewards/rejected": -3.050933837890625, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.22014426930269837, | |
| "grad_norm": 7.969619734080228, | |
| "learning_rate": 6.697409101749103e-07, | |
| "logits/chosen": -0.9213029146194458, | |
| "logits/rejected": -0.8857940435409546, | |
| "logps/chosen": -506.6609191894531, | |
| "logps/rejected": -597.8681640625, | |
| "loss": 0.4681, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.1986255645751953, | |
| "rewards/margins": 0.973738431930542, | |
| "rewards/rejected": -3.172363758087158, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.22228159230563718, | |
| "grad_norm": 8.749688164160379, | |
| "learning_rate": 6.686671413305856e-07, | |
| "logits/chosen": -1.1210659742355347, | |
| "logits/rejected": -1.0568785667419434, | |
| "logps/chosen": -527.4649658203125, | |
| "logps/rejected": -666.8826904296875, | |
| "loss": 0.4962, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.895803213119507, | |
| "rewards/margins": 1.2766165733337402, | |
| "rewards/rejected": -4.172420024871826, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.224418915308576, | |
| "grad_norm": 10.816785495540321, | |
| "learning_rate": 6.675755430987527e-07, | |
| "logits/chosen": -1.098203420639038, | |
| "logits/rejected": -1.043796181678772, | |
| "logps/chosen": -406.22186279296875, | |
| "logps/rejected": -515.1035766601562, | |
| "loss": 0.5235, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -2.0460591316223145, | |
| "rewards/margins": 0.9124148488044739, | |
| "rewards/rejected": -2.9584736824035645, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.22655623831151483, | |
| "grad_norm": 16.006357368038902, | |
| "learning_rate": 6.66466176554199e-07, | |
| "logits/chosen": -1.1074966192245483, | |
| "logits/rejected": -1.1250437498092651, | |
| "logps/chosen": -511.63519287109375, | |
| "logps/rejected": -583.53759765625, | |
| "loss": 0.6286, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -2.4097325801849365, | |
| "rewards/margins": 0.7932775020599365, | |
| "rewards/rejected": -3.203009843826294, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.22869356131445365, | |
| "grad_norm": 10.145999208049496, | |
| "learning_rate": 6.653391037658466e-07, | |
| "logits/chosen": -0.8575969934463501, | |
| "logits/rejected": -0.8596129417419434, | |
| "logps/chosen": -600.901123046875, | |
| "logps/rejected": -702.7192993164062, | |
| "loss": 0.4988, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.2228050231933594, | |
| "rewards/margins": 0.8592959642410278, | |
| "rewards/rejected": -4.082100868225098, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.23083088431739246, | |
| "grad_norm": 9.90205874296923, | |
| "learning_rate": 6.641943877932801e-07, | |
| "logits/chosen": -0.982778787612915, | |
| "logits/rejected": -0.967957615852356, | |
| "logps/chosen": -637.15478515625, | |
| "logps/rejected": -736.35546875, | |
| "loss": 0.5006, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.472803831100464, | |
| "rewards/margins": 0.9627398252487183, | |
| "rewards/rejected": -4.435543537139893, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.23296820732033127, | |
| "grad_norm": 10.737852627360994, | |
| "learning_rate": 6.630320926832178e-07, | |
| "logits/chosen": -1.0073411464691162, | |
| "logits/rejected": -0.9530993700027466, | |
| "logps/chosen": -543.8530883789062, | |
| "logps/rejected": -669.4638061523438, | |
| "loss": 0.5518, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.8368093967437744, | |
| "rewards/margins": 1.1880838871002197, | |
| "rewards/rejected": -4.024893283843994, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.2351055303232701, | |
| "grad_norm": 10.25518470798023, | |
| "learning_rate": 6.618522834659287e-07, | |
| "logits/chosen": -1.0670475959777832, | |
| "logits/rejected": -0.9968966245651245, | |
| "logps/chosen": -551.1478271484375, | |
| "logps/rejected": -710.794921875, | |
| "loss": 0.4774, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.1101412773132324, | |
| "rewards/margins": 1.3935075998306274, | |
| "rewards/rejected": -4.50364875793457, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.23724285332620892, | |
| "grad_norm": 10.88392855187359, | |
| "learning_rate": 6.606550261515939e-07, | |
| "logits/chosen": -1.1226112842559814, | |
| "logits/rejected": -1.1211527585983276, | |
| "logps/chosen": -496.4976806640625, | |
| "logps/rejected": -598.9983520507812, | |
| "loss": 0.5214, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.4094114303588867, | |
| "rewards/margins": 0.967483639717102, | |
| "rewards/rejected": -3.37689471244812, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.23938017632914774, | |
| "grad_norm": 14.5461852961629, | |
| "learning_rate": 6.594403877266134e-07, | |
| "logits/chosen": -0.9625245332717896, | |
| "logits/rejected": -1.0075706243515015, | |
| "logps/chosen": -479.2593688964844, | |
| "logps/rejected": -569.9885864257812, | |
| "loss": 0.5683, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -2.5545759201049805, | |
| "rewards/margins": 0.8655453324317932, | |
| "rewards/rejected": -3.420121431350708, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.24151749933208655, | |
| "grad_norm": 10.257155762835067, | |
| "learning_rate": 6.582084361498582e-07, | |
| "logits/chosen": -1.0782970190048218, | |
| "logits/rejected": -1.0272691249847412, | |
| "logps/chosen": -431.1832580566406, | |
| "logps/rejected": -507.42401123046875, | |
| "loss": 0.4517, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.3007423877716064, | |
| "rewards/margins": 0.6344060301780701, | |
| "rewards/rejected": -2.935148239135742, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.2436548223350254, | |
| "grad_norm": 12.281013811232151, | |
| "learning_rate": 6.569592403488683e-07, | |
| "logits/chosen": -0.8818793296813965, | |
| "logits/rejected": -0.8034133315086365, | |
| "logps/chosen": -575.656005859375, | |
| "logps/rejected": -665.9652709960938, | |
| "loss": 0.5607, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.9418675899505615, | |
| "rewards/margins": 0.8054768443107605, | |
| "rewards/rejected": -3.747344493865967, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.2457921453379642, | |
| "grad_norm": 12.328703519510336, | |
| "learning_rate": 6.556928702159954e-07, | |
| "logits/chosen": -1.0138773918151855, | |
| "logits/rejected": -0.9439175128936768, | |
| "logps/chosen": -389.8624267578125, | |
| "logps/rejected": -458.929443359375, | |
| "loss": 0.5415, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -1.944897174835205, | |
| "rewards/margins": 0.5071231126785278, | |
| "rewards/rejected": -2.4520201683044434, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.24792946834090301, | |
| "grad_norm": 11.019958842452935, | |
| "learning_rate": 6.544093966044933e-07, | |
| "logits/chosen": -1.0471012592315674, | |
| "logits/rejected": -1.046525001525879, | |
| "logps/chosen": -589.4481811523438, | |
| "logps/rejected": -666.8422241210938, | |
| "loss": 0.4744, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -2.920729160308838, | |
| "rewards/margins": 0.9037714004516602, | |
| "rewards/rejected": -3.824500560760498, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.25006679134384185, | |
| "grad_norm": 12.320353833685262, | |
| "learning_rate": 6.531088913245536e-07, | |
| "logits/chosen": -0.9775707721710205, | |
| "logits/rejected": -1.0437175035476685, | |
| "logps/chosen": -559.2511596679688, | |
| "logps/rejected": -629.9223022460938, | |
| "loss": 0.5403, | |
| "rewards/accuracies": 0.5625, | |
| "rewards/chosen": -2.863795280456543, | |
| "rewards/margins": 0.9548764228820801, | |
| "rewards/rejected": -3.818671703338623, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.25220411434678064, | |
| "grad_norm": 13.023997265472861, | |
| "learning_rate": 6.517914271392868e-07, | |
| "logits/chosen": -0.9753481149673462, | |
| "logits/rejected": -0.9616096019744873, | |
| "logps/chosen": -393.8183288574219, | |
| "logps/rejected": -467.1902770996094, | |
| "loss": 0.5117, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -2.0589523315429688, | |
| "rewards/margins": 0.6962214708328247, | |
| "rewards/rejected": -2.755174160003662, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.2543414373497195, | |
| "grad_norm": 10.577094903989462, | |
| "learning_rate": 6.504570777606531e-07, | |
| "logits/chosen": -1.1235861778259277, | |
| "logits/rejected": -1.0339062213897705, | |
| "logps/chosen": -567.0230712890625, | |
| "logps/rejected": -711.5052490234375, | |
| "loss": 0.5027, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.0050158500671387, | |
| "rewards/margins": 1.1521648168563843, | |
| "rewards/rejected": -4.1571807861328125, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.2564787603526583, | |
| "grad_norm": 10.49693020043088, | |
| "learning_rate": 6.491059178453366e-07, | |
| "logits/chosen": -0.9402109980583191, | |
| "logits/rejected": -0.9911918044090271, | |
| "logps/chosen": -524.8615112304688, | |
| "logps/rejected": -612.2509765625, | |
| "loss": 0.4693, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -2.6408395767211914, | |
| "rewards/margins": 0.9583368301391602, | |
| "rewards/rejected": -3.5991761684417725, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.2586160833555971, | |
| "grad_norm": 11.423558750572097, | |
| "learning_rate": 6.477380229905688e-07, | |
| "logits/chosen": -1.072409987449646, | |
| "logits/rejected": -1.1194437742233276, | |
| "logps/chosen": -561.1153564453125, | |
| "logps/rejected": -709.1156616210938, | |
| "loss": 0.4927, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -2.7978341579437256, | |
| "rewards/margins": 1.542366862297058, | |
| "rewards/rejected": -4.340201377868652, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.26075340635853594, | |
| "grad_norm": 11.084946490602835, | |
| "learning_rate": 6.463534697298995e-07, | |
| "logits/chosen": -0.9931835532188416, | |
| "logits/rejected": -0.8609973788261414, | |
| "logps/chosen": -422.6210632324219, | |
| "logps/rejected": -529.1079711914062, | |
| "loss": 0.4607, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -1.9958739280700684, | |
| "rewards/margins": 0.8684347867965698, | |
| "rewards/rejected": -2.8643088340759277, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.26289072936147473, | |
| "grad_norm": 12.662380910929564, | |
| "learning_rate": 6.449523355289137e-07, | |
| "logits/chosen": -1.0383917093276978, | |
| "logits/rejected": -1.0384039878845215, | |
| "logps/chosen": -552.278076171875, | |
| "logps/rejected": -629.6370849609375, | |
| "loss": 0.5599, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -2.6315407752990723, | |
| "rewards/margins": 0.8207972049713135, | |
| "rewards/rejected": -3.4523377418518066, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.26502805236441357, | |
| "grad_norm": 12.221280650614627, | |
| "learning_rate": 6.435346987808983e-07, | |
| "logits/chosen": -1.0482319593429565, | |
| "logits/rejected": -1.0103380680084229, | |
| "logps/chosen": -566.80078125, | |
| "logps/rejected": -658.3344116210938, | |
| "loss": 0.5172, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.72255539894104, | |
| "rewards/margins": 1.0078556537628174, | |
| "rewards/rejected": -3.7304112911224365, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.2671653753673524, | |
| "grad_norm": 10.864584700405295, | |
| "learning_rate": 6.421006388024559e-07, | |
| "logits/chosen": -0.9935693740844727, | |
| "logits/rejected": -1.0035309791564941, | |
| "logps/chosen": -450.3325500488281, | |
| "logps/rejected": -518.1874389648438, | |
| "loss": 0.4811, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.3707103729248047, | |
| "rewards/margins": 0.6181319952011108, | |
| "rewards/rejected": -2.988842248916626, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.2693026983702912, | |
| "grad_norm": 12.045454242902174, | |
| "learning_rate": 6.406502358290663e-07, | |
| "logits/chosen": -1.1015965938568115, | |
| "logits/rejected": -1.08380126953125, | |
| "logps/chosen": -496.258056640625, | |
| "logps/rejected": -653.9327392578125, | |
| "loss": 0.5494, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -2.305727481842041, | |
| "rewards/margins": 1.4904249906539917, | |
| "rewards/rejected": -3.796152353286743, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.27144002137323003, | |
| "grad_norm": 18.364703541342568, | |
| "learning_rate": 6.391835710105981e-07, | |
| "logits/chosen": -0.941745400428772, | |
| "logits/rejected": -0.9719420671463013, | |
| "logps/chosen": -511.3363952636719, | |
| "logps/rejected": -570.1834716796875, | |
| "loss": 0.5391, | |
| "rewards/accuracies": 0.5625, | |
| "rewards/chosen": -2.4952189922332764, | |
| "rewards/margins": 0.8485791683197021, | |
| "rewards/rejected": -3.3437979221343994, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.2735773443761689, | |
| "grad_norm": 11.918550043243753, | |
| "learning_rate": 6.377007264067684e-07, | |
| "logits/chosen": -0.9891614317893982, | |
| "logits/rejected": -0.9654213190078735, | |
| "logps/chosen": -455.00775146484375, | |
| "logps/rejected": -501.5223083496094, | |
| "loss": 0.5492, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.259676218032837, | |
| "rewards/margins": 0.5743985176086426, | |
| "rewards/rejected": -2.8340747356414795, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.27571466737910766, | |
| "grad_norm": 15.016960436290145, | |
| "learning_rate": 6.36201784982551e-07, | |
| "logits/chosen": -0.9479376673698425, | |
| "logits/rejected": -1.0105921030044556, | |
| "logps/chosen": -484.4723815917969, | |
| "logps/rejected": -579.9881591796875, | |
| "loss": 0.4852, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -2.237053871154785, | |
| "rewards/margins": 1.103590488433838, | |
| "rewards/rejected": -3.340644121170044, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.2778519903820465, | |
| "grad_norm": 10.336460981894195, | |
| "learning_rate": 6.346868306035347e-07, | |
| "logits/chosen": -1.0786710977554321, | |
| "logits/rejected": -1.0136293172836304, | |
| "logps/chosen": -531.4271850585938, | |
| "logps/rejected": -632.714599609375, | |
| "loss": 0.4997, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -2.405646562576294, | |
| "rewards/margins": 1.1792259216308594, | |
| "rewards/rejected": -3.5848724842071533, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.2799893133849853, | |
| "grad_norm": 13.583887221626354, | |
| "learning_rate": 6.331559480312316e-07, | |
| "logits/chosen": -0.8798739910125732, | |
| "logits/rejected": -0.9477823972702026, | |
| "logps/chosen": -596.0357055664062, | |
| "logps/rejected": -668.8670654296875, | |
| "loss": 0.4991, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -2.9465270042419434, | |
| "rewards/margins": 0.8481062054634094, | |
| "rewards/rejected": -3.794633150100708, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.2821266363879241, | |
| "grad_norm": 14.394323156381558, | |
| "learning_rate": 6.316092229183339e-07, | |
| "logits/chosen": -0.9489572048187256, | |
| "logits/rejected": -0.9146859645843506, | |
| "logps/chosen": -600.4107666015625, | |
| "logps/rejected": -661.3778076171875, | |
| "loss": 0.5529, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -3.2759668827056885, | |
| "rewards/margins": 0.545818567276001, | |
| "rewards/rejected": -3.8217854499816895, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.28426395939086296, | |
| "grad_norm": 11.385386930454487, | |
| "learning_rate": 6.300467418039222e-07, | |
| "logits/chosen": -0.960631787776947, | |
| "logits/rejected": -0.8502722382545471, | |
| "logps/chosen": -396.5357360839844, | |
| "logps/rejected": -541.553955078125, | |
| "loss": 0.5093, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -1.8977466821670532, | |
| "rewards/margins": 1.2623937129974365, | |
| "rewards/rejected": -3.1601402759552, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.28640128239380175, | |
| "grad_norm": 11.957264036760781, | |
| "learning_rate": 6.284685921086233e-07, | |
| "logits/chosen": -0.8640112280845642, | |
| "logits/rejected": -0.8963266611099243, | |
| "logps/chosen": -384.329345703125, | |
| "logps/rejected": -407.7789611816406, | |
| "loss": 0.498, | |
| "rewards/accuracies": 0.5625, | |
| "rewards/chosen": -2.0386128425598145, | |
| "rewards/margins": 0.3072957992553711, | |
| "rewards/rejected": -2.3459086418151855, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.2885386053967406, | |
| "grad_norm": 11.759955525639521, | |
| "learning_rate": 6.268748621297195e-07, | |
| "logits/chosen": -0.8902631998062134, | |
| "logits/rejected": -0.7775343656539917, | |
| "logps/chosen": -403.541015625, | |
| "logps/rejected": -534.9651489257812, | |
| "loss": 0.4668, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -1.9903576374053955, | |
| "rewards/margins": 1.1432865858078003, | |
| "rewards/rejected": -3.1336441040039062, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.29067592839967943, | |
| "grad_norm": 11.950216444117117, | |
| "learning_rate": 6.252656410362074e-07, | |
| "logits/chosen": -0.9792582988739014, | |
| "logits/rejected": -0.9614965915679932, | |
| "logps/chosen": -561.1026611328125, | |
| "logps/rejected": -663.7088012695312, | |
| "loss": 0.4924, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.82108998298645, | |
| "rewards/margins": 1.011682391166687, | |
| "rewards/rejected": -3.8327722549438477, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.2928132514026182, | |
| "grad_norm": 10.772908788991131, | |
| "learning_rate": 6.236410188638104e-07, | |
| "logits/chosen": -1.053274393081665, | |
| "logits/rejected": -0.9625363349914551, | |
| "logps/chosen": -572.0177001953125, | |
| "logps/rejected": -789.4164428710938, | |
| "loss": 0.5162, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -2.667423725128174, | |
| "rewards/margins": 2.069328546524048, | |
| "rewards/rejected": -4.736752510070801, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.29495057440555705, | |
| "grad_norm": 9.953041872331532, | |
| "learning_rate": 6.220010865099398e-07, | |
| "logits/chosen": -1.0631476640701294, | |
| "logits/rejected": -1.0828768014907837, | |
| "logps/chosen": -395.8948669433594, | |
| "logps/rejected": -478.61468505859375, | |
| "loss": 0.4991, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -2.0671029090881348, | |
| "rewards/margins": 0.7652881145477295, | |
| "rewards/rejected": -2.8323915004730225, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.29708789740849584, | |
| "grad_norm": 9.56597037699992, | |
| "learning_rate": 6.203459357286098e-07, | |
| "logits/chosen": -1.0143028497695923, | |
| "logits/rejected": -1.0154125690460205, | |
| "logps/chosen": -517.6961669921875, | |
| "logps/rejected": -596.7298583984375, | |
| "loss": 0.4613, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -2.6147890090942383, | |
| "rewards/margins": 0.7214634418487549, | |
| "rewards/rejected": -3.336252212524414, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.2992252204114347, | |
| "grad_norm": 13.08709351718772, | |
| "learning_rate": 6.186756591253038e-07, | |
| "logits/chosen": -0.9290235042572021, | |
| "logits/rejected": -0.9178426861763, | |
| "logps/chosen": -527.4471435546875, | |
| "logps/rejected": -568.3038330078125, | |
| "loss": 0.5523, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -2.790428638458252, | |
| "rewards/margins": 0.6073124408721924, | |
| "rewards/rejected": -3.3977413177490234, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.3013625434143735, | |
| "grad_norm": 9.930492612600617, | |
| "learning_rate": 6.169903501517932e-07, | |
| "logits/chosen": -0.9746488928794861, | |
| "logits/rejected": -0.9574599862098694, | |
| "logps/chosen": -426.73760986328125, | |
| "logps/rejected": -535.6727294921875, | |
| "loss": 0.5198, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.180694580078125, | |
| "rewards/margins": 0.977644145488739, | |
| "rewards/rejected": -3.158338785171509, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.3034998664173123, | |
| "grad_norm": 14.86115549874432, | |
| "learning_rate": 6.152901031009086e-07, | |
| "logits/chosen": -0.9366909265518188, | |
| "logits/rejected": -0.9158231019973755, | |
| "logps/chosen": -489.0951232910156, | |
| "logps/rejected": -574.0645141601562, | |
| "loss": 0.5373, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.401134967803955, | |
| "rewards/margins": 0.8239090442657471, | |
| "rewards/rejected": -3.2250442504882812, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.30563718942025114, | |
| "grad_norm": 18.2841162585259, | |
| "learning_rate": 6.135750131012639e-07, | |
| "logits/chosen": -0.9252594709396362, | |
| "logits/rejected": -0.8823320269584656, | |
| "logps/chosen": -500.0283203125, | |
| "logps/rejected": -582.243408203125, | |
| "loss": 0.5163, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.57209849357605, | |
| "rewards/margins": 0.7344849109649658, | |
| "rewards/rejected": -3.3065831661224365, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.3077745124231899, | |
| "grad_norm": 10.361990112128243, | |
| "learning_rate": 6.118451761119344e-07, | |
| "logits/chosen": -0.9340752363204956, | |
| "logits/rejected": -0.9739060997962952, | |
| "logps/chosen": -686.3743286132812, | |
| "logps/rejected": -827.3367919921875, | |
| "loss": 0.4274, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.355973243713379, | |
| "rewards/margins": 1.5688568353652954, | |
| "rewards/rejected": -4.924829483032227, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.30991183542612877, | |
| "grad_norm": 14.742711073001722, | |
| "learning_rate": 6.101006889170879e-07, | |
| "logits/chosen": -0.9780830144882202, | |
| "logits/rejected": -0.9924584031105042, | |
| "logps/chosen": -611.8880615234375, | |
| "logps/rejected": -783.9960327148438, | |
| "loss": 0.4892, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.101553440093994, | |
| "rewards/margins": 1.6389245986938477, | |
| "rewards/rejected": -4.740478038787842, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.3120491584290676, | |
| "grad_norm": 11.760151720162954, | |
| "learning_rate": 6.083416491205693e-07, | |
| "logits/chosen": -0.9051653146743774, | |
| "logits/rejected": -0.9137300252914429, | |
| "logps/chosen": -518.42333984375, | |
| "logps/rejected": -624.302490234375, | |
| "loss": 0.4934, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.439613103866577, | |
| "rewards/margins": 1.1922130584716797, | |
| "rewards/rejected": -3.631826162338257, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.3141864814320064, | |
| "grad_norm": 12.767953140987824, | |
| "learning_rate": 6.065681551404392e-07, | |
| "logits/chosen": -0.9767911434173584, | |
| "logits/rejected": -0.9007346034049988, | |
| "logps/chosen": -498.14361572265625, | |
| "logps/rejected": -551.3754272460938, | |
| "loss": 0.5575, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.6738038063049316, | |
| "rewards/margins": 0.7255844473838806, | |
| "rewards/rejected": -3.399388313293457, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.31632380443494523, | |
| "grad_norm": 11.792817714413513, | |
| "learning_rate": 6.047803062034688e-07, | |
| "logits/chosen": -1.0714683532714844, | |
| "logits/rejected": -1.0007489919662476, | |
| "logps/chosen": -524.3717041015625, | |
| "logps/rejected": -721.374267578125, | |
| "loss": 0.5315, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.6513381004333496, | |
| "rewards/margins": 1.8740673065185547, | |
| "rewards/rejected": -4.525405406951904, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.3184611274378841, | |
| "grad_norm": 10.394135670198176, | |
| "learning_rate": 6.029782023395871e-07, | |
| "logits/chosen": -0.99624103307724, | |
| "logits/rejected": -0.970421314239502, | |
| "logps/chosen": -520.40771484375, | |
| "logps/rejected": -551.2412109375, | |
| "loss": 0.4518, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.831399917602539, | |
| "rewards/margins": 0.5486774444580078, | |
| "rewards/rejected": -3.380077362060547, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.32059845044082286, | |
| "grad_norm": 14.89234343098954, | |
| "learning_rate": 6.011619443762841e-07, | |
| "logits/chosen": -1.1248900890350342, | |
| "logits/rejected": -1.1022447347640991, | |
| "logps/chosen": -546.7406005859375, | |
| "logps/rejected": -674.46484375, | |
| "loss": 0.5543, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.7904787063598633, | |
| "rewards/margins": 1.1329127550125122, | |
| "rewards/rejected": -3.923391342163086, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.3227357734437617, | |
| "grad_norm": 10.760811605793574, | |
| "learning_rate": 5.993316339329707e-07, | |
| "logits/chosen": -0.8464213013648987, | |
| "logits/rejected": -0.9315951466560364, | |
| "logps/chosen": -557.2722778320312, | |
| "logps/rejected": -562.7290649414062, | |
| "loss": 0.5073, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -3.0413103103637695, | |
| "rewards/margins": 0.24244722723960876, | |
| "rewards/rejected": -3.2837576866149902, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.3248730964467005, | |
| "grad_norm": 12.468671998929876, | |
| "learning_rate": 5.974873734152916e-07, | |
| "logits/chosen": -0.9382638931274414, | |
| "logits/rejected": -0.8903573751449585, | |
| "logps/chosen": -517.4348754882812, | |
| "logps/rejected": -586.1632690429688, | |
| "loss": 0.559, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.7158446311950684, | |
| "rewards/margins": 0.5773513317108154, | |
| "rewards/rejected": -3.293196201324463, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.3270104194496393, | |
| "grad_norm": 10.71297529597072, | |
| "learning_rate": 5.95629266009397e-07, | |
| "logits/chosen": -0.9615129232406616, | |
| "logits/rejected": -0.9363632202148438, | |
| "logps/chosen": -530.434814453125, | |
| "logps/rejected": -629.813720703125, | |
| "loss": 0.4965, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.732804775238037, | |
| "rewards/margins": 1.224104881286621, | |
| "rewards/rejected": -3.9569101333618164, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.32914774245257816, | |
| "grad_norm": 12.327388201247391, | |
| "learning_rate": 5.937574156761684e-07, | |
| "logits/chosen": -0.9693203568458557, | |
| "logits/rejected": -0.9414520263671875, | |
| "logps/chosen": -467.2398681640625, | |
| "logps/rejected": -549.6279296875, | |
| "loss": 0.558, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.324397087097168, | |
| "rewards/margins": 0.6942296624183655, | |
| "rewards/rejected": -3.018627166748047, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.33128506545551695, | |
| "grad_norm": 11.452079247996043, | |
| "learning_rate": 5.918719271454026e-07, | |
| "logits/chosen": -0.9070168137550354, | |
| "logits/rejected": -0.9717684984207153, | |
| "logps/chosen": -463.07073974609375, | |
| "logps/rejected": -531.9006958007812, | |
| "loss": 0.4757, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.246826171875, | |
| "rewards/margins": 0.8463835120201111, | |
| "rewards/rejected": -3.0932095050811768, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.3334223884584558, | |
| "grad_norm": 10.387010221236869, | |
| "learning_rate": 5.89972905909952e-07, | |
| "logits/chosen": -0.9737594127655029, | |
| "logits/rejected": -0.9816542267799377, | |
| "logps/chosen": -482.228515625, | |
| "logps/rejected": -550.7467041015625, | |
| "loss": 0.4944, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.8164095878601074, | |
| "rewards/margins": 0.6258185505867004, | |
| "rewards/rejected": -3.442227840423584, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.3355597114613946, | |
| "grad_norm": 13.81426757522365, | |
| "learning_rate": 5.880604582198217e-07, | |
| "logits/chosen": -0.8456093072891235, | |
| "logits/rejected": -0.831813395023346, | |
| "logps/chosen": -598.4921875, | |
| "logps/rejected": -722.5546875, | |
| "loss": 0.4857, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.0666327476501465, | |
| "rewards/margins": 1.3944388628005981, | |
| "rewards/rejected": -4.461071014404297, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.3376970344643334, | |
| "grad_norm": 12.739583339970373, | |
| "learning_rate": 5.86134691076226e-07, | |
| "logits/chosen": -0.9977485537528992, | |
| "logits/rejected": -0.9227933287620544, | |
| "logps/chosen": -458.33758544921875, | |
| "logps/rejected": -590.2880249023438, | |
| "loss": 0.4951, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.321625232696533, | |
| "rewards/margins": 1.1777639389038086, | |
| "rewards/rejected": -3.499389410018921, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.33983435746727225, | |
| "grad_norm": 11.40592684455961, | |
| "learning_rate": 5.841957122256004e-07, | |
| "logits/chosen": -0.9870405793190002, | |
| "logits/rejected": -0.9605780839920044, | |
| "logps/chosen": -460.610107421875, | |
| "logps/rejected": -543.5918579101562, | |
| "loss": 0.4798, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.591230869293213, | |
| "rewards/margins": 0.8468169569969177, | |
| "rewards/rejected": -3.4380481243133545, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.34197168047021104, | |
| "grad_norm": 11.54298974939068, | |
| "learning_rate": 5.822436301535739e-07, | |
| "logits/chosen": -0.9439342617988586, | |
| "logits/rejected": -0.896612823009491, | |
| "logps/chosen": -549.1138916015625, | |
| "logps/rejected": -611.7896728515625, | |
| "loss": 0.4867, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -2.6459498405456543, | |
| "rewards/margins": 0.6619236469268799, | |
| "rewards/rejected": -3.307873487472534, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.3441090034731499, | |
| "grad_norm": 10.028108962774086, | |
| "learning_rate": 5.802785540788994e-07, | |
| "logits/chosen": -1.018284797668457, | |
| "logits/rejected": -0.9662111401557922, | |
| "logps/chosen": -477.1954650878906, | |
| "logps/rejected": -642.691650390625, | |
| "loss": 0.4616, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -2.5217745304107666, | |
| "rewards/margins": 1.391313910484314, | |
| "rewards/rejected": -3.913088321685791, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.3462463264760887, | |
| "grad_norm": 10.240941578340253, | |
| "learning_rate": 5.783005939473424e-07, | |
| "logits/chosen": -0.8555217981338501, | |
| "logits/rejected": -0.8488653898239136, | |
| "logps/chosen": -490.9669189453125, | |
| "logps/rejected": -589.8624267578125, | |
| "loss": 0.474, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.722303628921509, | |
| "rewards/margins": 0.9005026817321777, | |
| "rewards/rejected": -3.6228065490722656, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.3483836494790275, | |
| "grad_norm": 11.409205514657348, | |
| "learning_rate": 5.763098604255298e-07, | |
| "logits/chosen": -0.8463901281356812, | |
| "logits/rejected": -0.8225279450416565, | |
| "logps/chosen": -470.57440185546875, | |
| "logps/rejected": -639.9716796875, | |
| "loss": 0.4536, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -2.5086305141448975, | |
| "rewards/margins": 1.750978946685791, | |
| "rewards/rejected": -4.259609222412109, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.35052097248196634, | |
| "grad_norm": 10.927823097709238, | |
| "learning_rate": 5.743064648947584e-07, | |
| "logits/chosen": -0.7492021322250366, | |
| "logits/rejected": -0.7506992816925049, | |
| "logps/chosen": -561.3836669921875, | |
| "logps/rejected": -710.4789428710938, | |
| "loss": 0.5088, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -2.9847238063812256, | |
| "rewards/margins": 1.428779125213623, | |
| "rewards/rejected": -4.413503170013428, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.3526582954849052, | |
| "grad_norm": 13.549577455647025, | |
| "learning_rate": 5.722905194447625e-07, | |
| "logits/chosen": -1.0247427225112915, | |
| "logits/rejected": -1.0630146265029907, | |
| "logps/chosen": -696.59228515625, | |
| "logps/rejected": -761.5941772460938, | |
| "loss": 0.5028, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.713867425918579, | |
| "rewards/margins": 0.7951303124427795, | |
| "rewards/rejected": -4.508997440338135, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.35479561848784397, | |
| "grad_norm": 15.326243524128424, | |
| "learning_rate": 5.702621368674431e-07, | |
| "logits/chosen": -0.9843580722808838, | |
| "logits/rejected": -0.9505161643028259, | |
| "logps/chosen": -439.31622314453125, | |
| "logps/rejected": -539.7349243164062, | |
| "loss": 0.5054, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.5671231746673584, | |
| "rewards/margins": 0.8157074451446533, | |
| "rewards/rejected": -3.3828306198120117, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.3569329414907828, | |
| "grad_norm": 13.81282026576703, | |
| "learning_rate": 5.682214306505567e-07, | |
| "logits/chosen": -0.7752354145050049, | |
| "logits/rejected": -0.831868052482605, | |
| "logps/chosen": -618.5186157226562, | |
| "logps/rejected": -734.19580078125, | |
| "loss": 0.4584, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.667295217514038, | |
| "rewards/margins": 1.2412769794464111, | |
| "rewards/rejected": -4.908572196960449, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.3590702644937216, | |
| "grad_norm": 12.331430233086655, | |
| "learning_rate": 5.661685149713663e-07, | |
| "logits/chosen": -0.8932394981384277, | |
| "logits/rejected": -0.8569494485855103, | |
| "logps/chosen": -504.1289367675781, | |
| "logps/rejected": -628.0838012695312, | |
| "loss": 0.5174, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.809536933898926, | |
| "rewards/margins": 1.0781981945037842, | |
| "rewards/rejected": -3.88773512840271, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.36120758749666043, | |
| "grad_norm": 13.419219251759696, | |
| "learning_rate": 5.641035046902522e-07, | |
| "logits/chosen": -0.8772855997085571, | |
| "logits/rejected": -0.8467257618904114, | |
| "logps/chosen": -470.7606201171875, | |
| "logps/rejected": -565.2396240234375, | |
| "loss": 0.4684, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.8519928455352783, | |
| "rewards/margins": 0.8433955907821655, | |
| "rewards/rejected": -3.695388078689575, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.36334491049959927, | |
| "grad_norm": 12.695684401858434, | |
| "learning_rate": 5.620265153442869e-07, | |
| "logits/chosen": -1.0130642652511597, | |
| "logits/rejected": -0.9847390055656433, | |
| "logps/chosen": -514.5957641601562, | |
| "logps/rejected": -585.168212890625, | |
| "loss": 0.469, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.0407803058624268, | |
| "rewards/margins": 0.6757410168647766, | |
| "rewards/rejected": -3.7165212631225586, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.36548223350253806, | |
| "grad_norm": 12.147780619595629, | |
| "learning_rate": 5.599376631407699e-07, | |
| "logits/chosen": -0.9069080948829651, | |
| "logits/rejected": -0.8009949922561646, | |
| "logps/chosen": -547.1576538085938, | |
| "logps/rejected": -679.2531127929688, | |
| "loss": 0.4299, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.197631359100342, | |
| "rewards/margins": 1.0914721488952637, | |
| "rewards/rejected": -4.2891035079956055, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.3676195565054769, | |
| "grad_norm": 18.51274657202258, | |
| "learning_rate": 5.578370649507255e-07, | |
| "logits/chosen": -0.6547280550003052, | |
| "logits/rejected": -0.6987937092781067, | |
| "logps/chosen": -577.7064819335938, | |
| "logps/rejected": -660.9352416992188, | |
| "loss": 0.5682, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.2265751361846924, | |
| "rewards/margins": 0.9371775388717651, | |
| "rewards/rejected": -4.163752555847168, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.36975687950841574, | |
| "grad_norm": 13.62299944344872, | |
| "learning_rate": 5.557248383023656e-07, | |
| "logits/chosen": -0.9801725745201111, | |
| "logits/rejected": -0.945483386516571, | |
| "logps/chosen": -449.10614013671875, | |
| "logps/rejected": -567.4400634765625, | |
| "loss": 0.4765, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -2.63840389251709, | |
| "rewards/margins": 1.1649121046066284, | |
| "rewards/rejected": -3.803316116333008, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.3718942025113545, | |
| "grad_norm": 12.16862925856691, | |
| "learning_rate": 5.536011013745121e-07, | |
| "logits/chosen": -0.9143638610839844, | |
| "logits/rejected": -0.9123052358627319, | |
| "logps/chosen": -483.41357421875, | |
| "logps/rejected": -624.1827392578125, | |
| "loss": 0.459, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -2.4780359268188477, | |
| "rewards/margins": 1.3084805011749268, | |
| "rewards/rejected": -3.7865164279937744, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.37403152551429336, | |
| "grad_norm": 11.507324455166394, | |
| "learning_rate": 5.514659729899857e-07, | |
| "logits/chosen": -0.7952810525894165, | |
| "logits/rejected": -0.7528566718101501, | |
| "logps/chosen": -498.6630554199219, | |
| "logps/rejected": -630.5216064453125, | |
| "loss": 0.4045, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -2.8591041564941406, | |
| "rewards/margins": 1.1144087314605713, | |
| "rewards/rejected": -3.973512649536133, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.37616884851723215, | |
| "grad_norm": 12.170272129622308, | |
| "learning_rate": 5.493195726089582e-07, | |
| "logits/chosen": -0.9865118265151978, | |
| "logits/rejected": -0.9716078639030457, | |
| "logps/chosen": -543.14501953125, | |
| "logps/rejected": -683.6224365234375, | |
| "loss": 0.4348, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.054544448852539, | |
| "rewards/margins": 1.5414537191390991, | |
| "rewards/rejected": -4.595998287200928, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.378306171520171, | |
| "grad_norm": 13.22726269166873, | |
| "learning_rate": 5.471620203222677e-07, | |
| "logits/chosen": -0.9932882189750671, | |
| "logits/rejected": -0.991521418094635, | |
| "logps/chosen": -507.20574951171875, | |
| "logps/rejected": -601.6077880859375, | |
| "loss": 0.4227, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.048675060272217, | |
| "rewards/margins": 0.8585546016693115, | |
| "rewards/rejected": -3.9072299003601074, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.3804434945231098, | |
| "grad_norm": 11.607213817333788, | |
| "learning_rate": 5.449934368447002e-07, | |
| "logits/chosen": -0.8857173919677734, | |
| "logits/rejected": -0.9733960628509521, | |
| "logps/chosen": -627.8464965820312, | |
| "logps/rejected": -713.438720703125, | |
| "loss": 0.4544, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.520371437072754, | |
| "rewards/margins": 1.112282156944275, | |
| "rewards/rejected": -4.63265323638916, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.3825808175260486, | |
| "grad_norm": 12.188016147935972, | |
| "learning_rate": 5.428139435082358e-07, | |
| "logits/chosen": -0.7584793567657471, | |
| "logits/rejected": -0.8010637760162354, | |
| "logps/chosen": -566.6948852539062, | |
| "logps/rejected": -713.6017456054688, | |
| "loss": 0.4106, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.4136064052581787, | |
| "rewards/margins": 1.416290044784546, | |
| "rewards/rejected": -4.829895973205566, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.38471814052898745, | |
| "grad_norm": 13.803380819286584, | |
| "learning_rate": 5.406236622552594e-07, | |
| "logits/chosen": -0.8117521405220032, | |
| "logits/rejected": -0.8442290425300598, | |
| "logps/chosen": -658.14990234375, | |
| "logps/rejected": -791.3058471679688, | |
| "loss": 0.4795, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.9962399005889893, | |
| "rewards/margins": 1.3741449117660522, | |
| "rewards/rejected": -5.370384693145752, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.38685546353192624, | |
| "grad_norm": 19.74453585010364, | |
| "learning_rate": 5.384227156317388e-07, | |
| "logits/chosen": -0.9625721573829651, | |
| "logits/rejected": -1.0246632099151611, | |
| "logps/chosen": -570.2412109375, | |
| "logps/rejected": -627.0347900390625, | |
| "loss": 0.5302, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -3.3639683723449707, | |
| "rewards/margins": 0.4717191457748413, | |
| "rewards/rejected": -3.8356876373291016, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.3889927865348651, | |
| "grad_norm": 12.180825522167304, | |
| "learning_rate": 5.362112267803678e-07, | |
| "logits/chosen": -1.1118032932281494, | |
| "logits/rejected": -1.0548309087753296, | |
| "logps/chosen": -561.20654296875, | |
| "logps/rejected": -801.3348388671875, | |
| "loss": 0.407, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -3.0669779777526855, | |
| "rewards/margins": 2.297924041748047, | |
| "rewards/rejected": -5.364901542663574, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.3911301095378039, | |
| "grad_norm": 15.493196005683158, | |
| "learning_rate": 5.339893194336765e-07, | |
| "logits/chosen": -0.9748885035514832, | |
| "logits/rejected": -0.9212169647216797, | |
| "logps/chosen": -556.37451171875, | |
| "logps/rejected": -747.4850463867188, | |
| "loss": 0.434, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.1176629066467285, | |
| "rewards/margins": 2.0257248878479004, | |
| "rewards/rejected": -5.143387794494629, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.3932674325407427, | |
| "grad_norm": 12.834279479584271, | |
| "learning_rate": 5.317571179071092e-07, | |
| "logits/chosen": -1.0676939487457275, | |
| "logits/rejected": -1.0940359830856323, | |
| "logps/chosen": -547.4573364257812, | |
| "logps/rejected": -625.5507202148438, | |
| "loss": 0.4441, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.1368393898010254, | |
| "rewards/margins": 0.7552158236503601, | |
| "rewards/rejected": -3.8920555114746094, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.39540475554368154, | |
| "grad_norm": 15.858497539317051, | |
| "learning_rate": 5.295147470920672e-07, | |
| "logits/chosen": -0.8825098276138306, | |
| "logits/rejected": -0.8697827458381653, | |
| "logps/chosen": -606.1354370117188, | |
| "logps/rejected": -752.2478637695312, | |
| "loss": 0.4948, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.499366521835327, | |
| "rewards/margins": 1.4578311443328857, | |
| "rewards/rejected": -4.957198143005371, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.3975420785466204, | |
| "grad_norm": 15.81390161765314, | |
| "learning_rate": 5.27262332448923e-07, | |
| "logits/chosen": -0.9162149429321289, | |
| "logits/rejected": -0.8807965517044067, | |
| "logps/chosen": -588.79736328125, | |
| "logps/rejected": -775.908447265625, | |
| "loss": 0.4836, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.620396852493286, | |
| "rewards/margins": 1.820716142654419, | |
| "rewards/rejected": -5.441113471984863, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.39967940154955917, | |
| "grad_norm": 11.930553509668082, | |
| "learning_rate": 5.25e-07, | |
| "logits/chosen": -0.878287136554718, | |
| "logits/rejected": -0.8447045683860779, | |
| "logps/chosen": -588.80908203125, | |
| "logps/rejected": -710.580078125, | |
| "loss": 0.4217, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.770730495452881, | |
| "rewards/margins": 1.0582211017608643, | |
| "rewards/rejected": -4.828950881958008, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.401816724552498, | |
| "grad_norm": 19.78691767583468, | |
| "learning_rate": 5.227278763225213e-07, | |
| "logits/chosen": -1.002894639968872, | |
| "logits/rejected": -0.9999670386314392, | |
| "logps/chosen": -729.7056884765625, | |
| "logps/rejected": -895.4324340820312, | |
| "loss": 0.488, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.301171779632568, | |
| "rewards/margins": 1.7632997035980225, | |
| "rewards/rejected": -6.06447172164917, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.4039540475554368, | |
| "grad_norm": 15.17666246755368, | |
| "learning_rate": 5.204460885415285e-07, | |
| "logits/chosen": -0.9192298650741577, | |
| "logits/rejected": -0.9201673269271851, | |
| "logps/chosen": -612.3890380859375, | |
| "logps/rejected": -722.1416625976562, | |
| "loss": 0.4896, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.5543270111083984, | |
| "rewards/margins": 1.225655198097229, | |
| "rewards/rejected": -4.779982089996338, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.40609137055837563, | |
| "grad_norm": 12.423293515766808, | |
| "learning_rate": 5.181547643227682e-07, | |
| "logits/chosen": -0.8379946947097778, | |
| "logits/rejected": -0.8106329441070557, | |
| "logps/chosen": -546.194580078125, | |
| "logps/rejected": -636.456298828125, | |
| "loss": 0.4722, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -3.497601270675659, | |
| "rewards/margins": 0.9037814736366272, | |
| "rewards/rejected": -4.4013824462890625, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.40822869356131447, | |
| "grad_norm": 16.79590350569075, | |
| "learning_rate": 5.158540318655495e-07, | |
| "logits/chosen": -1.0028809309005737, | |
| "logits/rejected": -0.9536342024803162, | |
| "logps/chosen": -669.4102783203125, | |
| "logps/rejected": -854.4193115234375, | |
| "loss": 0.4629, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.044215679168701, | |
| "rewards/margins": 1.6450707912445068, | |
| "rewards/rejected": -5.689286708831787, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.41036601656425326, | |
| "grad_norm": 14.164634270897476, | |
| "learning_rate": 5.135440198955717e-07, | |
| "logits/chosen": -0.97246915102005, | |
| "logits/rejected": -0.9198447465896606, | |
| "logps/chosen": -513.80859375, | |
| "logps/rejected": -543.1536865234375, | |
| "loss": 0.4983, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -3.1685631275177, | |
| "rewards/margins": 0.3603837192058563, | |
| "rewards/rejected": -3.528946876525879, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.4125033395671921, | |
| "grad_norm": 15.329408000607371, | |
| "learning_rate": 5.112248576577213e-07, | |
| "logits/chosen": -0.7749034762382507, | |
| "logits/rejected": -0.8009266257286072, | |
| "logps/chosen": -596.5591430664062, | |
| "logps/rejected": -797.4959716796875, | |
| "loss": 0.4978, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.661137342453003, | |
| "rewards/margins": 1.7927237749099731, | |
| "rewards/rejected": -5.453861236572266, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.41464066257013094, | |
| "grad_norm": 14.042005201292891, | |
| "learning_rate": 5.088966749088414e-07, | |
| "logits/chosen": -1.116222858428955, | |
| "logits/rejected": -1.090318202972412, | |
| "logps/chosen": -603.7730102539062, | |
| "logps/rejected": -657.4857177734375, | |
| "loss": 0.4895, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -3.191464424133301, | |
| "rewards/margins": 0.7639944553375244, | |
| "rewards/rejected": -3.9554591178894043, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.4167779855730697, | |
| "grad_norm": 13.675394100746447, | |
| "learning_rate": 5.065596019104713e-07, | |
| "logits/chosen": -1.0994431972503662, | |
| "logits/rejected": -1.091101884841919, | |
| "logps/chosen": -700.6111450195312, | |
| "logps/rejected": -847.7877197265625, | |
| "loss": 0.4107, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.003596305847168, | |
| "rewards/margins": 1.4772356748580933, | |
| "rewards/rejected": -5.480831623077393, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.41891530857600856, | |
| "grad_norm": 18.55076809192975, | |
| "learning_rate": 5.042137694215592e-07, | |
| "logits/chosen": -0.9493328928947449, | |
| "logits/rejected": -1.022355079650879, | |
| "logps/chosen": -641.4091796875, | |
| "logps/rejected": -696.9602661132812, | |
| "loss": 0.618, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.780533790588379, | |
| "rewards/margins": 0.6779485940933228, | |
| "rewards/rejected": -4.458482265472412, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.42105263157894735, | |
| "grad_norm": 12.163973418481676, | |
| "learning_rate": 5.018593086911453e-07, | |
| "logits/chosen": -1.0403311252593994, | |
| "logits/rejected": -1.069019079208374, | |
| "logps/chosen": -671.8824462890625, | |
| "logps/rejected": -849.7545166015625, | |
| "loss": 0.4197, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.7160799503326416, | |
| "rewards/margins": 1.7244200706481934, | |
| "rewards/rejected": -5.440499782562256, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.4231899545818862, | |
| "grad_norm": 13.736070000914967, | |
| "learning_rate": 4.994963514510188e-07, | |
| "logits/chosen": -1.0466636419296265, | |
| "logits/rejected": -0.9742124676704407, | |
| "logps/chosen": -536.3780517578125, | |
| "logps/rejected": -622.6585083007812, | |
| "loss": 0.4457, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.165748119354248, | |
| "rewards/margins": 0.7600272297859192, | |
| "rewards/rejected": -3.9257755279541016, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.425327277584825, | |
| "grad_norm": 11.871725745717265, | |
| "learning_rate": 4.971250299083479e-07, | |
| "logits/chosen": -0.9257810711860657, | |
| "logits/rejected": -0.9252554774284363, | |
| "logps/chosen": -556.8322143554688, | |
| "logps/rejected": -656.6981201171875, | |
| "loss": 0.4155, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.365008592605591, | |
| "rewards/margins": 1.075837254524231, | |
| "rewards/rejected": -4.440845966339111, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.4274646005877638, | |
| "grad_norm": 12.50996465065088, | |
| "learning_rate": 4.947454767382822e-07, | |
| "logits/chosen": -0.9889892339706421, | |
| "logits/rejected": -1.004003882408142, | |
| "logps/chosen": -499.806884765625, | |
| "logps/rejected": -617.0054321289062, | |
| "loss": 0.5021, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -2.9506893157958984, | |
| "rewards/margins": 1.0361154079437256, | |
| "rewards/rejected": -3.986804485321045, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.4274646005877638, | |
| "eval_logits/chosen": -1.084425926208496, | |
| "eval_logits/rejected": -1.103520393371582, | |
| "eval_logps/chosen": -613.6640625, | |
| "eval_logps/rejected": -715.6145629882812, | |
| "eval_loss": 0.44827935099601746, | |
| "eval_rewards/accuracies": 0.8024193644523621, | |
| "eval_rewards/chosen": -3.4053404331207275, | |
| "eval_rewards/margins": 1.080467700958252, | |
| "eval_rewards/rejected": -4.4858078956604, | |
| "eval_runtime": 126.9937, | |
| "eval_samples_per_second": 15.442, | |
| "eval_steps_per_second": 0.488, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.42960192359070265, | |
| "grad_norm": 13.246746821586836, | |
| "learning_rate": 4.923578250765301e-07, | |
| "logits/chosen": -1.0361019372940063, | |
| "logits/rejected": -1.0069564580917358, | |
| "logps/chosen": -630.4815063476562, | |
| "logps/rejected": -819.7576293945312, | |
| "loss": 0.4381, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.616086483001709, | |
| "rewards/margins": 1.9921762943267822, | |
| "rewards/rejected": -5.608262538909912, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.4317392465936415, | |
| "grad_norm": 13.986462198608523, | |
| "learning_rate": 4.899622085119092e-07, | |
| "logits/chosen": -1.0016858577728271, | |
| "logits/rejected": -1.0090727806091309, | |
| "logps/chosen": -639.8945922851562, | |
| "logps/rejected": -797.4194946289062, | |
| "loss": 0.389, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.3241751194000244, | |
| "rewards/margins": 1.6805753707885742, | |
| "rewards/rejected": -5.0047502517700195, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.4338765695965803, | |
| "grad_norm": 12.938400299110988, | |
| "learning_rate": 4.875587610788733e-07, | |
| "logits/chosen": -1.0168877840042114, | |
| "logits/rejected": -1.0365827083587646, | |
| "logps/chosen": -593.1552124023438, | |
| "logps/rejected": -752.401611328125, | |
| "loss": 0.4125, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.574855327606201, | |
| "rewards/margins": 1.3029221296310425, | |
| "rewards/rejected": -4.877778053283691, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.4360138925995191, | |
| "grad_norm": 12.350696519100337, | |
| "learning_rate": 4.851476172500112e-07, | |
| "logits/chosen": -1.0002954006195068, | |
| "logits/rejected": -0.9592558145523071, | |
| "logps/chosen": -691.1764526367188, | |
| "logps/rejected": -894.4859619140625, | |
| "loss": 0.4075, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.358014106750488, | |
| "rewards/margins": 1.7022799253463745, | |
| "rewards/rejected": -6.060294151306152, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.4381512156024579, | |
| "grad_norm": 14.780877866432451, | |
| "learning_rate": 4.827289119285249e-07, | |
| "logits/chosen": -0.9795655608177185, | |
| "logits/rejected": -0.9753063917160034, | |
| "logps/chosen": -560.5239868164062, | |
| "logps/rejected": -701.3786010742188, | |
| "loss": 0.466, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.2604007720947266, | |
| "rewards/margins": 1.3381632566452026, | |
| "rewards/rejected": -4.5985636711120605, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.44028853860539674, | |
| "grad_norm": 22.9673286565427, | |
| "learning_rate": 4.803027804406808e-07, | |
| "logits/chosen": -0.7228925228118896, | |
| "logits/rejected": -0.8121004700660706, | |
| "logps/chosen": -680.5167236328125, | |
| "logps/rejected": -703.938232421875, | |
| "loss": 0.559, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.237366199493408, | |
| "rewards/margins": 0.45848214626312256, | |
| "rewards/rejected": -4.695847988128662, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.4424258616083356, | |
| "grad_norm": 18.405605350322162, | |
| "learning_rate": 4.778693585282384e-07, | |
| "logits/chosen": -1.0253394842147827, | |
| "logits/rejected": -1.0438570976257324, | |
| "logps/chosen": -631.9534912109375, | |
| "logps/rejected": -705.13623046875, | |
| "loss": 0.5002, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.731057643890381, | |
| "rewards/margins": 0.8837456107139587, | |
| "rewards/rejected": -4.614803314208984, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.44456318461127436, | |
| "grad_norm": 15.70410271681006, | |
| "learning_rate": 4.75428782340855e-07, | |
| "logits/chosen": -1.0454537868499756, | |
| "logits/rejected": -1.1345258951187134, | |
| "logps/chosen": -641.6697998046875, | |
| "logps/rejected": -731.5447387695312, | |
| "loss": 0.4413, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -3.9197850227355957, | |
| "rewards/margins": 0.858109712600708, | |
| "rewards/rejected": -4.777895450592041, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.4467005076142132, | |
| "grad_norm": 13.037823833248444, | |
| "learning_rate": 4.7298118842846993e-07, | |
| "logits/chosen": -0.9263153076171875, | |
| "logits/rejected": -0.9316372871398926, | |
| "logps/chosen": -522.490966796875, | |
| "logps/rejected": -635.503173828125, | |
| "loss": 0.3848, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.2573442459106445, | |
| "rewards/margins": 0.9754186272621155, | |
| "rewards/rejected": -4.232762336730957, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.448837830617152, | |
| "grad_norm": 15.711193208519104, | |
| "learning_rate": 4.705267137336622e-07, | |
| "logits/chosen": -1.0621023178100586, | |
| "logits/rejected": -1.0041028261184692, | |
| "logps/chosen": -799.73974609375, | |
| "logps/rejected": -927.7545776367188, | |
| "loss": 0.4351, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.812376022338867, | |
| "rewards/margins": 1.1584749221801758, | |
| "rewards/rejected": -5.970850944519043, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.45097515362009083, | |
| "grad_norm": 13.021464052671469, | |
| "learning_rate": 4.680654955839905e-07, | |
| "logits/chosen": -1.0352355241775513, | |
| "logits/rejected": -1.0864734649658203, | |
| "logps/chosen": -679.844970703125, | |
| "logps/rejected": -842.2390747070312, | |
| "loss": 0.3926, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.970327138900757, | |
| "rewards/margins": 1.7980788946151733, | |
| "rewards/rejected": -5.768405914306641, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.45311247662302967, | |
| "grad_norm": 13.400793392671966, | |
| "learning_rate": 4.655976716843085e-07, | |
| "logits/chosen": -0.9447561502456665, | |
| "logits/rejected": -0.9419984221458435, | |
| "logps/chosen": -577.1600341796875, | |
| "logps/rejected": -626.2113647460938, | |
| "loss": 0.4097, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -3.421278476715088, | |
| "rewards/margins": 0.7281789183616638, | |
| "rewards/rejected": -4.1494574546813965, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.45524979962596845, | |
| "grad_norm": 15.76259483264769, | |
| "learning_rate": 4.63123380109061e-07, | |
| "logits/chosen": -0.9420887231826782, | |
| "logits/rejected": -0.9258626699447632, | |
| "logps/chosen": -604.66064453125, | |
| "logps/rejected": -688.6736450195312, | |
| "loss": 0.5374, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -3.727839708328247, | |
| "rewards/margins": 0.6107276082038879, | |
| "rewards/rejected": -4.33856725692749, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.4573871226289073, | |
| "grad_norm": 14.47796429675634, | |
| "learning_rate": 4.6064275929455854e-07, | |
| "logits/chosen": -1.012191653251648, | |
| "logits/rejected": -0.9834191799163818, | |
| "logps/chosen": -539.5471801757812, | |
| "logps/rejected": -643.7244873046875, | |
| "loss": 0.3953, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.355208396911621, | |
| "rewards/margins": 0.9154967665672302, | |
| "rewards/rejected": -4.270705223083496, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.45952444563184613, | |
| "grad_norm": 18.127660409487774, | |
| "learning_rate": 4.581559480312316e-07, | |
| "logits/chosen": -0.9188220500946045, | |
| "logits/rejected": -0.8601971864700317, | |
| "logps/chosen": -562.9246826171875, | |
| "logps/rejected": -646.825439453125, | |
| "loss": 0.5093, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -3.4996066093444824, | |
| "rewards/margins": 0.7593184113502502, | |
| "rewards/rejected": -4.258925437927246, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.4616617686347849, | |
| "grad_norm": 14.332440562045605, | |
| "learning_rate": 4.556630854558657e-07, | |
| "logits/chosen": -1.0821930170059204, | |
| "logits/rejected": -1.0690282583236694, | |
| "logps/chosen": -612.6143188476562, | |
| "logps/rejected": -785.2466430664062, | |
| "loss": 0.4259, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -3.40315580368042, | |
| "rewards/margins": 1.751339316368103, | |
| "rewards/rejected": -5.1544952392578125, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.46379909163772376, | |
| "grad_norm": 13.6646330391651, | |
| "learning_rate": 4.5316431104381647e-07, | |
| "logits/chosen": -1.0587506294250488, | |
| "logits/rejected": -1.0372809171676636, | |
| "logps/chosen": -491.8074951171875, | |
| "logps/rejected": -638.18212890625, | |
| "loss": 0.4567, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.595409870147705, | |
| "rewards/margins": 1.3147836923599243, | |
| "rewards/rejected": -3.910193681716919, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.46593641464066254, | |
| "grad_norm": 16.27324594458174, | |
| "learning_rate": 4.50659764601206e-07, | |
| "logits/chosen": -1.1704292297363281, | |
| "logits/rejected": -1.1407532691955566, | |
| "logps/chosen": -635.9962158203125, | |
| "logps/rejected": -709.2085571289062, | |
| "loss": 0.5169, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -3.4440693855285645, | |
| "rewards/margins": 0.8417525887489319, | |
| "rewards/rejected": -4.285822868347168, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.4680737376436014, | |
| "grad_norm": 14.30405716244965, | |
| "learning_rate": 4.4814958625710095e-07, | |
| "logits/chosen": -0.9373888969421387, | |
| "logits/rejected": -0.9414705634117126, | |
| "logps/chosen": -641.3363037109375, | |
| "logps/rejected": -782.9466552734375, | |
| "loss": 0.4214, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.41910719871521, | |
| "rewards/margins": 1.5010746717453003, | |
| "rewards/rejected": -4.920182228088379, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.4702110606465402, | |
| "grad_norm": 12.888240768536523, | |
| "learning_rate": 4.4563391645567203e-07, | |
| "logits/chosen": -0.9721742272377014, | |
| "logits/rejected": -0.9880850315093994, | |
| "logps/chosen": -563.5908813476562, | |
| "logps/rejected": -648.4769897460938, | |
| "loss": 0.4525, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.4122231006622314, | |
| "rewards/margins": 0.8691967725753784, | |
| "rewards/rejected": -4.28141975402832, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.472348383649479, | |
| "grad_norm": 13.808905917996086, | |
| "learning_rate": 4.4311289594833626e-07, | |
| "logits/chosen": -1.062425136566162, | |
| "logits/rejected": -1.0666816234588623, | |
| "logps/chosen": -425.4259338378906, | |
| "logps/rejected": -508.6617736816406, | |
| "loss": 0.3815, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -2.534140110015869, | |
| "rewards/margins": 0.9357590079307556, | |
| "rewards/rejected": -3.4698996543884277, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.47448570665241785, | |
| "grad_norm": 14.257159662099976, | |
| "learning_rate": 4.405866657858822e-07, | |
| "logits/chosen": -1.0133253335952759, | |
| "logits/rejected": -1.0463453531265259, | |
| "logps/chosen": -577.285888671875, | |
| "logps/rejected": -644.2965087890625, | |
| "loss": 0.4452, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -3.20715594291687, | |
| "rewards/margins": 0.7728344202041626, | |
| "rewards/rejected": -3.979990243911743, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.4766230296553567, | |
| "grad_norm": 12.854510596268446, | |
| "learning_rate": 4.38055367310578e-07, | |
| "logits/chosen": -0.985044538974762, | |
| "logits/rejected": -0.9488839507102966, | |
| "logps/chosen": -645.5025024414062, | |
| "logps/rejected": -869.3623046875, | |
| "loss": 0.3977, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.6057004928588867, | |
| "rewards/margins": 1.8732173442840576, | |
| "rewards/rejected": -5.478918075561523, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.4787603526582955, | |
| "grad_norm": 14.869442818796234, | |
| "learning_rate": 4.355191421482629e-07, | |
| "logits/chosen": -1.1076860427856445, | |
| "logits/rejected": -1.0764734745025635, | |
| "logps/chosen": -516.0100708007812, | |
| "logps/rejected": -644.3858642578125, | |
| "loss": 0.4573, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.1032497882843018, | |
| "rewards/margins": 1.1473891735076904, | |
| "rewards/rejected": -4.250638961791992, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.4808976756612343, | |
| "grad_norm": 16.43448798480314, | |
| "learning_rate": 4.329781322004241e-07, | |
| "logits/chosen": -1.0606333017349243, | |
| "logits/rejected": -1.0501140356063843, | |
| "logps/chosen": -663.1077880859375, | |
| "logps/rejected": -874.1226806640625, | |
| "loss": 0.4973, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.6670713424682617, | |
| "rewards/margins": 1.8036597967147827, | |
| "rewards/rejected": -5.470730781555176, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.4830349986641731, | |
| "grad_norm": 15.477278289904351, | |
| "learning_rate": 4.304324796362568e-07, | |
| "logits/chosen": -1.026381015777588, | |
| "logits/rejected": -0.9911444187164307, | |
| "logps/chosen": -616.499267578125, | |
| "logps/rejected": -757.9083862304688, | |
| "loss": 0.4483, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.533419609069824, | |
| "rewards/margins": 1.2801600694656372, | |
| "rewards/rejected": -4.81358003616333, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.48517232166711194, | |
| "grad_norm": 14.181376472254154, | |
| "learning_rate": 4.2788232688471e-07, | |
| "logits/chosen": -1.0257827043533325, | |
| "logits/rejected": -1.030511498451233, | |
| "logps/chosen": -623.668212890625, | |
| "logps/rejected": -722.494873046875, | |
| "loss": 0.4315, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.7645962238311768, | |
| "rewards/margins": 1.0766507387161255, | |
| "rewards/rejected": -4.841247081756592, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.4873096446700508, | |
| "grad_norm": 18.128852851106277, | |
| "learning_rate": 4.253278166265177e-07, | |
| "logits/chosen": -1.0358821153640747, | |
| "logits/rejected": -1.0768516063690186, | |
| "logps/chosen": -567.903564453125, | |
| "logps/rejected": -689.8753051757812, | |
| "loss": 0.4725, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -3.21205997467041, | |
| "rewards/margins": 1.3463387489318848, | |
| "rewards/rejected": -4.558398723602295, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.48944696767298956, | |
| "grad_norm": 13.499906597454904, | |
| "learning_rate": 4.2276909178621567e-07, | |
| "logits/chosen": -0.9607468843460083, | |
| "logits/rejected": -0.9298718571662903, | |
| "logps/chosen": -749.897705078125, | |
| "logps/rejected": -971.5405883789062, | |
| "loss": 0.3638, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.647284030914307, | |
| "rewards/margins": 2.187747001647949, | |
| "rewards/rejected": -6.835031032562256, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.4915842906759284, | |
| "grad_norm": 15.151757518766736, | |
| "learning_rate": 4.2020629552414545e-07, | |
| "logits/chosen": -1.0237997770309448, | |
| "logits/rejected": -0.9573734998703003, | |
| "logps/chosen": -746.3142700195312, | |
| "logps/rejected": -845.5519409179688, | |
| "loss": 0.4511, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.615874767303467, | |
| "rewards/margins": 1.053206205368042, | |
| "rewards/rejected": -5.669081211090088, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.49372161367886724, | |
| "grad_norm": 15.54384712409316, | |
| "learning_rate": 4.176395712284436e-07, | |
| "logits/chosen": -1.0997838973999023, | |
| "logits/rejected": -1.0569286346435547, | |
| "logps/chosen": -650.4686279296875, | |
| "logps/rejected": -801.4287109375, | |
| "loss": 0.4244, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.111543655395508, | |
| "rewards/margins": 1.6189815998077393, | |
| "rewards/rejected": -5.730525970458984, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.49585893668180603, | |
| "grad_norm": 20.212574699658393, | |
| "learning_rate": 4.1506906250702017e-07, | |
| "logits/chosen": -0.9895690083503723, | |
| "logits/rejected": -0.9893757700920105, | |
| "logps/chosen": -653.65673828125, | |
| "logps/rejected": -774.2328491210938, | |
| "loss": 0.4351, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -4.012841701507568, | |
| "rewards/margins": 1.2045834064483643, | |
| "rewards/rejected": -5.217425346374512, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.49799625968474487, | |
| "grad_norm": 19.25316743472589, | |
| "learning_rate": 4.124949131795228e-07, | |
| "logits/chosen": -0.9950495362281799, | |
| "logits/rejected": -0.9298911690711975, | |
| "logps/chosen": -593.9339599609375, | |
| "logps/rejected": -793.3387451171875, | |
| "loss": 0.3809, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.7703654766082764, | |
| "rewards/margins": 1.754205346107483, | |
| "rewards/rejected": -5.524570941925049, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.5001335826876837, | |
| "grad_norm": 18.738876938950654, | |
| "learning_rate": 4.0991726726929135e-07, | |
| "logits/chosen": -1.06260085105896, | |
| "logits/rejected": -1.1126683950424194, | |
| "logps/chosen": -576.23583984375, | |
| "logps/rejected": -658.715576171875, | |
| "loss": 0.5016, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -3.6776123046875, | |
| "rewards/margins": 0.8450465202331543, | |
| "rewards/rejected": -4.5226593017578125, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.5022709056906225, | |
| "grad_norm": 15.19487228526955, | |
| "learning_rate": 4.073362689952982e-07, | |
| "logits/chosen": -0.895923912525177, | |
| "logits/rejected": -0.9245651364326477, | |
| "logps/chosen": -693.7532348632812, | |
| "logps/rejected": -714.86865234375, | |
| "loss": 0.3918, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -4.3402838706970215, | |
| "rewards/margins": 0.3496522903442383, | |
| "rewards/rejected": -4.68993616104126, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.5044082286935613, | |
| "grad_norm": 23.63383518264876, | |
| "learning_rate": 4.047520627640808e-07, | |
| "logits/chosen": -1.0637860298156738, | |
| "logits/rejected": -1.0301114320755005, | |
| "logps/chosen": -648.5076293945312, | |
| "logps/rejected": -772.224609375, | |
| "loss": 0.4001, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.6958110332489014, | |
| "rewards/margins": 1.3828696012496948, | |
| "rewards/rejected": -5.078680515289307, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.5065455516965002, | |
| "grad_norm": 26.41816558524286, | |
| "learning_rate": 4.02164793161661e-07, | |
| "logits/chosen": -0.9920711517333984, | |
| "logits/rejected": -0.9205961227416992, | |
| "logps/chosen": -606.1458129882812, | |
| "logps/rejected": -882.5693969726562, | |
| "loss": 0.4818, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -3.58052396774292, | |
| "rewards/margins": 0.41025519371032715, | |
| "rewards/rejected": -3.9907796382904053, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.508682874699439, | |
| "grad_norm": 15.74897212135201, | |
| "learning_rate": 3.9957460494545616e-07, | |
| "logits/chosen": -0.9549577236175537, | |
| "logits/rejected": -0.9211697578430176, | |
| "logps/chosen": -669.394287109375, | |
| "logps/rejected": -709.4468994140625, | |
| "loss": 0.4783, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -4.0179338455200195, | |
| "rewards/margins": 0.4471871554851532, | |
| "rewards/rejected": -4.465120792388916, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.5108201977023777, | |
| "grad_norm": 16.016701961339525, | |
| "learning_rate": 3.969816430361794e-07, | |
| "logits/chosen": -1.0792049169540405, | |
| "logits/rejected": -0.9834136366844177, | |
| "logps/chosen": -673.969970703125, | |
| "logps/rejected": -964.8361206054688, | |
| "loss": 0.3911, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -4.139585018157959, | |
| "rewards/margins": 2.691437244415283, | |
| "rewards/rejected": -6.831022262573242, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.5129575207053166, | |
| "grad_norm": 17.928342726480253, | |
| "learning_rate": 3.9438605250973193e-07, | |
| "logits/chosen": -0.9689183831214905, | |
| "logits/rejected": -0.9091028571128845, | |
| "logps/chosen": -529.1758422851562, | |
| "logps/rejected": -694.4168090820312, | |
| "loss": 0.4055, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.4839565753936768, | |
| "rewards/margins": 1.5238869190216064, | |
| "rewards/rejected": -5.007843494415283, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.5150948437082554, | |
| "grad_norm": 13.751574315942849, | |
| "learning_rate": 3.9178797858908544e-07, | |
| "logits/chosen": -0.9920436143875122, | |
| "logits/rejected": -0.9840397834777832, | |
| "logps/chosen": -706.0386352539062, | |
| "logps/rejected": -849.04345703125, | |
| "loss": 0.3468, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.169567108154297, | |
| "rewards/margins": 1.2529165744781494, | |
| "rewards/rejected": -5.422484397888184, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.5172321667111942, | |
| "grad_norm": 13.92465051011581, | |
| "learning_rate": 3.891875666361577e-07, | |
| "logits/chosen": -0.9852272868156433, | |
| "logits/rejected": -0.9285619258880615, | |
| "logps/chosen": -515.2431030273438, | |
| "logps/rejected": -713.544189453125, | |
| "loss": 0.3992, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.9760236740112305, | |
| "rewards/margins": 1.73358952999115, | |
| "rewards/rejected": -4.709612846374512, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.5193694897141331, | |
| "grad_norm": 15.781523786771364, | |
| "learning_rate": 3.865849621436787e-07, | |
| "logits/chosen": -1.0415364503860474, | |
| "logits/rejected": -1.0579416751861572, | |
| "logps/chosen": -614.417236328125, | |
| "logps/rejected": -755.8591918945312, | |
| "loss": 0.4447, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.9967360496520996, | |
| "rewards/margins": 1.3200428485870361, | |
| "rewards/rejected": -5.316779136657715, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.5215068127170719, | |
| "grad_norm": 13.998215826829304, | |
| "learning_rate": 3.839803107270509e-07, | |
| "logits/chosen": -0.8844782710075378, | |
| "logits/rejected": -0.7867937088012695, | |
| "logps/chosen": -551.641845703125, | |
| "logps/rejected": -744.5933227539062, | |
| "loss": 0.4074, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -3.376842737197876, | |
| "rewards/margins": 1.7052807807922363, | |
| "rewards/rejected": -5.082123756408691, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.5236441357200107, | |
| "grad_norm": 14.905532531142882, | |
| "learning_rate": 3.813737581162017e-07, | |
| "logits/chosen": -1.0912367105484009, | |
| "logits/rejected": -1.0676392316818237, | |
| "logps/chosen": -624.3543701171875, | |
| "logps/rejected": -729.902099609375, | |
| "loss": 0.413, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.5566658973693848, | |
| "rewards/margins": 1.0893948078155518, | |
| "rewards/rejected": -4.646060943603516, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.5257814587229495, | |
| "grad_norm": 13.609157264860004, | |
| "learning_rate": 3.7876545014743065e-07, | |
| "logits/chosen": -0.898338258266449, | |
| "logits/rejected": -0.9181092381477356, | |
| "logps/chosen": -624.335205078125, | |
| "logps/rejected": -792.426025390625, | |
| "loss": 0.4068, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.840670108795166, | |
| "rewards/margins": 1.4304872751235962, | |
| "rewards/rejected": -5.2711567878723145, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.5279187817258884, | |
| "grad_norm": 15.394561748839626, | |
| "learning_rate": 3.761555327552485e-07, | |
| "logits/chosen": -1.0056723356246948, | |
| "logits/rejected": -1.0341317653656006, | |
| "logps/chosen": -591.2755737304688, | |
| "logps/rejected": -716.7362060546875, | |
| "loss": 0.3879, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.308074951171875, | |
| "rewards/margins": 1.277435302734375, | |
| "rewards/rejected": -4.585509777069092, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.5300561047288271, | |
| "grad_norm": 16.757770907784614, | |
| "learning_rate": 3.735441519642139e-07, | |
| "logits/chosen": -1.1397889852523804, | |
| "logits/rejected": -1.1101233959197998, | |
| "logps/chosen": -517.30810546875, | |
| "logps/rejected": -605.835205078125, | |
| "loss": 0.4748, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.128363847732544, | |
| "rewards/margins": 0.9252746105194092, | |
| "rewards/rejected": -4.053638935089111, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.5321934277317659, | |
| "grad_norm": 14.363837462668146, | |
| "learning_rate": 3.70931453880762e-07, | |
| "logits/chosen": -0.9677891135215759, | |
| "logits/rejected": -0.9882450699806213, | |
| "logps/chosen": -697.7656860351562, | |
| "logps/rejected": -846.583740234375, | |
| "loss": 0.4142, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.01542329788208, | |
| "rewards/margins": 1.4604833126068115, | |
| "rewards/rejected": -5.4759063720703125, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.5343307507347048, | |
| "grad_norm": 18.905238772507477, | |
| "learning_rate": 3.6831758468503035e-07, | |
| "logits/chosen": -1.1917240619659424, | |
| "logits/rejected": -1.1658238172531128, | |
| "logps/chosen": -629.4384155273438, | |
| "logps/rejected": -744.068603515625, | |
| "loss": 0.5198, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.910032272338867, | |
| "rewards/margins": 1.0619316101074219, | |
| "rewards/rejected": -4.971964359283447, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.5364680737376436, | |
| "grad_norm": 13.775541697952745, | |
| "learning_rate": 3.6570269062268025e-07, | |
| "logits/chosen": -0.8680559992790222, | |
| "logits/rejected": -0.8241602182388306, | |
| "logps/chosen": -813.6458740234375, | |
| "logps/rejected": -992.6861572265625, | |
| "loss": 0.3533, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -5.124260425567627, | |
| "rewards/margins": 1.764125108718872, | |
| "rewards/rejected": -6.88838529586792, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.5386053967405824, | |
| "grad_norm": 22.37517195362321, | |
| "learning_rate": 3.6308691799671404e-07, | |
| "logits/chosen": -0.7893460392951965, | |
| "logits/rejected": -0.8846604824066162, | |
| "logps/chosen": -585.7125854492188, | |
| "logps/rejected": -700.5401611328125, | |
| "loss": 0.4298, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -3.845160961151123, | |
| "rewards/margins": 1.1050822734832764, | |
| "rewards/rejected": -4.9502434730529785, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.5407427197435213, | |
| "grad_norm": 20.261154485800315, | |
| "learning_rate": 3.604704131592895e-07, | |
| "logits/chosen": -1.1260894536972046, | |
| "logits/rejected": -1.116862177848816, | |
| "logps/chosen": -766.9825439453125, | |
| "logps/rejected": -869.4150390625, | |
| "loss": 0.4666, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.645950794219971, | |
| "rewards/margins": 1.14671790599823, | |
| "rewards/rejected": -5.79266881942749, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.5428800427464601, | |
| "grad_norm": 24.637498347289263, | |
| "learning_rate": 3.578533225035318e-07, | |
| "logits/chosen": -1.0664944648742676, | |
| "logits/rejected": -1.0649199485778809, | |
| "logps/chosen": -772.8311157226562, | |
| "logps/rejected": -995.858642578125, | |
| "loss": 0.3884, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.625110149383545, | |
| "rewards/margins": 2.349405527114868, | |
| "rewards/rejected": -6.974515914916992, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.5450173657493989, | |
| "grad_norm": 18.22340364830364, | |
| "learning_rate": 3.552357924553423e-07, | |
| "logits/chosen": -0.9930300712585449, | |
| "logits/rejected": -1.0258715152740479, | |
| "logps/chosen": -745.672119140625, | |
| "logps/rejected": -877.061279296875, | |
| "loss": 0.4139, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -4.158998489379883, | |
| "rewards/margins": 1.6085114479064941, | |
| "rewards/rejected": -5.767509937286377, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.5471546887523377, | |
| "grad_norm": 16.289391048607587, | |
| "learning_rate": 3.526179694652065e-07, | |
| "logits/chosen": -1.0990363359451294, | |
| "logits/rejected": -1.0807216167449951, | |
| "logps/chosen": -646.4122314453125, | |
| "logps/rejected": -707.5439453125, | |
| "loss": 0.4572, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -4.093816757202148, | |
| "rewards/margins": 0.4628375768661499, | |
| "rewards/rejected": -4.556654453277588, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.5492920117552765, | |
| "grad_norm": 20.631321302576914, | |
| "learning_rate": 3.5e-07, | |
| "logits/chosen": -1.1180474758148193, | |
| "logits/rejected": -1.0840083360671997, | |
| "logps/chosen": -714.1011352539062, | |
| "logps/rejected": -901.5518798828125, | |
| "loss": 0.5129, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.527349948883057, | |
| "rewards/margins": 1.6243555545806885, | |
| "rewards/rejected": -6.151705265045166, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.5514293347582153, | |
| "grad_norm": 17.30980385185334, | |
| "learning_rate": 3.4738203053479345e-07, | |
| "logits/chosen": -1.1147381067276, | |
| "logits/rejected": -1.0951800346374512, | |
| "logps/chosen": -671.677978515625, | |
| "logps/rejected": -869.009033203125, | |
| "loss": 0.3768, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -4.159421443939209, | |
| "rewards/margins": 1.6203829050064087, | |
| "rewards/rejected": -5.77980375289917, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.5535666577611541, | |
| "grad_norm": 16.77670843921571, | |
| "learning_rate": 3.447642075446578e-07, | |
| "logits/chosen": -1.027138113975525, | |
| "logits/rejected": -1.0414177179336548, | |
| "logps/chosen": -600.3823852539062, | |
| "logps/rejected": -688.4371337890625, | |
| "loss": 0.3932, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.747048854827881, | |
| "rewards/margins": 0.837365448474884, | |
| "rewards/rejected": -4.584414482116699, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.555703980764093, | |
| "grad_norm": 13.731293304469197, | |
| "learning_rate": 3.421466774964683e-07, | |
| "logits/chosen": -1.060163140296936, | |
| "logits/rejected": -1.1123335361480713, | |
| "logps/chosen": -769.50244140625, | |
| "logps/rejected": -878.5049438476562, | |
| "loss": 0.4123, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.471192359924316, | |
| "rewards/margins": 1.6601918935775757, | |
| "rewards/rejected": -6.131383895874023, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.5578413037670318, | |
| "grad_norm": 21.94893337829568, | |
| "learning_rate": 3.395295868407106e-07, | |
| "logits/chosen": -0.7995742559432983, | |
| "logits/rejected": -0.7949234247207642, | |
| "logps/chosen": -540.7459716796875, | |
| "logps/rejected": -606.6634521484375, | |
| "loss": 0.4368, | |
| "rewards/accuracies": 0.5625, | |
| "rewards/chosen": -3.475754976272583, | |
| "rewards/margins": 0.5799713134765625, | |
| "rewards/rejected": -4.055726051330566, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.5599786267699706, | |
| "grad_norm": 18.73889178987264, | |
| "learning_rate": 3.3691308200328604e-07, | |
| "logits/chosen": -0.8902988433837891, | |
| "logits/rejected": -0.9296258687973022, | |
| "logps/chosen": -738.5679931640625, | |
| "logps/rejected": -910.6119384765625, | |
| "loss": 0.4267, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.609591007232666, | |
| "rewards/margins": 1.8606202602386475, | |
| "rewards/rejected": -6.470211029052734, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.5621159497729095, | |
| "grad_norm": 17.495724524128903, | |
| "learning_rate": 3.342973093773199e-07, | |
| "logits/chosen": -1.0284425020217896, | |
| "logits/rejected": -1.0994598865509033, | |
| "logps/chosen": -763.5560302734375, | |
| "logps/rejected": -916.3192749023438, | |
| "loss": 0.4193, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -4.318557262420654, | |
| "rewards/margins": 1.6308737993240356, | |
| "rewards/rejected": -5.949431419372559, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.5642532727758482, | |
| "grad_norm": 21.8003372321031, | |
| "learning_rate": 3.316824153149696e-07, | |
| "logits/chosen": -0.9204785823822021, | |
| "logits/rejected": -0.9681702852249146, | |
| "logps/chosen": -642.7691650390625, | |
| "logps/rejected": -805.9739990234375, | |
| "loss": 0.4396, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -4.044567108154297, | |
| "rewards/margins": 1.6253376007080078, | |
| "rewards/rejected": -5.669903755187988, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.566390595778787, | |
| "grad_norm": 29.221196391296576, | |
| "learning_rate": 3.29068546119238e-07, | |
| "logits/chosen": -0.8801388740539551, | |
| "logits/rejected": -0.8704509139060974, | |
| "logps/chosen": -707.189453125, | |
| "logps/rejected": -804.7384033203125, | |
| "loss": 0.597, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -4.639124870300293, | |
| "rewards/margins": 1.1081290245056152, | |
| "rewards/rejected": -5.74725341796875, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.5685279187817259, | |
| "grad_norm": 21.440238859390956, | |
| "learning_rate": 3.2645584803578614e-07, | |
| "logits/chosen": -0.9265174865722656, | |
| "logits/rejected": -0.9847102761268616, | |
| "logps/chosen": -483.4872131347656, | |
| "logps/rejected": -527.4281005859375, | |
| "loss": 0.44, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.243124008178711, | |
| "rewards/margins": 0.43065446615219116, | |
| "rewards/rejected": -3.6737782955169678, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.5706652417846647, | |
| "grad_norm": 15.539782944871728, | |
| "learning_rate": 3.238444672447515e-07, | |
| "logits/chosen": -0.8670752048492432, | |
| "logits/rejected": -0.8761327862739563, | |
| "logps/chosen": -669.8744506835938, | |
| "logps/rejected": -792.420166015625, | |
| "loss": 0.4026, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.5057759284973145, | |
| "rewards/margins": 1.2092504501342773, | |
| "rewards/rejected": -5.715025901794434, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.5728025647876035, | |
| "grad_norm": 15.529372167608297, | |
| "learning_rate": 3.2123454985256943e-07, | |
| "logits/chosen": -0.9794555902481079, | |
| "logits/rejected": -0.9547824859619141, | |
| "logps/chosen": -610.2971801757812, | |
| "logps/rejected": -825.6856079101562, | |
| "loss": 0.3279, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -3.7575111389160156, | |
| "rewards/margins": 2.0598204135894775, | |
| "rewards/rejected": -5.8173322677612305, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.5749398877905424, | |
| "grad_norm": 22.72130516871319, | |
| "learning_rate": 3.186262418837983e-07, | |
| "logits/chosen": -1.082902431488037, | |
| "logits/rejected": -1.1030265092849731, | |
| "logps/chosen": -545.2824096679688, | |
| "logps/rejected": -609.6895141601562, | |
| "loss": 0.5641, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -3.404395580291748, | |
| "rewards/margins": 0.5698573589324951, | |
| "rewards/rejected": -3.974252939224243, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.5770772107934812, | |
| "grad_norm": 16.252579751792183, | |
| "learning_rate": 3.1601968927294914e-07, | |
| "logits/chosen": -0.9424474239349365, | |
| "logits/rejected": -0.9634624123573303, | |
| "logps/chosen": -599.9176025390625, | |
| "logps/rejected": -702.5125732421875, | |
| "loss": 0.4418, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.833646774291992, | |
| "rewards/margins": 0.9874881505966187, | |
| "rewards/rejected": -4.821134567260742, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.57921453379642, | |
| "grad_norm": 15.206487367829114, | |
| "learning_rate": 3.134150378563213e-07, | |
| "logits/chosen": -0.9836793541908264, | |
| "logits/rejected": -0.9680826663970947, | |
| "logps/chosen": -637.9475708007812, | |
| "logps/rejected": -767.5353393554688, | |
| "loss": 0.3601, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.120051383972168, | |
| "rewards/margins": 1.235574722290039, | |
| "rewards/rejected": -5.355626106262207, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.5813518567993589, | |
| "grad_norm": 21.078417332237123, | |
| "learning_rate": 3.1081243336384226e-07, | |
| "logits/chosen": -0.9045091271400452, | |
| "logits/rejected": -0.9048910140991211, | |
| "logps/chosen": -615.5716552734375, | |
| "logps/rejected": -752.4681396484375, | |
| "loss": 0.3887, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.629241704940796, | |
| "rewards/margins": 1.4587962627410889, | |
| "rewards/rejected": -5.088037490844727, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.5834891798022976, | |
| "grad_norm": 17.81077343546014, | |
| "learning_rate": 3.0821202141091453e-07, | |
| "logits/chosen": -1.1009454727172852, | |
| "logits/rejected": -0.9980039596557617, | |
| "logps/chosen": -591.7613525390625, | |
| "logps/rejected": -779.107421875, | |
| "loss": 0.4669, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.549128770828247, | |
| "rewards/margins": 1.750599980354309, | |
| "rewards/rejected": -5.299729824066162, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.5856265028052364, | |
| "grad_norm": 19.256523288494076, | |
| "learning_rate": 3.0561394749026814e-07, | |
| "logits/chosen": -1.008500576019287, | |
| "logits/rejected": -0.9249081611633301, | |
| "logps/chosen": -631.9851684570312, | |
| "logps/rejected": -834.7702026367188, | |
| "loss": 0.4028, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.780165195465088, | |
| "rewards/margins": 1.7658891677856445, | |
| "rewards/rejected": -5.546054363250732, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.5877638258081752, | |
| "grad_norm": 21.051631953087565, | |
| "learning_rate": 3.030183569638207e-07, | |
| "logits/chosen": -1.1362390518188477, | |
| "logits/rejected": -1.0792036056518555, | |
| "logps/chosen": -576.5594482421875, | |
| "logps/rejected": -694.7816162109375, | |
| "loss": 0.4611, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.658750534057617, | |
| "rewards/margins": 1.0495951175689697, | |
| "rewards/rejected": -4.708345890045166, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.5899011488111141, | |
| "grad_norm": 15.76447884471548, | |
| "learning_rate": 3.0042539505454386e-07, | |
| "logits/chosen": -0.9091801643371582, | |
| "logits/rejected": -0.9027244448661804, | |
| "logps/chosen": -592.4364013671875, | |
| "logps/rejected": -717.2314453125, | |
| "loss": 0.3772, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.7294676303863525, | |
| "rewards/margins": 1.2672703266143799, | |
| "rewards/rejected": -4.996737957000732, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.5920384718140529, | |
| "grad_norm": 14.607600516553031, | |
| "learning_rate": 2.978352068383389e-07, | |
| "logits/chosen": -0.9434270858764648, | |
| "logits/rejected": -0.9658694267272949, | |
| "logps/chosen": -426.6009216308594, | |
| "logps/rejected": -538.635009765625, | |
| "loss": 0.3716, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.5574347972869873, | |
| "rewards/margins": 1.0464235544204712, | |
| "rewards/rejected": -3.603858470916748, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.5941757948169917, | |
| "grad_norm": 20.680956313480134, | |
| "learning_rate": 2.952479372359191e-07, | |
| "logits/chosen": -0.9002370238304138, | |
| "logits/rejected": -0.9020406007766724, | |
| "logps/chosen": -564.5108642578125, | |
| "logps/rejected": -660.9511108398438, | |
| "loss": 0.3884, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -3.561006784439087, | |
| "rewards/margins": 0.8000065088272095, | |
| "rewards/rejected": -4.361013412475586, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.5963131178199306, | |
| "grad_norm": 27.6993585691743, | |
| "learning_rate": 2.926637310047018e-07, | |
| "logits/chosen": -1.1155834197998047, | |
| "logits/rejected": -1.105792760848999, | |
| "logps/chosen": -596.0205078125, | |
| "logps/rejected": -664.81640625, | |
| "loss": 0.3998, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.8352932929992676, | |
| "rewards/margins": 0.8193560838699341, | |
| "rewards/rejected": -4.654649257659912, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.5984504408228694, | |
| "grad_norm": 18.916329978828376, | |
| "learning_rate": 2.900827327307087e-07, | |
| "logits/chosen": -1.2077603340148926, | |
| "logits/rejected": -1.0835894346237183, | |
| "logps/chosen": -778.5531005859375, | |
| "logps/rejected": -890.9183349609375, | |
| "loss": 0.4658, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -4.6639227867126465, | |
| "rewards/margins": 1.0659689903259277, | |
| "rewards/rejected": -5.729891777038574, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.6005877638258081, | |
| "grad_norm": 18.365489297240394, | |
| "learning_rate": 2.8750508682047717e-07, | |
| "logits/chosen": -1.1601464748382568, | |
| "logits/rejected": -1.1277116537094116, | |
| "logps/chosen": -677.4540405273438, | |
| "logps/rejected": -838.7645874023438, | |
| "loss": 0.4918, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.8409969806671143, | |
| "rewards/margins": 1.480958342552185, | |
| "rewards/rejected": -5.321955680847168, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.602725086828747, | |
| "grad_norm": 20.141526963645916, | |
| "learning_rate": 2.849309374929799e-07, | |
| "logits/chosen": -1.1542294025421143, | |
| "logits/rejected": -1.1290372610092163, | |
| "logps/chosen": -638.0623779296875, | |
| "logps/rejected": -759.2158203125, | |
| "loss": 0.4268, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.9765753746032715, | |
| "rewards/margins": 1.2076830863952637, | |
| "rewards/rejected": -5.184258460998535, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.6048624098316858, | |
| "grad_norm": 17.776205208549275, | |
| "learning_rate": 2.823604287715564e-07, | |
| "logits/chosen": -0.9990053772926331, | |
| "logits/rejected": -1.038635015487671, | |
| "logps/chosen": -641.4173583984375, | |
| "logps/rejected": -752.838623046875, | |
| "loss": 0.3768, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -3.788572072982788, | |
| "rewards/margins": 1.1508256196975708, | |
| "rewards/rejected": -4.939397811889648, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.6069997328346246, | |
| "grad_norm": 26.51717001409862, | |
| "learning_rate": 2.797937044758545e-07, | |
| "logits/chosen": -0.9398884177207947, | |
| "logits/rejected": -0.9421570897102356, | |
| "logps/chosen": -616.0810546875, | |
| "logps/rejected": -697.7490844726562, | |
| "loss": 0.4983, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -4.009638786315918, | |
| "rewards/margins": 0.752707302570343, | |
| "rewards/rejected": -4.762345790863037, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.6091370558375635, | |
| "grad_norm": 35.30303590344862, | |
| "learning_rate": 2.772309082137843e-07, | |
| "logits/chosen": -1.085971474647522, | |
| "logits/rejected": -1.0797044038772583, | |
| "logps/chosen": -635.1014404296875, | |
| "logps/rejected": -687.0366821289062, | |
| "loss": 0.5866, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -3.941286325454712, | |
| "rewards/margins": 0.7452540397644043, | |
| "rewards/rejected": -4.686540603637695, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.6112743788405023, | |
| "grad_norm": 26.601600287549388, | |
| "learning_rate": 2.7467218337348233e-07, | |
| "logits/chosen": -0.8175097703933716, | |
| "logits/rejected": -0.7498407959938049, | |
| "logps/chosen": -589.5838623046875, | |
| "logps/rejected": -743.729248046875, | |
| "loss": 0.4928, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.852874279022217, | |
| "rewards/margins": 1.4163994789123535, | |
| "rewards/rejected": -5.2692742347717285, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.6134117018434411, | |
| "grad_norm": 17.687526214496195, | |
| "learning_rate": 2.7211767311529e-07, | |
| "logits/chosen": -0.8146362900733948, | |
| "logits/rejected": -0.8423186540603638, | |
| "logps/chosen": -607.890869140625, | |
| "logps/rejected": -708.3568115234375, | |
| "loss": 0.4853, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -3.738939046859741, | |
| "rewards/margins": 0.9543510675430298, | |
| "rewards/rejected": -4.693289756774902, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.6155490248463799, | |
| "grad_norm": 18.930801664854705, | |
| "learning_rate": 2.695675203637432e-07, | |
| "logits/chosen": -1.0421204566955566, | |
| "logits/rejected": -1.0792558193206787, | |
| "logps/chosen": -602.5674438476562, | |
| "logps/rejected": -736.7977905273438, | |
| "loss": 0.4063, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.712022304534912, | |
| "rewards/margins": 1.2383270263671875, | |
| "rewards/rejected": -4.9503493309021, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.6176863478493188, | |
| "grad_norm": 17.90358421564856, | |
| "learning_rate": 2.670218677995759e-07, | |
| "logits/chosen": -1.0456268787384033, | |
| "logits/rejected": -0.9645753502845764, | |
| "logps/chosen": -618.1510009765625, | |
| "logps/rejected": -819.269287109375, | |
| "loss": 0.4206, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -3.5795562267303467, | |
| "rewards/margins": 2.006627321243286, | |
| "rewards/rejected": -5.586184024810791, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.6198236708522575, | |
| "grad_norm": 23.380000713489068, | |
| "learning_rate": 2.6448085785173703e-07, | |
| "logits/chosen": -1.005885124206543, | |
| "logits/rejected": -0.964793860912323, | |
| "logps/chosen": -665.0631103515625, | |
| "logps/rejected": -754.4349975585938, | |
| "loss": 0.4141, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -4.058958053588867, | |
| "rewards/margins": 0.8748332262039185, | |
| "rewards/rejected": -4.933791160583496, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.6219609938551963, | |
| "grad_norm": 16.59292554666212, | |
| "learning_rate": 2.6194463268942203e-07, | |
| "logits/chosen": -1.0780235528945923, | |
| "logits/rejected": -1.097104787826538, | |
| "logps/chosen": -590.5000610351562, | |
| "logps/rejected": -796.20166015625, | |
| "loss": 0.3928, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.4696645736694336, | |
| "rewards/margins": 1.882230281829834, | |
| "rewards/rejected": -5.351894855499268, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.6240983168581352, | |
| "grad_norm": 14.429424103488177, | |
| "learning_rate": 2.594133342141177e-07, | |
| "logits/chosen": -0.9270427227020264, | |
| "logits/rejected": -0.8782027959823608, | |
| "logps/chosen": -498.6771545410156, | |
| "logps/rejected": -686.7008056640625, | |
| "loss": 0.3666, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -2.730340003967285, | |
| "rewards/margins": 1.7681416273117065, | |
| "rewards/rejected": -4.498481273651123, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.626235639861074, | |
| "grad_norm": 21.064661843353115, | |
| "learning_rate": 2.568871040516637e-07, | |
| "logits/chosen": -1.0022443532943726, | |
| "logits/rejected": -0.9993023872375488, | |
| "logps/chosen": -685.757080078125, | |
| "logps/rejected": -779.1624755859375, | |
| "loss": 0.4839, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -4.1098246574401855, | |
| "rewards/margins": 0.944150984287262, | |
| "rewards/rejected": -5.053975582122803, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.6283729628640128, | |
| "grad_norm": 19.087071179960915, | |
| "learning_rate": 2.54366083544328e-07, | |
| "logits/chosen": -0.8884956240653992, | |
| "logits/rejected": -0.9064831137657166, | |
| "logps/chosen": -571.4553833007812, | |
| "logps/rejected": -674.0166625976562, | |
| "loss": 0.504, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.6071152687072754, | |
| "rewards/margins": 0.9564936757087708, | |
| "rewards/rejected": -4.563608646392822, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.6305102858669517, | |
| "grad_norm": 21.794412344839774, | |
| "learning_rate": 2.5185041374289907e-07, | |
| "logits/chosen": -1.1455574035644531, | |
| "logits/rejected": -1.1415196657180786, | |
| "logps/chosen": -497.8589782714844, | |
| "logps/rejected": -751.036376953125, | |
| "loss": 0.4041, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -2.744825839996338, | |
| "rewards/margins": 2.1894543170928955, | |
| "rewards/rejected": -4.934279918670654, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.6326476088698905, | |
| "grad_norm": 19.554010929778244, | |
| "learning_rate": 2.4934023539879395e-07, | |
| "logits/chosen": -0.979045569896698, | |
| "logits/rejected": -0.9661969542503357, | |
| "logps/chosen": -601.6738891601562, | |
| "logps/rejected": -756.327392578125, | |
| "loss": 0.3696, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.3881332874298096, | |
| "rewards/margins": 1.3837690353393555, | |
| "rewards/rejected": -4.771902561187744, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.6347849318728293, | |
| "grad_norm": 20.678635374709422, | |
| "learning_rate": 2.468356889561835e-07, | |
| "logits/chosen": -1.0623462200164795, | |
| "logits/rejected": -1.0434157848358154, | |
| "logps/chosen": -569.0560302734375, | |
| "logps/rejected": -698.5628051757812, | |
| "loss": 0.3802, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.12467622756958, | |
| "rewards/margins": 1.3073335886001587, | |
| "rewards/rejected": -4.432010173797607, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.6369222548757681, | |
| "grad_norm": 15.887777705629667, | |
| "learning_rate": 2.443369145441343e-07, | |
| "logits/chosen": -1.0596822500228882, | |
| "logits/rejected": -1.044713020324707, | |
| "logps/chosen": -531.6007080078125, | |
| "logps/rejected": -636.859130859375, | |
| "loss": 0.3491, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.2483232021331787, | |
| "rewards/margins": 1.0054796934127808, | |
| "rewards/rejected": -4.253802299499512, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.6390595778787069, | |
| "grad_norm": 20.882809623209052, | |
| "learning_rate": 2.418440519687684e-07, | |
| "logits/chosen": -1.0757739543914795, | |
| "logits/rejected": -1.0567525625228882, | |
| "logps/chosen": -511.7173156738281, | |
| "logps/rejected": -668.195556640625, | |
| "loss": 0.4593, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -3.052992820739746, | |
| "rewards/margins": 1.4922919273376465, | |
| "rewards/rejected": -4.545284271240234, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.6411969008816457, | |
| "grad_norm": 14.618996262053995, | |
| "learning_rate": 2.393572407054416e-07, | |
| "logits/chosen": -0.8958306312561035, | |
| "logits/rejected": -0.8867323398590088, | |
| "logps/chosen": -659.5025634765625, | |
| "logps/rejected": -834.076171875, | |
| "loss": 0.3802, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.8700366020202637, | |
| "rewards/margins": 1.7134443521499634, | |
| "rewards/rejected": -5.5834808349609375, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.6411969008816457, | |
| "eval_logits/chosen": -1.1104539632797241, | |
| "eval_logits/rejected": -1.130950927734375, | |
| "eval_logps/chosen": -652.87158203125, | |
| "eval_logps/rejected": -784.0881958007812, | |
| "eval_loss": 0.40694576501846313, | |
| "eval_rewards/accuracies": 0.8427419066429138, | |
| "eval_rewards/chosen": -3.7974159717559814, | |
| "eval_rewards/margins": 1.3731279373168945, | |
| "eval_rewards/rejected": -5.170544147491455, | |
| "eval_runtime": 127.0172, | |
| "eval_samples_per_second": 15.439, | |
| "eval_steps_per_second": 0.488, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.6433342238845846, | |
| "grad_norm": 19.43800281730569, | |
| "learning_rate": 2.3687661989093904e-07, | |
| "logits/chosen": -0.9887138605117798, | |
| "logits/rejected": -1.0439525842666626, | |
| "logps/chosen": -595.8880615234375, | |
| "logps/rejected": -732.6058349609375, | |
| "loss": 0.4198, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.636826753616333, | |
| "rewards/margins": 1.4018120765686035, | |
| "rewards/rejected": -5.038639068603516, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.6454715468875234, | |
| "grad_norm": 22.456716854896456, | |
| "learning_rate": 2.3440232831569162e-07, | |
| "logits/chosen": -1.0877864360809326, | |
| "logits/rejected": -1.1251873970031738, | |
| "logps/chosen": -665.487548828125, | |
| "logps/rejected": -819.5171508789062, | |
| "loss": 0.4536, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.9592628479003906, | |
| "rewards/margins": 1.4847943782806396, | |
| "rewards/rejected": -5.444056987762451, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.6476088698904622, | |
| "grad_norm": 16.0762099994994, | |
| "learning_rate": 2.319345044160095e-07, | |
| "logits/chosen": -1.2076233625411987, | |
| "logits/rejected": -1.230367660522461, | |
| "logps/chosen": -769.3408203125, | |
| "logps/rejected": -1005.293701171875, | |
| "loss": 0.4059, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -4.611854076385498, | |
| "rewards/margins": 2.47025465965271, | |
| "rewards/rejected": -7.082108974456787, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.649746192893401, | |
| "grad_norm": 17.96340906701677, | |
| "learning_rate": 2.2947328626633766e-07, | |
| "logits/chosen": -1.058774709701538, | |
| "logits/rejected": -1.0470080375671387, | |
| "logps/chosen": -590.3370361328125, | |
| "logps/rejected": -681.3898315429688, | |
| "loss": 0.4649, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.8478586673736572, | |
| "rewards/margins": 0.7897017002105713, | |
| "rewards/rejected": -4.637560844421387, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.6518835158963399, | |
| "grad_norm": 22.719738681967296, | |
| "learning_rate": 2.2701881157153e-07, | |
| "logits/chosen": -1.1047711372375488, | |
| "logits/rejected": -1.0974376201629639, | |
| "logps/chosen": -653.9512329101562, | |
| "logps/rejected": -839.6368408203125, | |
| "loss": 0.3939, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.9316320419311523, | |
| "rewards/margins": 1.9634764194488525, | |
| "rewards/rejected": -5.895109176635742, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.6540208388992786, | |
| "grad_norm": 17.780668266782456, | |
| "learning_rate": 2.245712176591449e-07, | |
| "logits/chosen": -1.0736472606658936, | |
| "logits/rejected": -1.0763237476348877, | |
| "logps/chosen": -615.7779541015625, | |
| "logps/rejected": -687.5728759765625, | |
| "loss": 0.4217, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.957977533340454, | |
| "rewards/margins": 0.773561954498291, | |
| "rewards/rejected": -4.731539726257324, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.6561581619022174, | |
| "grad_norm": 23.74418488307736, | |
| "learning_rate": 2.2213064147176174e-07, | |
| "logits/chosen": -1.094491958618164, | |
| "logits/rejected": -1.1138453483581543, | |
| "logps/chosen": -725.0467529296875, | |
| "logps/rejected": -822.287353515625, | |
| "loss": 0.4067, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -4.600536823272705, | |
| "rewards/margins": 1.1523770093917847, | |
| "rewards/rejected": -5.752913951873779, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.6582954849051563, | |
| "grad_norm": 15.5978905581422, | |
| "learning_rate": 2.196972195593192e-07, | |
| "logits/chosen": -1.1050280332565308, | |
| "logits/rejected": -1.0931296348571777, | |
| "logps/chosen": -582.18115234375, | |
| "logps/rejected": -741.2723388671875, | |
| "loss": 0.3451, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.750410318374634, | |
| "rewards/margins": 1.5357167720794678, | |
| "rewards/rejected": -5.28612756729126, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.6604328079080951, | |
| "grad_norm": 17.049613660365612, | |
| "learning_rate": 2.1727108807147506e-07, | |
| "logits/chosen": -1.0725035667419434, | |
| "logits/rejected": -1.1041122674942017, | |
| "logps/chosen": -610.0225219726562, | |
| "logps/rejected": -762.8485107421875, | |
| "loss": 0.4524, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.9435532093048096, | |
| "rewards/margins": 1.720339059829712, | |
| "rewards/rejected": -5.663891792297363, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.6625701309110339, | |
| "grad_norm": 21.589679932544243, | |
| "learning_rate": 2.148523827499888e-07, | |
| "logits/chosen": -0.9776382446289062, | |
| "logits/rejected": -1.0473860502243042, | |
| "logps/chosen": -845.308837890625, | |
| "logps/rejected": -1063.19482421875, | |
| "loss": 0.4161, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -5.683403491973877, | |
| "rewards/margins": 1.968895673751831, | |
| "rewards/rejected": -7.652299404144287, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.6647074539139728, | |
| "grad_norm": 20.546655366561865, | |
| "learning_rate": 2.1244123892112674e-07, | |
| "logits/chosen": -0.971989095211029, | |
| "logits/rejected": -0.9635407328605652, | |
| "logps/chosen": -752.0794677734375, | |
| "logps/rejected": -896.9614868164062, | |
| "loss": 0.4412, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.961668968200684, | |
| "rewards/margins": 1.418576717376709, | |
| "rewards/rejected": -6.380245685577393, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.6668447769169116, | |
| "grad_norm": 16.64812094871957, | |
| "learning_rate": 2.1003779148809072e-07, | |
| "logits/chosen": -0.9267586469650269, | |
| "logits/rejected": -0.9304214715957642, | |
| "logps/chosen": -742.295166015625, | |
| "logps/rejected": -886.9481201171875, | |
| "loss": 0.3719, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.672271251678467, | |
| "rewards/margins": 1.6211450099945068, | |
| "rewards/rejected": -6.2934160232543945, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.6689820999198504, | |
| "grad_norm": 21.010619081649534, | |
| "learning_rate": 2.0764217492346996e-07, | |
| "logits/chosen": -1.0951707363128662, | |
| "logits/rejected": -1.0666468143463135, | |
| "logps/chosen": -790.2965087890625, | |
| "logps/rejected": -910.8655395507812, | |
| "loss": 0.4601, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -5.142488956451416, | |
| "rewards/margins": 1.2062201499938965, | |
| "rewards/rejected": -6.348708629608154, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.6711194229227893, | |
| "grad_norm": 19.84084802689762, | |
| "learning_rate": 2.0525452326171784e-07, | |
| "logits/chosen": -1.0453927516937256, | |
| "logits/rejected": -1.0439743995666504, | |
| "logps/chosen": -865.6875610351562, | |
| "logps/rejected": -1150.561767578125, | |
| "loss": 0.4036, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -5.613504409790039, | |
| "rewards/margins": 2.756046772003174, | |
| "rewards/rejected": -8.369551658630371, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.673256745925728, | |
| "grad_norm": 17.469930397973854, | |
| "learning_rate": 2.028749700916521e-07, | |
| "logits/chosen": -0.955051600933075, | |
| "logits/rejected": -0.944666862487793, | |
| "logps/chosen": -787.5214233398438, | |
| "logps/rejected": -877.498779296875, | |
| "loss": 0.403, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -5.070991039276123, | |
| "rewards/margins": 0.8326537013053894, | |
| "rewards/rejected": -5.903645038604736, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.6753940689286668, | |
| "grad_norm": 31.931815437246083, | |
| "learning_rate": 2.0050364854898127e-07, | |
| "logits/chosen": -1.066016674041748, | |
| "logits/rejected": -1.063683032989502, | |
| "logps/chosen": -740.451416015625, | |
| "logps/rejected": -935.1329345703125, | |
| "loss": 0.497, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.710479736328125, | |
| "rewards/margins": 1.9114099740982056, | |
| "rewards/rejected": -6.621890068054199, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.6775313919316056, | |
| "grad_norm": 23.195446430699608, | |
| "learning_rate": 1.9814069130885467e-07, | |
| "logits/chosen": -0.7954821586608887, | |
| "logits/rejected": -0.8669359683990479, | |
| "logps/chosen": -679.1461791992188, | |
| "logps/rejected": -868.0947875976562, | |
| "loss": 0.4159, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.395887851715088, | |
| "rewards/margins": 2.0239951610565186, | |
| "rewards/rejected": -6.4198832511901855, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.6796687149345445, | |
| "grad_norm": 27.100165283554862, | |
| "learning_rate": 1.9578623057844064e-07, | |
| "logits/chosen": -1.1223851442337036, | |
| "logits/rejected": -1.1304949522018433, | |
| "logps/chosen": -606.3875732421875, | |
| "logps/rejected": -768.2408447265625, | |
| "loss": 0.4085, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -3.690753936767578, | |
| "rewards/margins": 1.4597177505493164, | |
| "rewards/rejected": -5.1504716873168945, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.6818060379374833, | |
| "grad_norm": 19.59121701664144, | |
| "learning_rate": 1.9344039808952853e-07, | |
| "logits/chosen": -0.9872533082962036, | |
| "logits/rejected": -0.934008002281189, | |
| "logps/chosen": -740.6689453125, | |
| "logps/rejected": -1047.7109375, | |
| "loss": 0.3775, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -4.856841087341309, | |
| "rewards/margins": 2.8627772331237793, | |
| "rewards/rejected": -7.71961784362793, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.6839433609404221, | |
| "grad_norm": 20.56349471919662, | |
| "learning_rate": 1.9110332509115864e-07, | |
| "logits/chosen": -1.0536935329437256, | |
| "logits/rejected": -1.111939549446106, | |
| "logps/chosen": -713.4016723632812, | |
| "logps/rejected": -897.5386962890625, | |
| "loss": 0.4077, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.394659996032715, | |
| "rewards/margins": 1.7375450134277344, | |
| "rewards/rejected": -6.132205009460449, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.686080683943361, | |
| "grad_norm": 21.59919721276828, | |
| "learning_rate": 1.8877514234227874e-07, | |
| "logits/chosen": -0.8561320900917053, | |
| "logits/rejected": -0.9341483116149902, | |
| "logps/chosen": -669.7821044921875, | |
| "logps/rejected": -811.287353515625, | |
| "loss": 0.4224, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.155871391296387, | |
| "rewards/margins": 1.4691696166992188, | |
| "rewards/rejected": -5.6250410079956055, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.6882180069462998, | |
| "grad_norm": 19.329148607505996, | |
| "learning_rate": 1.8645598010442828e-07, | |
| "logits/chosen": -0.9411246180534363, | |
| "logits/rejected": -0.9587753415107727, | |
| "logps/chosen": -699.0866088867188, | |
| "logps/rejected": -864.09375, | |
| "loss": 0.3697, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.35011100769043, | |
| "rewards/margins": 1.8306729793548584, | |
| "rewards/rejected": -6.180783748626709, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.6903553299492385, | |
| "grad_norm": 16.796887314919882, | |
| "learning_rate": 1.8414596813445047e-07, | |
| "logits/chosen": -0.9905204772949219, | |
| "logits/rejected": -1.024290680885315, | |
| "logps/chosen": -622.2347412109375, | |
| "logps/rejected": -708.5049438476562, | |
| "loss": 0.407, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -3.9594295024871826, | |
| "rewards/margins": 0.9213578104972839, | |
| "rewards/rejected": -4.8807878494262695, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.6924926529521774, | |
| "grad_norm": 15.1418538158911, | |
| "learning_rate": 1.8184523567723194e-07, | |
| "logits/chosen": -1.0796232223510742, | |
| "logits/rejected": -1.0750352144241333, | |
| "logps/chosen": -628.67041015625, | |
| "logps/rejected": -812.5968017578125, | |
| "loss": 0.351, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -3.8600831031799316, | |
| "rewards/margins": 1.7709476947784424, | |
| "rewards/rejected": -5.631031036376953, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.6946299759551162, | |
| "grad_norm": 18.026973060407244, | |
| "learning_rate": 1.7955391145847145e-07, | |
| "logits/chosen": -0.9602807760238647, | |
| "logits/rejected": -0.9538272619247437, | |
| "logps/chosen": -533.8419189453125, | |
| "logps/rejected": -605.8224487304688, | |
| "loss": 0.338, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.626598596572876, | |
| "rewards/margins": 0.6452375054359436, | |
| "rewards/rejected": -4.271836280822754, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.696767298958055, | |
| "grad_norm": 20.22808705229649, | |
| "learning_rate": 1.772721236774786e-07, | |
| "logits/chosen": -0.8902896642684937, | |
| "logits/rejected": -0.891946017742157, | |
| "logps/chosen": -547.3910522460938, | |
| "logps/rejected": -737.225830078125, | |
| "loss": 0.3568, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.2871787548065186, | |
| "rewards/margins": 1.8806804418563843, | |
| "rewards/rejected": -5.167859077453613, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.6989046219609939, | |
| "grad_norm": 24.61898214966437, | |
| "learning_rate": 1.7500000000000007e-07, | |
| "logits/chosen": -1.0208594799041748, | |
| "logits/rejected": -0.9971433877944946, | |
| "logps/chosen": -628.2722778320312, | |
| "logps/rejected": -766.7388916015625, | |
| "loss": 0.4498, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.7911126613616943, | |
| "rewards/margins": 1.395633339881897, | |
| "rewards/rejected": -5.186745643615723, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.7010419449639327, | |
| "grad_norm": 23.202132151091234, | |
| "learning_rate": 1.7273766755107707e-07, | |
| "logits/chosen": -1.00026535987854, | |
| "logits/rejected": -1.0049934387207031, | |
| "logps/chosen": -618.7490234375, | |
| "logps/rejected": -784.1082153320312, | |
| "loss": 0.4044, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.9590413570404053, | |
| "rewards/margins": 1.4626474380493164, | |
| "rewards/rejected": -5.421689033508301, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.7031792679668715, | |
| "grad_norm": 22.040873223414177, | |
| "learning_rate": 1.7048525290793284e-07, | |
| "logits/chosen": -1.151913046836853, | |
| "logits/rejected": -1.1811047792434692, | |
| "logps/chosen": -676.8132934570312, | |
| "logps/rejected": -835.91064453125, | |
| "loss": 0.4432, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.948080062866211, | |
| "rewards/margins": 1.632805347442627, | |
| "rewards/rejected": -5.580885887145996, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.7053165909698104, | |
| "grad_norm": 16.489066064805076, | |
| "learning_rate": 1.6824288209289077e-07, | |
| "logits/chosen": -0.9843920469284058, | |
| "logits/rejected": -0.9963368773460388, | |
| "logps/chosen": -695.6856689453125, | |
| "logps/rejected": -835.2234497070312, | |
| "loss": 0.3754, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.025155067443848, | |
| "rewards/margins": 1.4948614835739136, | |
| "rewards/rejected": -5.520016193389893, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.7074539139727491, | |
| "grad_norm": 18.164643888570296, | |
| "learning_rate": 1.660106805663233e-07, | |
| "logits/chosen": -0.9996950626373291, | |
| "logits/rejected": -1.0081021785736084, | |
| "logps/chosen": -585.10009765625, | |
| "logps/rejected": -739.9697265625, | |
| "loss": 0.3477, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.4289603233337402, | |
| "rewards/margins": 1.5738253593444824, | |
| "rewards/rejected": -5.002785682678223, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.7095912369756879, | |
| "grad_norm": 18.104299917352034, | |
| "learning_rate": 1.6378877321963224e-07, | |
| "logits/chosen": -1.0804736614227295, | |
| "logits/rejected": -1.0568981170654297, | |
| "logps/chosen": -713.994140625, | |
| "logps/rejected": -918.4098510742188, | |
| "loss": 0.3097, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.44266414642334, | |
| "rewards/margins": 1.9913928508758545, | |
| "rewards/rejected": -6.434057712554932, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.7117285599786267, | |
| "grad_norm": 19.250777560445798, | |
| "learning_rate": 1.6157728436826125e-07, | |
| "logits/chosen": -1.0584830045700073, | |
| "logits/rejected": -1.0436077117919922, | |
| "logps/chosen": -645.3274536132812, | |
| "logps/rejected": -729.302001953125, | |
| "loss": 0.3813, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.107142925262451, | |
| "rewards/margins": 0.8789306879043579, | |
| "rewards/rejected": -4.9860734939575195, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.7138658829815656, | |
| "grad_norm": 18.363434689830413, | |
| "learning_rate": 1.5937633774474053e-07, | |
| "logits/chosen": -0.964330792427063, | |
| "logits/rejected": -1.0107676982879639, | |
| "logps/chosen": -681.0447387695312, | |
| "logps/rejected": -893.3312377929688, | |
| "loss": 0.4289, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.495457172393799, | |
| "rewards/margins": 2.035775661468506, | |
| "rewards/rejected": -6.531231880187988, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.7160032059845044, | |
| "grad_norm": 19.36554980671261, | |
| "learning_rate": 1.5718605649176415e-07, | |
| "logits/chosen": -1.08909273147583, | |
| "logits/rejected": -1.029083013534546, | |
| "logps/chosen": -606.0968017578125, | |
| "logps/rejected": -801.4658203125, | |
| "loss": 0.3958, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.555162191390991, | |
| "rewards/margins": 1.8386735916137695, | |
| "rewards/rejected": -5.393835544586182, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.7181405289874432, | |
| "grad_norm": 19.53200074466149, | |
| "learning_rate": 1.5500656315529971e-07, | |
| "logits/chosen": -0.9612129330635071, | |
| "logits/rejected": -1.0220246315002441, | |
| "logps/chosen": -689.0376586914062, | |
| "logps/rejected": -871.2042846679688, | |
| "loss": 0.4066, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.374889850616455, | |
| "rewards/margins": 1.9528312683105469, | |
| "rewards/rejected": -6.327720642089844, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.7202778519903821, | |
| "grad_norm": 16.525872426498882, | |
| "learning_rate": 1.5283797967773228e-07, | |
| "logits/chosen": -0.9871097803115845, | |
| "logits/rejected": -1.0105241537094116, | |
| "logps/chosen": -629.0919799804688, | |
| "logps/rejected": -870.4136352539062, | |
| "loss": 0.3259, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.831345558166504, | |
| "rewards/margins": 2.3878016471862793, | |
| "rewards/rejected": -6.219147205352783, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.7224151749933209, | |
| "grad_norm": 17.894860864055556, | |
| "learning_rate": 1.506804273910418e-07, | |
| "logits/chosen": -1.0389041900634766, | |
| "logits/rejected": -0.9967656135559082, | |
| "logps/chosen": -597.2425537109375, | |
| "logps/rejected": -690.931884765625, | |
| "loss": 0.3708, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.6145212650299072, | |
| "rewards/margins": 1.04227876663208, | |
| "rewards/rejected": -4.656799793243408, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.7245524979962596, | |
| "grad_norm": 24.07466390909236, | |
| "learning_rate": 1.485340270100143e-07, | |
| "logits/chosen": -0.8043147921562195, | |
| "logits/rejected": -0.8566476106643677, | |
| "logps/chosen": -721.7520751953125, | |
| "logps/rejected": -766.68896484375, | |
| "loss": 0.384, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -4.88438081741333, | |
| "rewards/margins": 0.5339137315750122, | |
| "rewards/rejected": -5.418294906616211, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.7266898209991985, | |
| "grad_norm": 23.369533031738897, | |
| "learning_rate": 1.4639889862548805e-07, | |
| "logits/chosen": -1.1091313362121582, | |
| "logits/rejected": -1.1004701852798462, | |
| "logps/chosen": -672.283447265625, | |
| "logps/rejected": -803.2059936523438, | |
| "loss": 0.4195, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.266483306884766, | |
| "rewards/margins": 1.2370930910110474, | |
| "rewards/rejected": -5.503575801849365, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.7288271440021373, | |
| "grad_norm": 18.450003910880596, | |
| "learning_rate": 1.4427516169763443e-07, | |
| "logits/chosen": -0.9054552912712097, | |
| "logits/rejected": -0.9048701524734497, | |
| "logps/chosen": -747.39013671875, | |
| "logps/rejected": -927.0717163085938, | |
| "loss": 0.3928, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -5.073859214782715, | |
| "rewards/margins": 1.7517752647399902, | |
| "rewards/rejected": -6.825634479522705, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.7309644670050761, | |
| "grad_norm": 23.60596160513393, | |
| "learning_rate": 1.4216293504927448e-07, | |
| "logits/chosen": -1.067272424697876, | |
| "logits/rejected": -1.0798040628433228, | |
| "logps/chosen": -676.371826171875, | |
| "logps/rejected": -869.149169921875, | |
| "loss": 0.379, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -4.598742961883545, | |
| "rewards/margins": 1.6217763423919678, | |
| "rewards/rejected": -6.220519065856934, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.733101790008015, | |
| "grad_norm": 21.614156690463627, | |
| "learning_rate": 1.400623368592302e-07, | |
| "logits/chosen": -0.9705036878585815, | |
| "logits/rejected": -0.9520074725151062, | |
| "logps/chosen": -678.7489013671875, | |
| "logps/rejected": -816.748779296875, | |
| "loss": 0.4281, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.580175876617432, | |
| "rewards/margins": 1.3712455034255981, | |
| "rewards/rejected": -5.951421737670898, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.7352391130109538, | |
| "grad_norm": 23.05746652103131, | |
| "learning_rate": 1.3797348465571306e-07, | |
| "logits/chosen": -0.9307321310043335, | |
| "logits/rejected": -0.9591343402862549, | |
| "logps/chosen": -746.435546875, | |
| "logps/rejected": -1008.3950805664062, | |
| "loss": 0.3567, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -4.570581912994385, | |
| "rewards/margins": 2.517709255218506, | |
| "rewards/rejected": -7.088291168212891, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.7373764360138926, | |
| "grad_norm": 18.10148284802051, | |
| "learning_rate": 1.3589649530974773e-07, | |
| "logits/chosen": -1.0869290828704834, | |
| "logits/rejected": -1.0850858688354492, | |
| "logps/chosen": -658.4027099609375, | |
| "logps/rejected": -982.6063842773438, | |
| "loss": 0.3781, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.2937188148498535, | |
| "rewards/margins": 2.9649014472961426, | |
| "rewards/rejected": -7.258620262145996, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.7395137590168315, | |
| "grad_norm": 23.41128054527833, | |
| "learning_rate": 1.3383148502863378e-07, | |
| "logits/chosen": -1.0312128067016602, | |
| "logits/rejected": -1.1049691438674927, | |
| "logps/chosen": -814.4356689453125, | |
| "logps/rejected": -798.0429077148438, | |
| "loss": 0.514, | |
| "rewards/accuracies": 0.4375, | |
| "rewards/chosen": -5.450601577758789, | |
| "rewards/margins": 0.045907363295555115, | |
| "rewards/rejected": -5.496509552001953, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.7416510820197703, | |
| "grad_norm": 18.50899288865318, | |
| "learning_rate": 1.317785693494433e-07, | |
| "logits/chosen": -1.0586178302764893, | |
| "logits/rejected": -1.029137134552002, | |
| "logps/chosen": -691.7874755859375, | |
| "logps/rejected": -813.152099609375, | |
| "loss": 0.3592, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.9073405265808105, | |
| "rewards/margins": 1.246219515800476, | |
| "rewards/rejected": -5.153559684753418, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.743788405022709, | |
| "grad_norm": 20.32668011498763, | |
| "learning_rate": 1.2973786313255695e-07, | |
| "logits/chosen": -0.8740507960319519, | |
| "logits/rejected": -0.9160584211349487, | |
| "logps/chosen": -662.5859375, | |
| "logps/rejected": -873.8433227539062, | |
| "loss": 0.4097, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.367182731628418, | |
| "rewards/margins": 2.035503625869751, | |
| "rewards/rejected": -6.4026875495910645, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.7459257280256478, | |
| "grad_norm": 23.02424339442235, | |
| "learning_rate": 1.2770948055523742e-07, | |
| "logits/chosen": -1.0587798357009888, | |
| "logits/rejected": -1.0669668912887573, | |
| "logps/chosen": -809.7626953125, | |
| "logps/rejected": -983.58984375, | |
| "loss": 0.3906, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -5.2913947105407715, | |
| "rewards/margins": 1.9067097902297974, | |
| "rewards/rejected": -7.1981048583984375, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.7480630510285867, | |
| "grad_norm": 25.77622479036058, | |
| "learning_rate": 1.2569353510524156e-07, | |
| "logits/chosen": -1.069049596786499, | |
| "logits/rejected": -1.1337969303131104, | |
| "logps/chosen": -720.7489013671875, | |
| "logps/rejected": -938.7985229492188, | |
| "loss": 0.4103, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.965020179748535, | |
| "rewards/margins": 2.0392074584960938, | |
| "rewards/rejected": -7.004227638244629, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.7502003740315255, | |
| "grad_norm": 17.60740131186996, | |
| "learning_rate": 1.2369013957447015e-07, | |
| "logits/chosen": -1.0574181079864502, | |
| "logits/rejected": -1.0570802688598633, | |
| "logps/chosen": -694.0639038085938, | |
| "logps/rejected": -829.4335327148438, | |
| "loss": 0.3407, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.506197452545166, | |
| "rewards/margins": 1.4661719799041748, | |
| "rewards/rejected": -5.97236967086792, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.7523376970344643, | |
| "grad_norm": 17.276020504905794, | |
| "learning_rate": 1.216994060526577e-07, | |
| "logits/chosen": -1.1085654497146606, | |
| "logits/rejected": -1.0444343090057373, | |
| "logps/chosen": -699.8790893554688, | |
| "logps/rejected": -857.7581176757812, | |
| "loss": 0.3333, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.61436128616333, | |
| "rewards/margins": 1.533206582069397, | |
| "rewards/rejected": -6.147567272186279, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.7544750200374032, | |
| "grad_norm": 20.20530099751691, | |
| "learning_rate": 1.197214459211006e-07, | |
| "logits/chosen": -1.028286099433899, | |
| "logits/rejected": -1.0080163478851318, | |
| "logps/chosen": -729.8060913085938, | |
| "logps/rejected": -908.0136108398438, | |
| "loss": 0.3995, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -4.445433616638184, | |
| "rewards/margins": 1.915571689605713, | |
| "rewards/rejected": -6.361005783081055, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.756612343040342, | |
| "grad_norm": 21.65892045821199, | |
| "learning_rate": 1.1775636984642612e-07, | |
| "logits/chosen": -1.001996636390686, | |
| "logits/rejected": -0.9989675283432007, | |
| "logps/chosen": -658.6278076171875, | |
| "logps/rejected": -824.7811279296875, | |
| "loss": 0.3989, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.4795989990234375, | |
| "rewards/margins": 1.533904790878296, | |
| "rewards/rejected": -6.0135040283203125, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.7587496660432808, | |
| "grad_norm": 18.745090346580444, | |
| "learning_rate": 1.1580428777439973e-07, | |
| "logits/chosen": -1.1561853885650635, | |
| "logits/rejected": -1.0848287343978882, | |
| "logps/chosen": -622.0913696289062, | |
| "logps/rejected": -787.5831298828125, | |
| "loss": 0.3734, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.140936851501465, | |
| "rewards/margins": 1.524781346321106, | |
| "rewards/rejected": -5.665718078613281, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.7608869890462197, | |
| "grad_norm": 18.598176758293135, | |
| "learning_rate": 1.1386530892377399e-07, | |
| "logits/chosen": -0.9880399107933044, | |
| "logits/rejected": -0.9888783097267151, | |
| "logps/chosen": -613.0370483398438, | |
| "logps/rejected": -777.3377075195312, | |
| "loss": 0.3762, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.9182024002075195, | |
| "rewards/margins": 1.6418043375015259, | |
| "rewards/rejected": -5.560007095336914, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.7630243120491584, | |
| "grad_norm": 20.075874653172413, | |
| "learning_rate": 1.1193954178017814e-07, | |
| "logits/chosen": -0.9616910219192505, | |
| "logits/rejected": -1.0504767894744873, | |
| "logps/chosen": -782.3270874023438, | |
| "logps/rejected": -1025.231689453125, | |
| "loss": 0.4392, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.93320894241333, | |
| "rewards/margins": 2.5418636798858643, | |
| "rewards/rejected": -7.475072383880615, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.7651616350520972, | |
| "grad_norm": 18.869618371734305, | |
| "learning_rate": 1.1002709409004803e-07, | |
| "logits/chosen": -1.1053738594055176, | |
| "logits/rejected": -1.0864824056625366, | |
| "logps/chosen": -768.8507080078125, | |
| "logps/rejected": -939.0321044921875, | |
| "loss": 0.3226, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -5.1586713790893555, | |
| "rewards/margins": 1.6820204257965088, | |
| "rewards/rejected": -6.840692043304443, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.7672989580550361, | |
| "grad_norm": 22.39575999598468, | |
| "learning_rate": 1.0812807285459737e-07, | |
| "logits/chosen": -0.8523386120796204, | |
| "logits/rejected": -0.8684836030006409, | |
| "logps/chosen": -634.4469604492188, | |
| "logps/rejected": -910.783447265625, | |
| "loss": 0.4391, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.161304950714111, | |
| "rewards/margins": 2.574713945388794, | |
| "rewards/rejected": -6.736019134521484, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.7694362810579749, | |
| "grad_norm": 24.655379460289804, | |
| "learning_rate": 1.0624258432383165e-07, | |
| "logits/chosen": -0.9855338931083679, | |
| "logits/rejected": -0.9286175966262817, | |
| "logps/chosen": -821.5938110351562, | |
| "logps/rejected": -976.9266357421875, | |
| "loss": 0.4425, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -5.609586715698242, | |
| "rewards/margins": 1.46958327293396, | |
| "rewards/rejected": -7.0791707038879395, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.7715736040609137, | |
| "grad_norm": 23.07044684161756, | |
| "learning_rate": 1.0437073399060298e-07, | |
| "logits/chosen": -0.9206562638282776, | |
| "logits/rejected": -0.8095078468322754, | |
| "logps/chosen": -618.7409057617188, | |
| "logps/rejected": -815.9017333984375, | |
| "loss": 0.396, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.211564064025879, | |
| "rewards/margins": 1.776960849761963, | |
| "rewards/rejected": -5.988524913787842, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.7737109270638525, | |
| "grad_norm": 20.255683251830362, | |
| "learning_rate": 1.0251262658470838e-07, | |
| "logits/chosen": -1.0115031003952026, | |
| "logits/rejected": -0.9697583317756653, | |
| "logps/chosen": -671.9834594726562, | |
| "logps/rejected": -856.5125122070312, | |
| "loss": 0.4114, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.316826343536377, | |
| "rewards/margins": 1.731043815612793, | |
| "rewards/rejected": -6.04787015914917, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.7758482500667914, | |
| "grad_norm": 36.27259699112537, | |
| "learning_rate": 1.0066836606702937e-07, | |
| "logits/chosen": -0.9073708057403564, | |
| "logits/rejected": -1.013731598854065, | |
| "logps/chosen": -687.530029296875, | |
| "logps/rejected": -816.8972778320312, | |
| "loss": 0.5571, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.503976821899414, | |
| "rewards/margins": 1.4232851266860962, | |
| "rewards/rejected": -5.927262306213379, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.7779855730697302, | |
| "grad_norm": 18.487130214617952, | |
| "learning_rate": 9.883805562371581e-08, | |
| "logits/chosen": -1.096664547920227, | |
| "logits/rejected": -1.151119351387024, | |
| "logps/chosen": -627.2655029296875, | |
| "logps/rejected": -766.6741943359375, | |
| "loss": 0.3511, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.355429649353027, | |
| "rewards/margins": 1.6355644464492798, | |
| "rewards/rejected": -5.990993976593018, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.7801228960726689, | |
| "grad_norm": 17.467778162939954, | |
| "learning_rate": 9.702179766041299e-08, | |
| "logits/chosen": -1.1129746437072754, | |
| "logits/rejected": -1.0882641077041626, | |
| "logps/chosen": -609.0090942382812, | |
| "logps/rejected": -835.59375, | |
| "loss": 0.3284, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.666825294494629, | |
| "rewards/margins": 2.0798864364624023, | |
| "rewards/rejected": -5.746711730957031, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.7822602190756078, | |
| "grad_norm": 26.497135911369625, | |
| "learning_rate": 9.521969379653121e-08, | |
| "logits/chosen": -1.0410820245742798, | |
| "logits/rejected": -1.075594425201416, | |
| "logps/chosen": -861.631591796875, | |
| "logps/rejected": -1022.0664672851562, | |
| "loss": 0.4385, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -5.093762397766113, | |
| "rewards/margins": 1.7978633642196655, | |
| "rewards/rejected": -6.891625881195068, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.7843975420785466, | |
| "grad_norm": 19.54124775721202, | |
| "learning_rate": 9.343184485956086e-08, | |
| "logits/chosen": -1.0408751964569092, | |
| "logits/rejected": -1.1065208911895752, | |
| "logps/chosen": -597.0819702148438, | |
| "logps/rejected": -755.0318603515625, | |
| "loss": 0.3985, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.7190897464752197, | |
| "rewards/margins": 1.8094736337661743, | |
| "rewards/rejected": -5.528563499450684, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.7865348650814854, | |
| "grad_norm": 23.844802367349075, | |
| "learning_rate": 9.16583508794308e-08, | |
| "logits/chosen": -0.9044683575630188, | |
| "logits/rejected": -0.8596857786178589, | |
| "logps/chosen": -730.1447143554688, | |
| "logps/rejected": -878.2034912109375, | |
| "loss": 0.3, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.852618217468262, | |
| "rewards/margins": 1.4661897420883179, | |
| "rewards/rejected": -6.318808078765869, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.7886721880844243, | |
| "grad_norm": 20.562190767658738, | |
| "learning_rate": 8.989931108291197e-08, | |
| "logits/chosen": -0.9048526287078857, | |
| "logits/rejected": -0.9187178611755371, | |
| "logps/chosen": -699.8194580078125, | |
| "logps/rejected": -844.4905395507812, | |
| "loss": 0.3635, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -4.6599884033203125, | |
| "rewards/margins": 1.4540899991989136, | |
| "rewards/rejected": -6.114078521728516, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.7908095110873631, | |
| "grad_norm": 20.73718481631679, | |
| "learning_rate": 8.815482388806554e-08, | |
| "logits/chosen": -1.0599019527435303, | |
| "logits/rejected": -1.065873622894287, | |
| "logps/chosen": -760.4301147460938, | |
| "logps/rejected": -946.1746215820312, | |
| "loss": 0.4287, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -5.120336532592773, | |
| "rewards/margins": 1.838600993156433, | |
| "rewards/rejected": -6.958937168121338, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.7929468340903019, | |
| "grad_norm": 30.164533991222857, | |
| "learning_rate": 8.642498689873619e-08, | |
| "logits/chosen": -1.0670329332351685, | |
| "logits/rejected": -1.1095917224884033, | |
| "logps/chosen": -768.016357421875, | |
| "logps/rejected": -985.7068481445312, | |
| "loss": 0.4338, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -4.948676586151123, | |
| "rewards/margins": 2.0501699447631836, | |
| "rewards/rejected": -6.998846530914307, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.7950841570932408, | |
| "grad_norm": 18.533158040155175, | |
| "learning_rate": 8.47098968990914e-08, | |
| "logits/chosen": -1.212889313697815, | |
| "logits/rejected": -1.174148678779602, | |
| "logps/chosen": -575.9932250976562, | |
| "logps/rejected": -694.8380126953125, | |
| "loss": 0.3543, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.408289670944214, | |
| "rewards/margins": 1.2301357984542847, | |
| "rewards/rejected": -4.638425350189209, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.7972214800961795, | |
| "grad_norm": 19.740449187608878, | |
| "learning_rate": 8.300964984820672e-08, | |
| "logits/chosen": -0.9624559879302979, | |
| "logits/rejected": -0.9103982448577881, | |
| "logps/chosen": -758.9552001953125, | |
| "logps/rejected": -952.0904541015625, | |
| "loss": 0.385, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.812871932983398, | |
| "rewards/margins": 1.8450303077697754, | |
| "rewards/rejected": -6.657902240753174, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.7993588030991183, | |
| "grad_norm": 17.857229527882023, | |
| "learning_rate": 8.13243408746961e-08, | |
| "logits/chosen": -1.044665813446045, | |
| "logits/rejected": -1.0194975137710571, | |
| "logps/chosen": -696.4033813476562, | |
| "logps/rejected": -845.8680419921875, | |
| "loss": 0.3787, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -4.486112594604492, | |
| "rewards/margins": 1.4695513248443604, | |
| "rewards/rejected": -5.955664157867432, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.8014961261020572, | |
| "grad_norm": 27.402209380453634, | |
| "learning_rate": 7.96540642713902e-08, | |
| "logits/chosen": -1.0042835474014282, | |
| "logits/rejected": -0.9937084317207336, | |
| "logps/chosen": -798.3321533203125, | |
| "logps/rejected": -931.4829711914062, | |
| "loss": 0.4162, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -5.05294132232666, | |
| "rewards/margins": 1.302819848060608, | |
| "rewards/rejected": -6.35576057434082, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.803633449104996, | |
| "grad_norm": 28.3566059208529, | |
| "learning_rate": 7.799891349006019e-08, | |
| "logits/chosen": -1.1246654987335205, | |
| "logits/rejected": -1.1276860237121582, | |
| "logps/chosen": -656.9721069335938, | |
| "logps/rejected": -813.6206665039062, | |
| "loss": 0.4719, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -4.687532901763916, | |
| "rewards/margins": 1.3982094526290894, | |
| "rewards/rejected": -6.085742473602295, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.8057707721079348, | |
| "grad_norm": 18.856881189055258, | |
| "learning_rate": 7.635898113618957e-08, | |
| "logits/chosen": -1.2034432888031006, | |
| "logits/rejected": -1.1600438356399536, | |
| "logps/chosen": -585.2130737304688, | |
| "logps/rejected": -785.3953247070312, | |
| "loss": 0.3801, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.6061196327209473, | |
| "rewards/margins": 1.8955157995224, | |
| "rewards/rejected": -5.501636028289795, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.8079080951108736, | |
| "grad_norm": 21.136594852281714, | |
| "learning_rate": 7.473435896379258e-08, | |
| "logits/chosen": -1.1049622297286987, | |
| "logits/rejected": -1.025770902633667, | |
| "logps/chosen": -697.8673706054688, | |
| "logps/rejected": -841.3628540039062, | |
| "loss": 0.4047, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.117053985595703, | |
| "rewards/margins": 1.336838960647583, | |
| "rewards/rejected": -5.453892707824707, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.8100454181138125, | |
| "grad_norm": 21.622158153196573, | |
| "learning_rate": 7.312513787028065e-08, | |
| "logits/chosen": -1.018204927444458, | |
| "logits/rejected": -1.0209660530090332, | |
| "logps/chosen": -551.69091796875, | |
| "logps/rejected": -765.5928344726562, | |
| "loss": 0.3705, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -3.391167402267456, | |
| "rewards/margins": 1.9136013984680176, | |
| "rewards/rejected": -5.304769039154053, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.8121827411167513, | |
| "grad_norm": 25.692774884542565, | |
| "learning_rate": 7.153140789137667e-08, | |
| "logits/chosen": -1.1156607866287231, | |
| "logits/rejected": -1.1090834140777588, | |
| "logps/chosen": -587.258056640625, | |
| "logps/rejected": -709.8887329101562, | |
| "loss": 0.417, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -3.669807195663452, | |
| "rewards/margins": 1.1046210527420044, | |
| "rewards/rejected": -4.774428367614746, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.81432006411969, | |
| "grad_norm": 22.676599376198244, | |
| "learning_rate": 6.995325819607783e-08, | |
| "logits/chosen": -0.8035112619400024, | |
| "logits/rejected": -0.7905966639518738, | |
| "logps/chosen": -724.717529296875, | |
| "logps/rejected": -951.2332153320312, | |
| "loss": 0.3864, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.354774475097656, | |
| "rewards/margins": 2.2559542655944824, | |
| "rewards/rejected": -6.6107282638549805, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.8164573871226289, | |
| "grad_norm": 20.060527238580573, | |
| "learning_rate": 6.839077708166608e-08, | |
| "logits/chosen": -0.9624785780906677, | |
| "logits/rejected": -0.9216483235359192, | |
| "logps/chosen": -759.6564331054688, | |
| "logps/rejected": -846.4254760742188, | |
| "loss": 0.4242, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.706582546234131, | |
| "rewards/margins": 0.9559379816055298, | |
| "rewards/rejected": -5.662520408630371, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.8185947101255677, | |
| "grad_norm": 20.450745536438863, | |
| "learning_rate": 6.684405196876843e-08, | |
| "logits/chosen": -0.9839971661567688, | |
| "logits/rejected": -0.9619154930114746, | |
| "logps/chosen": -659.962890625, | |
| "logps/rejected": -834.504638671875, | |
| "loss": 0.4453, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -4.435713768005371, | |
| "rewards/margins": 1.5213212966918945, | |
| "rewards/rejected": -5.957034587860107, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.8207320331285065, | |
| "grad_norm": 21.923861116889775, | |
| "learning_rate": 6.531316939646522e-08, | |
| "logits/chosen": -1.0496439933776855, | |
| "logits/rejected": -1.0298651456832886, | |
| "logps/chosen": -652.1452026367188, | |
| "logps/rejected": -761.5989990234375, | |
| "loss": 0.4117, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.186628341674805, | |
| "rewards/margins": 1.098493218421936, | |
| "rewards/rejected": -5.285121917724609, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.8228693561314454, | |
| "grad_norm": 15.872158871462359, | |
| "learning_rate": 6.379821501744902e-08, | |
| "logits/chosen": -0.9291130900382996, | |
| "logits/rejected": -0.9559027552604675, | |
| "logps/chosen": -637.105712890625, | |
| "logps/rejected": -784.9781494140625, | |
| "loss": 0.3618, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.092860698699951, | |
| "rewards/margins": 1.487864375114441, | |
| "rewards/rejected": -5.580724716186523, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.8250066791343842, | |
| "grad_norm": 21.170940984738564, | |
| "learning_rate": 6.229927359323156e-08, | |
| "logits/chosen": -1.013567566871643, | |
| "logits/rejected": -0.9818376302719116, | |
| "logps/chosen": -581.1494140625, | |
| "logps/rejected": -750.385009765625, | |
| "loss": 0.3848, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -3.5493507385253906, | |
| "rewards/margins": 1.4864931106567383, | |
| "rewards/rejected": -5.035844326019287, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.827144002137323, | |
| "grad_norm": 20.203562100764046, | |
| "learning_rate": 6.081642898940186e-08, | |
| "logits/chosen": -0.9751057624816895, | |
| "logits/rejected": -0.9479618072509766, | |
| "logps/chosen": -535.7874145507812, | |
| "logps/rejected": -717.0790405273438, | |
| "loss": 0.3476, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -3.4093728065490723, | |
| "rewards/margins": 1.679715871810913, | |
| "rewards/rejected": -5.089088439941406, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.8292813251402619, | |
| "grad_norm": 16.226862763327308, | |
| "learning_rate": 5.9349764170933674e-08, | |
| "logits/chosen": -0.9501919746398926, | |
| "logits/rejected": -0.9817054867744446, | |
| "logps/chosen": -545.6821899414062, | |
| "logps/rejected": -637.3515625, | |
| "loss": 0.4066, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -3.451007843017578, | |
| "rewards/margins": 1.0472853183746338, | |
| "rewards/rejected": -4.498293399810791, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.8314186481432007, | |
| "grad_norm": 21.301359722964037, | |
| "learning_rate": 5.789936119754407e-08, | |
| "logits/chosen": -1.070288062095642, | |
| "logits/rejected": -1.03058660030365, | |
| "logps/chosen": -773.905517578125, | |
| "logps/rejected": -884.7736206054688, | |
| "loss": 0.3989, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.410757064819336, | |
| "rewards/margins": 1.2033069133758545, | |
| "rewards/rejected": -5.6140642166137695, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.8335559711461394, | |
| "grad_norm": 16.155956822127486, | |
| "learning_rate": 5.646530121910162e-08, | |
| "logits/chosen": -0.9560039043426514, | |
| "logits/rejected": -0.9677013158798218, | |
| "logps/chosen": -555.0410766601562, | |
| "logps/rejected": -647.4817504882812, | |
| "loss": 0.3702, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.367013931274414, | |
| "rewards/margins": 1.1322979927062988, | |
| "rewards/rejected": -4.499312400817871, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.8356932941490782, | |
| "grad_norm": 17.445283127579135, | |
| "learning_rate": 5.5047664471086284e-08, | |
| "logits/chosen": -1.0367249250411987, | |
| "logits/rejected": -1.0191857814788818, | |
| "logps/chosen": -623.27978515625, | |
| "logps/rejected": -845.25048828125, | |
| "loss": 0.354, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.8465986251831055, | |
| "rewards/margins": 2.2664730548858643, | |
| "rewards/rejected": -6.113072395324707, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.8378306171520171, | |
| "grad_norm": 17.389408198261865, | |
| "learning_rate": 5.3646530270100556e-08, | |
| "logits/chosen": -1.1824917793273926, | |
| "logits/rejected": -1.1786624193191528, | |
| "logps/chosen": -629.4869384765625, | |
| "logps/rejected": -910.8460693359375, | |
| "loss": 0.3307, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.6656405925750732, | |
| "rewards/margins": 2.664851188659668, | |
| "rewards/rejected": -6.33049201965332, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.8399679401549559, | |
| "grad_norm": 16.107124952267284, | |
| "learning_rate": 5.2261977009431193e-08, | |
| "logits/chosen": -1.2528231143951416, | |
| "logits/rejected": -1.2414242029190063, | |
| "logps/chosen": -679.169189453125, | |
| "logps/rejected": -904.231689453125, | |
| "loss": 0.3437, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -3.949631690979004, | |
| "rewards/margins": 1.9819270372390747, | |
| "rewards/rejected": -5.931559085845947, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.8421052631578947, | |
| "grad_norm": 14.65787560767015, | |
| "learning_rate": 5.089408215466349e-08, | |
| "logits/chosen": -1.1920839548110962, | |
| "logits/rejected": -1.1330435276031494, | |
| "logps/chosen": -591.995849609375, | |
| "logps/rejected": -754.32421875, | |
| "loss": 0.3777, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -3.461061954498291, | |
| "rewards/margins": 1.4271600246429443, | |
| "rewards/rejected": -4.8882222175598145, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.8442425861608336, | |
| "grad_norm": 19.973781830012914, | |
| "learning_rate": 4.9542922239346865e-08, | |
| "logits/chosen": -0.9532175660133362, | |
| "logits/rejected": -0.9305418729782104, | |
| "logps/chosen": -700.0613403320312, | |
| "logps/rejected": -854.81201171875, | |
| "loss": 0.4104, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.186321258544922, | |
| "rewards/margins": 1.5199509859085083, | |
| "rewards/rejected": -5.706272602081299, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.8463799091637724, | |
| "grad_norm": 26.08541321380337, | |
| "learning_rate": 4.8208572860713086e-08, | |
| "logits/chosen": -1.149269461631775, | |
| "logits/rejected": -1.1293056011199951, | |
| "logps/chosen": -618.3247680664062, | |
| "logps/rejected": -777.4251708984375, | |
| "loss": 0.4851, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.4766008853912354, | |
| "rewards/margins": 1.5825449228286743, | |
| "rewards/rejected": -5.059146404266357, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.8485172321667112, | |
| "grad_norm": 29.339785563765055, | |
| "learning_rate": 4.689110867544645e-08, | |
| "logits/chosen": -1.015437126159668, | |
| "logits/rejected": -0.9925183057785034, | |
| "logps/chosen": -679.9268188476562, | |
| "logps/rejected": -870.37890625, | |
| "loss": 0.344, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.385658264160156, | |
| "rewards/margins": 1.9707388877868652, | |
| "rewards/rejected": -6.35639762878418, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.85065455516965, | |
| "grad_norm": 19.625508223329977, | |
| "learning_rate": 4.5590603395506626e-08, | |
| "logits/chosen": -1.0667216777801514, | |
| "logits/rejected": -1.049599289894104, | |
| "logps/chosen": -589.4276123046875, | |
| "logps/rejected": -712.3489379882812, | |
| "loss": 0.4014, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.779334306716919, | |
| "rewards/margins": 1.0566929578781128, | |
| "rewards/rejected": -4.836027145385742, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.8527918781725888, | |
| "grad_norm": 19.959692153012657, | |
| "learning_rate": 4.430712978400462e-08, | |
| "logits/chosen": -1.0223028659820557, | |
| "logits/rejected": -1.0172353982925415, | |
| "logps/chosen": -494.1994323730469, | |
| "logps/rejected": -618.9550170898438, | |
| "loss": 0.4284, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -3.038055658340454, | |
| "rewards/margins": 1.1104681491851807, | |
| "rewards/rejected": -4.148523807525635, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.8549292011755276, | |
| "grad_norm": 22.852143621544176, | |
| "learning_rate": 4.304075965113171e-08, | |
| "logits/chosen": -1.2095826864242554, | |
| "logits/rejected": -1.1936359405517578, | |
| "logps/chosen": -572.5497436523438, | |
| "logps/rejected": -702.556396484375, | |
| "loss": 0.3827, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -3.307962656021118, | |
| "rewards/margins": 1.31619131565094, | |
| "rewards/rejected": -4.624154090881348, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.8549292011755276, | |
| "eval_logits/chosen": -1.128338098526001, | |
| "eval_logits/rejected": -1.1495376825332642, | |
| "eval_logps/chosen": -710.064697265625, | |
| "eval_logps/rejected": -863.7308349609375, | |
| "eval_loss": 0.3871653079986572, | |
| "eval_rewards/accuracies": 0.8709677457809448, | |
| "eval_rewards/chosen": -4.3693461418151855, | |
| "eval_rewards/margins": 1.5976247787475586, | |
| "eval_rewards/rejected": -5.966970920562744, | |
| "eval_runtime": 127.1231, | |
| "eval_samples_per_second": 15.426, | |
| "eval_steps_per_second": 0.488, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.8570665241784665, | |
| "grad_norm": 20.918890009764137, | |
| "learning_rate": 4.1791563850141695e-08, | |
| "logits/chosen": -0.9988707900047302, | |
| "logits/rejected": -1.0228999853134155, | |
| "logps/chosen": -705.0687255859375, | |
| "logps/rejected": -993.6966552734375, | |
| "loss": 0.3109, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -4.109606742858887, | |
| "rewards/margins": 2.890212297439575, | |
| "rewards/rejected": -6.999818801879883, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 0.8592038471814053, | |
| "grad_norm": 22.50167745931539, | |
| "learning_rate": 4.0559612273386616e-08, | |
| "logits/chosen": -1.098199725151062, | |
| "logits/rejected": -1.1250154972076416, | |
| "logps/chosen": -785.855224609375, | |
| "logps/rejected": -917.9451904296875, | |
| "loss": 0.3891, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.712646484375, | |
| "rewards/margins": 1.341720700263977, | |
| "rewards/rejected": -6.0543670654296875, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.8613411701843441, | |
| "grad_norm": 16.355069952426618, | |
| "learning_rate": 3.934497384840607e-08, | |
| "logits/chosen": -0.9095240235328674, | |
| "logits/rejected": -0.8507342338562012, | |
| "logps/chosen": -621.5723876953125, | |
| "logps/rejected": -791.5116577148438, | |
| "loss": 0.4111, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -3.8870487213134766, | |
| "rewards/margins": 1.6334927082061768, | |
| "rewards/rejected": -5.520541667938232, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 0.863478493187283, | |
| "grad_norm": 20.03294448892495, | |
| "learning_rate": 3.8147716534071274e-08, | |
| "logits/chosen": -1.0887154340744019, | |
| "logits/rejected": -1.095017671585083, | |
| "logps/chosen": -605.3096923828125, | |
| "logps/rejected": -669.5009155273438, | |
| "loss": 0.4435, | |
| "rewards/accuracies": 0.5625, | |
| "rewards/chosen": -3.492586374282837, | |
| "rewards/margins": 0.6767375469207764, | |
| "rewards/rejected": -4.169323921203613, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.8656158161902218, | |
| "grad_norm": 19.638131692176458, | |
| "learning_rate": 3.696790731678217e-08, | |
| "logits/chosen": -1.0262949466705322, | |
| "logits/rejected": -1.0150538682937622, | |
| "logps/chosen": -797.56689453125, | |
| "logps/rejected": -929.2540893554688, | |
| "loss": 0.4095, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.939427852630615, | |
| "rewards/margins": 1.520028829574585, | |
| "rewards/rejected": -6.459456920623779, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.8677531391931605, | |
| "grad_norm": 19.64316736105873, | |
| "learning_rate": 3.5805612206719885e-08, | |
| "logits/chosen": -0.9092820882797241, | |
| "logits/rejected": -0.9483243227005005, | |
| "logps/chosen": -681.32958984375, | |
| "logps/rejected": -848.2050170898438, | |
| "loss": 0.349, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -4.447098255157471, | |
| "rewards/margins": 1.769409418106079, | |
| "rewards/rejected": -6.216507911682129, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 0.8698904621960993, | |
| "grad_norm": 16.2039010339986, | |
| "learning_rate": 3.466089623415333e-08, | |
| "logits/chosen": -0.9941847324371338, | |
| "logits/rejected": -0.9800211787223816, | |
| "logps/chosen": -583.2901000976562, | |
| "logps/rejected": -831.3629150390625, | |
| "loss": 0.3707, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -3.81510329246521, | |
| "rewards/margins": 2.43589448928833, | |
| "rewards/rejected": -6.250997543334961, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 0.8720277851990382, | |
| "grad_norm": 17.593791491624575, | |
| "learning_rate": 3.3533823445800965e-08, | |
| "logits/chosen": -1.0939316749572754, | |
| "logits/rejected": -1.0538439750671387, | |
| "logps/chosen": -563.0457763671875, | |
| "logps/rejected": -702.2050170898438, | |
| "loss": 0.3908, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.7340145111083984, | |
| "rewards/margins": 1.5319364070892334, | |
| "rewards/rejected": -5.2659502029418945, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.874165108201977, | |
| "grad_norm": 17.753144065442136, | |
| "learning_rate": 3.242445690124729e-08, | |
| "logits/chosen": -1.0609853267669678, | |
| "logits/rejected": -1.0725332498550415, | |
| "logps/chosen": -745.48046875, | |
| "logps/rejected": -863.9573974609375, | |
| "loss": 0.3536, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -4.6711931228637695, | |
| "rewards/margins": 1.1300690174102783, | |
| "rewards/rejected": -5.801262378692627, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 0.8763024312049158, | |
| "grad_norm": 22.5406397148522, | |
| "learning_rate": 3.133285866941445e-08, | |
| "logits/chosen": -0.9890793561935425, | |
| "logits/rejected": -1.017478585243225, | |
| "logps/chosen": -489.8434143066406, | |
| "logps/rejected": -649.6760864257812, | |
| "loss": 0.463, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.189424991607666, | |
| "rewards/margins": 1.429877758026123, | |
| "rewards/rejected": -4.619302749633789, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.8784397542078547, | |
| "grad_norm": 16.06885831310049, | |
| "learning_rate": 3.025908982508966e-08, | |
| "logits/chosen": -1.04689359664917, | |
| "logits/rejected": -0.9806681871414185, | |
| "logps/chosen": -767.8428344726562, | |
| "logps/rejected": -1035.4447021484375, | |
| "loss": 0.3126, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -4.467133045196533, | |
| "rewards/margins": 2.5061354637145996, | |
| "rewards/rejected": -6.973268985748291, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 0.8805770772107935, | |
| "grad_norm": 19.662285734659836, | |
| "learning_rate": 2.9203210445508332e-08, | |
| "logits/chosen": -1.0549553632736206, | |
| "logits/rejected": -1.008999228477478, | |
| "logps/chosen": -739.88720703125, | |
| "logps/rejected": -918.3170166015625, | |
| "loss": 0.4546, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.464229583740234, | |
| "rewards/margins": 1.5558303594589233, | |
| "rewards/rejected": -6.0200605392456055, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.8827144002137323, | |
| "grad_norm": 27.990757995621482, | |
| "learning_rate": 2.8165279606992298e-08, | |
| "logits/chosen": -0.9684846997261047, | |
| "logits/rejected": -0.9231318235397339, | |
| "logps/chosen": -719.86083984375, | |
| "logps/rejected": -912.4309692382812, | |
| "loss": 0.3979, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.8175811767578125, | |
| "rewards/margins": 1.912165641784668, | |
| "rewards/rejected": -6.729746341705322, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 0.8848517232166712, | |
| "grad_norm": 17.729091244130924, | |
| "learning_rate": 2.7145355381644793e-08, | |
| "logits/chosen": -1.0500560998916626, | |
| "logits/rejected": -1.0563499927520752, | |
| "logps/chosen": -557.0240478515625, | |
| "logps/rejected": -717.523681640625, | |
| "loss": 0.3655, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -3.6325652599334717, | |
| "rewards/margins": 1.5925527811050415, | |
| "rewards/rejected": -5.225118160247803, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.88698904621961, | |
| "grad_norm": 19.8351911008517, | |
| "learning_rate": 2.6143494834101153e-08, | |
| "logits/chosen": -0.9061897993087769, | |
| "logits/rejected": -0.9052097797393799, | |
| "logps/chosen": -704.2842407226562, | |
| "logps/rejected": -743.07958984375, | |
| "loss": 0.3664, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -4.4450836181640625, | |
| "rewards/margins": 0.44577062129974365, | |
| "rewards/rejected": -4.890854835510254, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.8891263692225487, | |
| "grad_norm": 37.85808985299137, | |
| "learning_rate": 2.5159754018336403e-08, | |
| "logits/chosen": -1.1449795961380005, | |
| "logits/rejected": -1.1638754606246948, | |
| "logps/chosen": -737.4818115234375, | |
| "logps/rejected": -989.4339599609375, | |
| "loss": 0.4523, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.679790019989014, | |
| "rewards/margins": 2.3331949710845947, | |
| "rewards/rejected": -7.012984275817871, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.8912636922254876, | |
| "grad_norm": 21.579885808296115, | |
| "learning_rate": 2.4194187974528553e-08, | |
| "logits/chosen": -1.104061484336853, | |
| "logits/rejected": -1.0894217491149902, | |
| "logps/chosen": -679.447509765625, | |
| "logps/rejected": -833.4044189453125, | |
| "loss": 0.3835, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.530219554901123, | |
| "rewards/margins": 1.5757548809051514, | |
| "rewards/rejected": -6.105974197387695, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 0.8934010152284264, | |
| "grad_norm": 18.32956314016428, | |
| "learning_rate": 2.3246850725979427e-08, | |
| "logits/chosen": -1.0385122299194336, | |
| "logits/rejected": -1.0385818481445312, | |
| "logps/chosen": -666.721435546875, | |
| "logps/rejected": -787.4942626953125, | |
| "loss": 0.3856, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.093487739562988, | |
| "rewards/margins": 1.256799340248108, | |
| "rewards/rejected": -5.350286960601807, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 0.8955383382313652, | |
| "grad_norm": 26.323713896853675, | |
| "learning_rate": 2.2317795276091977e-08, | |
| "logits/chosen": -0.9605960249900818, | |
| "logits/rejected": -0.9666808247566223, | |
| "logps/chosen": -626.7743530273438, | |
| "logps/rejected": -742.9835205078125, | |
| "loss": 0.4861, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -3.809014320373535, | |
| "rewards/margins": 1.3489842414855957, | |
| "rewards/rejected": -5.157999038696289, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 0.897675661234304, | |
| "grad_norm": 29.92038686319786, | |
| "learning_rate": 2.1407073605404862e-08, | |
| "logits/chosen": -1.0402368307113647, | |
| "logits/rejected": -0.9794477224349976, | |
| "logps/chosen": -743.0745849609375, | |
| "logps/rejected": -964.0947875976562, | |
| "loss": 0.4477, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -4.6502366065979, | |
| "rewards/margins": 2.061614990234375, | |
| "rewards/rejected": -6.711851119995117, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.8998129842372429, | |
| "grad_norm": 18.598312258529717, | |
| "learning_rate": 2.0514736668683917e-08, | |
| "logits/chosen": -1.1911952495574951, | |
| "logits/rejected": -1.2088751792907715, | |
| "logps/chosen": -688.5134887695312, | |
| "logps/rejected": -876.8580932617188, | |
| "loss": 0.4087, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.259193420410156, | |
| "rewards/margins": 1.567337155342102, | |
| "rewards/rejected": -5.826530456542969, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 0.9019503072401817, | |
| "grad_norm": 23.213199916317794, | |
| "learning_rate": 1.964083439207135e-08, | |
| "logits/chosen": -1.1022056341171265, | |
| "logits/rejected": -1.1523065567016602, | |
| "logps/chosen": -729.16259765625, | |
| "logps/rejected": -873.5059204101562, | |
| "loss": 0.4305, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -4.697969913482666, | |
| "rewards/margins": 1.660890817642212, | |
| "rewards/rejected": -6.358860969543457, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 0.9040876302431204, | |
| "grad_norm": 17.78006990070099, | |
| "learning_rate": 1.878541567029251e-08, | |
| "logits/chosen": -1.0770941972732544, | |
| "logits/rejected": -1.0126092433929443, | |
| "logps/chosen": -606.9859008789062, | |
| "logps/rejected": -791.86474609375, | |
| "loss": 0.3209, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -4.224216461181641, | |
| "rewards/margins": 1.6672827005386353, | |
| "rewards/rejected": -5.891498565673828, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 0.9062249532460593, | |
| "grad_norm": 28.8620946540648, | |
| "learning_rate": 1.794852836392009e-08, | |
| "logits/chosen": -0.9167647957801819, | |
| "logits/rejected": -0.9531383514404297, | |
| "logps/chosen": -565.0890502929688, | |
| "logps/rejected": -742.736328125, | |
| "loss": 0.4255, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -3.4048359394073486, | |
| "rewards/margins": 1.6988798379898071, | |
| "rewards/rejected": -5.103715896606445, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.9083622762489981, | |
| "grad_norm": 20.499467136576563, | |
| "learning_rate": 1.7130219296696264e-08, | |
| "logits/chosen": -0.9955915212631226, | |
| "logits/rejected": -1.0456228256225586, | |
| "logps/chosen": -657.0418090820312, | |
| "logps/rejected": -762.1922607421875, | |
| "loss": 0.3629, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -3.7994446754455566, | |
| "rewards/margins": 1.2025730609893799, | |
| "rewards/rejected": -5.002017974853516, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.9104995992519369, | |
| "grad_norm": 17.206845074940055, | |
| "learning_rate": 1.6330534252913086e-08, | |
| "logits/chosen": -1.0955368280410767, | |
| "logits/rejected": -1.081055998802185, | |
| "logps/chosen": -706.4891967773438, | |
| "logps/rejected": -900.5104370117188, | |
| "loss": 0.3426, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.797817230224609, | |
| "rewards/margins": 1.9311976432800293, | |
| "rewards/rejected": -6.729015350341797, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.9126369222548758, | |
| "grad_norm": 17.119036708717413, | |
| "learning_rate": 1.5549517974850724e-08, | |
| "logits/chosen": -0.8513895273208618, | |
| "logits/rejected": -0.8118205070495605, | |
| "logps/chosen": -611.9838256835938, | |
| "logps/rejected": -755.0986328125, | |
| "loss": 0.3786, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.1998372077941895, | |
| "rewards/margins": 1.362562656402588, | |
| "rewards/rejected": -5.562399864196777, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 0.9147742452578146, | |
| "grad_norm": 30.666142274697034, | |
| "learning_rate": 1.4787214160274341e-08, | |
| "logits/chosen": -1.1465550661087036, | |
| "logits/rejected": -1.1435879468917847, | |
| "logps/chosen": -637.793701171875, | |
| "logps/rejected": -803.086669921875, | |
| "loss": 0.4319, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.044920444488525, | |
| "rewards/margins": 1.6718565225601196, | |
| "rewards/rejected": -5.7167768478393555, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.9169115682607534, | |
| "grad_norm": 17.479325335479295, | |
| "learning_rate": 1.4043665459988868e-08, | |
| "logits/chosen": -1.1191812753677368, | |
| "logits/rejected": -1.0620485544204712, | |
| "logps/chosen": -617.3558349609375, | |
| "logps/rejected": -858.5902099609375, | |
| "loss": 0.3978, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -4.054196357727051, | |
| "rewards/margins": 2.3324999809265137, | |
| "rewards/rejected": -6.3866963386535645, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 0.9190488912636923, | |
| "grad_norm": 19.870927489532274, | |
| "learning_rate": 1.3318913475452925e-08, | |
| "logits/chosen": -1.1520562171936035, | |
| "logits/rejected": -1.0596178770065308, | |
| "logps/chosen": -827.845947265625, | |
| "logps/rejected": -1057.0477294921875, | |
| "loss": 0.3633, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -5.0968828201293945, | |
| "rewards/margins": 2.2637462615966797, | |
| "rewards/rejected": -7.360629081726074, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.921186214266631, | |
| "grad_norm": 19.001198221145234, | |
| "learning_rate": 1.2612998756451366e-08, | |
| "logits/chosen": -1.0448460578918457, | |
| "logits/rejected": -1.0036524534225464, | |
| "logps/chosen": -565.2469482421875, | |
| "logps/rejected": -768.9799194335938, | |
| "loss": 0.3764, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.4013609886169434, | |
| "rewards/margins": 2.0993523597717285, | |
| "rewards/rejected": -5.500713348388672, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 0.9233235372695698, | |
| "grad_norm": 24.185158250695462, | |
| "learning_rate": 1.192596079882613e-08, | |
| "logits/chosen": -1.016557216644287, | |
| "logits/rejected": -0.9917696714401245, | |
| "logps/chosen": -754.6041259765625, | |
| "logps/rejected": -1039.5848388671875, | |
| "loss": 0.4467, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -4.575927257537842, | |
| "rewards/margins": 2.5993194580078125, | |
| "rewards/rejected": -7.175246715545654, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.9254608602725087, | |
| "grad_norm": 18.630615195275038, | |
| "learning_rate": 1.1257838042266754e-08, | |
| "logits/chosen": -0.9740123152732849, | |
| "logits/rejected": -0.9422295689582825, | |
| "logps/chosen": -684.7318115234375, | |
| "logps/rejected": -853.54931640625, | |
| "loss": 0.3922, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.472279071807861, | |
| "rewards/margins": 1.5442816019058228, | |
| "rewards/rejected": -6.016561031341553, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 0.9275981832754475, | |
| "grad_norm": 19.666494529888574, | |
| "learning_rate": 1.0608667868159592e-08, | |
| "logits/chosen": -1.0905375480651855, | |
| "logits/rejected": -1.0876115560531616, | |
| "logps/chosen": -828.9867553710938, | |
| "logps/rejected": -1100.01318359375, | |
| "loss": 0.3426, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.9841694831848145, | |
| "rewards/margins": 2.305903911590576, | |
| "rewards/rejected": -7.290073394775391, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 0.9297355062783863, | |
| "grad_norm": 18.138692746396494, | |
| "learning_rate": 9.978486597496372e-09, | |
| "logits/chosen": -0.9997820854187012, | |
| "logits/rejected": -0.997348427772522, | |
| "logps/chosen": -644.03955078125, | |
| "logps/rejected": -883.5596923828125, | |
| "loss": 0.3441, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.433194160461426, | |
| "rewards/margins": 2.0547845363616943, | |
| "rewards/rejected": -6.487978935241699, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.9318728292813251, | |
| "grad_norm": 22.806909647633304, | |
| "learning_rate": 9.367329488842001e-09, | |
| "logits/chosen": -1.0480362176895142, | |
| "logits/rejected": -1.025497317314148, | |
| "logps/chosen": -719.160400390625, | |
| "logps/rejected": -870.7994384765625, | |
| "loss": 0.3881, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.514778137207031, | |
| "rewards/margins": 1.344374656677246, | |
| "rewards/rejected": -5.859152317047119, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 0.934010152284264, | |
| "grad_norm": 34.30034026086993, | |
| "learning_rate": 8.775230736361733e-09, | |
| "logits/chosen": -0.9019148349761963, | |
| "logits/rejected": -1.021255612373352, | |
| "logps/chosen": -610.4000244140625, | |
| "logps/rejected": -666.5303344726562, | |
| "loss": 0.5228, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -3.8257713317871094, | |
| "rewards/margins": 0.6458057165145874, | |
| "rewards/rejected": -4.471577167510986, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 0.9361474752872028, | |
| "grad_norm": 26.008895134527737, | |
| "learning_rate": 8.202223467908299e-09, | |
| "logits/chosen": -1.208720088005066, | |
| "logits/rejected": -1.1576639413833618, | |
| "logps/chosen": -704.7356567382812, | |
| "logps/rejected": -935.2647094726562, | |
| "loss": 0.429, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.226789474487305, | |
| "rewards/margins": 2.07527494430542, | |
| "rewards/rejected": -6.302064418792725, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 0.9382847982901416, | |
| "grad_norm": 23.1664224383978, | |
| "learning_rate": 7.64833974316801e-09, | |
| "logits/chosen": -0.9366050958633423, | |
| "logits/rejected": -1.0036077499389648, | |
| "logps/chosen": -692.1240844726562, | |
| "logps/rejected": -962.029541015625, | |
| "loss": 0.3745, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.578640937805176, | |
| "rewards/margins": 2.751880407333374, | |
| "rewards/rejected": -7.330521583557129, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 0.9404221212930804, | |
| "grad_norm": 25.822513994850507, | |
| "learning_rate": 7.113610551867583e-09, | |
| "logits/chosen": -1.1607967615127563, | |
| "logits/rejected": -1.162611484527588, | |
| "logps/chosen": -727.8070678710938, | |
| "logps/rejected": -895.8046875, | |
| "loss": 0.4345, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -4.304068565368652, | |
| "rewards/margins": 1.6857025623321533, | |
| "rewards/rejected": -5.989770889282227, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.9425594442960192, | |
| "grad_norm": 24.174938576336224, | |
| "learning_rate": 6.5980658120396106e-09, | |
| "logits/chosen": -1.013208270072937, | |
| "logits/rejected": -0.9189801216125488, | |
| "logps/chosen": -747.7554321289062, | |
| "logps/rejected": -1023.7532348632812, | |
| "loss": 0.4307, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.824457168579102, | |
| "rewards/margins": 2.6137688159942627, | |
| "rewards/rejected": -7.438226222991943, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 0.944696767298958, | |
| "grad_norm": 21.66924362068404, | |
| "learning_rate": 6.101734368349104e-09, | |
| "logits/chosen": -1.089997410774231, | |
| "logits/rejected": -1.1180981397628784, | |
| "logps/chosen": -580.3588256835938, | |
| "logps/rejected": -653.803955078125, | |
| "loss": 0.3454, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.4357962608337402, | |
| "rewards/margins": 0.8316761255264282, | |
| "rewards/rejected": -4.267472267150879, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.9468340903018969, | |
| "grad_norm": 18.92185198276508, | |
| "learning_rate": 5.624643990479616e-09, | |
| "logits/chosen": -0.9731349349021912, | |
| "logits/rejected": -1.0123114585876465, | |
| "logps/chosen": -744.20068359375, | |
| "logps/rejected": -880.1783447265625, | |
| "loss": 0.4027, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.492010116577148, | |
| "rewards/margins": 1.5610084533691406, | |
| "rewards/rejected": -6.053019046783447, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 0.9489714133048357, | |
| "grad_norm": 19.40472791239129, | |
| "learning_rate": 5.166821371579277e-09, | |
| "logits/chosen": -0.8690105676651001, | |
| "logits/rejected": -0.9122138023376465, | |
| "logps/chosen": -592.9273071289062, | |
| "logps/rejected": -692.8623046875, | |
| "loss": 0.3917, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -3.9664011001586914, | |
| "rewards/margins": 1.128931999206543, | |
| "rewards/rejected": -5.095333099365234, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 0.9511087363077745, | |
| "grad_norm": 17.606366912659396, | |
| "learning_rate": 4.7282921267675306e-09, | |
| "logits/chosen": -1.1353975534439087, | |
| "logits/rejected": -1.069037914276123, | |
| "logps/chosen": -617.16845703125, | |
| "logps/rejected": -833.6134033203125, | |
| "loss": 0.3496, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.9898786544799805, | |
| "rewards/margins": 1.8373284339904785, | |
| "rewards/rejected": -5.827206611633301, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.9532460593107134, | |
| "grad_norm": 22.334810036861253, | |
| "learning_rate": 4.309080791701819e-09, | |
| "logits/chosen": -0.9705511927604675, | |
| "logits/rejected": -0.9864068031311035, | |
| "logps/chosen": -772.231689453125, | |
| "logps/rejected": -924.7553100585938, | |
| "loss": 0.3883, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.6818671226501465, | |
| "rewards/margins": 1.767922043800354, | |
| "rewards/rejected": -6.449788570404053, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 0.9553833823136522, | |
| "grad_norm": 18.928358142495153, | |
| "learning_rate": 3.909210821205016e-09, | |
| "logits/chosen": -1.0940232276916504, | |
| "logits/rejected": -1.1222940683364868, | |
| "logps/chosen": -569.94482421875, | |
| "logps/rejected": -690.92578125, | |
| "loss": 0.3731, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.7058024406433105, | |
| "rewards/margins": 1.2121186256408691, | |
| "rewards/rejected": -4.91792106628418, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 0.957520705316591, | |
| "grad_norm": 19.387388769733043, | |
| "learning_rate": 3.528704587952963e-09, | |
| "logits/chosen": -0.9721835255622864, | |
| "logits/rejected": -0.9633288383483887, | |
| "logps/chosen": -691.5062255859375, | |
| "logps/rejected": -864.804443359375, | |
| "loss": 0.3357, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.21799898147583, | |
| "rewards/margins": 1.6493499279022217, | |
| "rewards/rejected": -5.867349147796631, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 0.9596580283195298, | |
| "grad_norm": 15.631110311383365, | |
| "learning_rate": 3.1675833812228547e-09, | |
| "logits/chosen": -1.143801212310791, | |
| "logits/rejected": -1.1312353610992432, | |
| "logps/chosen": -591.1502685546875, | |
| "logps/rejected": -746.0208129882812, | |
| "loss": 0.4032, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.8137049674987793, | |
| "rewards/margins": 1.4380216598510742, | |
| "rewards/rejected": -5.2517266273498535, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 0.9617953513224686, | |
| "grad_norm": 23.63954938708458, | |
| "learning_rate": 2.825867405701826e-09, | |
| "logits/chosen": -1.0871686935424805, | |
| "logits/rejected": -1.0616105794906616, | |
| "logps/chosen": -785.9561157226562, | |
| "logps/rejected": -949.911376953125, | |
| "loss": 0.4622, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.555358409881592, | |
| "rewards/margins": 1.6683781147003174, | |
| "rewards/rejected": -6.22373628616333, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.9639326743254074, | |
| "grad_norm": 22.04528865685746, | |
| "learning_rate": 2.503575780356926e-09, | |
| "logits/chosen": -1.1157352924346924, | |
| "logits/rejected": -1.122366189956665, | |
| "logps/chosen": -687.6873779296875, | |
| "logps/rejected": -789.6235961914062, | |
| "loss": 0.4453, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -4.202605724334717, | |
| "rewards/margins": 1.0868072509765625, | |
| "rewards/rejected": -5.289412975311279, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 0.9660699973283462, | |
| "grad_norm": 20.620417690774662, | |
| "learning_rate": 2.2007265373650885e-09, | |
| "logits/chosen": -1.181131362915039, | |
| "logits/rejected": -1.1398506164550781, | |
| "logps/chosen": -752.2744750976562, | |
| "logps/rejected": -903.8162231445312, | |
| "loss": 0.4466, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.831580638885498, | |
| "rewards/margins": 1.5455623865127563, | |
| "rewards/rejected": -6.377142906188965, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 0.9682073203312851, | |
| "grad_norm": 21.990068975057824, | |
| "learning_rate": 1.9173366211043483e-09, | |
| "logits/chosen": -1.066049337387085, | |
| "logits/rejected": -1.145350694656372, | |
| "logps/chosen": -732.5936889648438, | |
| "logps/rejected": -820.995361328125, | |
| "loss": 0.4411, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -4.793266296386719, | |
| "rewards/margins": 0.9948696494102478, | |
| "rewards/rejected": -5.788136005401611, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 0.9703446433342239, | |
| "grad_norm": 18.52277369573019, | |
| "learning_rate": 1.653421887205747e-09, | |
| "logits/chosen": -1.002540111541748, | |
| "logits/rejected": -1.0223251581192017, | |
| "logps/chosen": -650.927978515625, | |
| "logps/rejected": -728.2699584960938, | |
| "loss": 0.3655, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.248347759246826, | |
| "rewards/margins": 0.7103351354598999, | |
| "rewards/rejected": -4.958683013916016, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 0.9724819663371627, | |
| "grad_norm": 33.05665306914434, | |
| "learning_rate": 1.408997101666326e-09, | |
| "logits/chosen": -1.1706193685531616, | |
| "logits/rejected": -1.108201265335083, | |
| "logps/chosen": -693.7764282226562, | |
| "logps/rejected": -911.11962890625, | |
| "loss": 0.4464, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.211030006408691, | |
| "rewards/margins": 1.9548094272613525, | |
| "rewards/rejected": -6.165839195251465, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.9746192893401016, | |
| "grad_norm": 18.73958161040263, | |
| "learning_rate": 1.1840759400229328e-09, | |
| "logits/chosen": -0.9501774907112122, | |
| "logits/rejected": -0.9423767924308777, | |
| "logps/chosen": -872.7874755859375, | |
| "logps/rejected": -1080.9110107421875, | |
| "loss": 0.4383, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -5.329258441925049, | |
| "rewards/margins": 1.8498148918151855, | |
| "rewards/rejected": -7.179073810577393, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 0.9767566123430403, | |
| "grad_norm": 23.633155328899434, | |
| "learning_rate": 9.786709865869547e-10, | |
| "logits/chosen": -1.0224264860153198, | |
| "logits/rejected": -0.9899497032165527, | |
| "logps/chosen": -691.5697631835938, | |
| "logps/rejected": -1032.654052734375, | |
| "loss": 0.3647, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -4.561187744140625, | |
| "rewards/margins": 3.524183750152588, | |
| "rewards/rejected": -8.085371971130371, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 0.9788939353459791, | |
| "grad_norm": 20.02815998548278, | |
| "learning_rate": 7.927937337403711e-10, | |
| "logits/chosen": -1.1305997371673584, | |
| "logits/rejected": -1.1589406728744507, | |
| "logps/chosen": -720.5179443359375, | |
| "logps/rejected": -914.910400390625, | |
| "loss": 0.413, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -4.451896667480469, | |
| "rewards/margins": 2.01759672164917, | |
| "rewards/rejected": -6.4694929122924805, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 0.981031258348918, | |
| "grad_norm": 21.2779508066291, | |
| "learning_rate": 6.264545812926958e-10, | |
| "logits/chosen": -1.0647788047790527, | |
| "logits/rejected": -1.044015884399414, | |
| "logps/chosen": -644.20263671875, | |
| "logps/rejected": -890.7341918945312, | |
| "loss": 0.4476, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -3.727494239807129, | |
| "rewards/margins": 2.3570847511291504, | |
| "rewards/rejected": -6.0845794677734375, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 0.9831685813518568, | |
| "grad_norm": 19.021711845408657, | |
| "learning_rate": 4.796628358991583e-10, | |
| "logits/chosen": -1.0134469270706177, | |
| "logits/rejected": -1.0272549390792847, | |
| "logps/chosen": -688.755859375, | |
| "logps/rejected": -888.3029174804688, | |
| "loss": 0.3752, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -4.3303303718566895, | |
| "rewards/margins": 1.7046260833740234, | |
| "rewards/rejected": -6.034955978393555, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.9853059043547956, | |
| "grad_norm": 22.296101394440072, | |
| "learning_rate": 3.524267105398937e-10, | |
| "logits/chosen": -1.0911214351654053, | |
| "logits/rejected": -1.1538598537445068, | |
| "logps/chosen": -693.211181640625, | |
| "logps/rejected": -855.1487426757812, | |
| "loss": 0.3781, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -4.124536514282227, | |
| "rewards/margins": 1.6871302127838135, | |
| "rewards/rejected": -5.811666965484619, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 0.9874432273577345, | |
| "grad_norm": 18.18277828641047, | |
| "learning_rate": 2.447533240604871e-10, | |
| "logits/chosen": -1.0473253726959229, | |
| "logits/rejected": -1.0253962278366089, | |
| "logps/chosen": -531.8089599609375, | |
| "logps/rejected": -790.6080322265625, | |
| "loss": 0.3867, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -3.4056613445281982, | |
| "rewards/margins": 2.266097068786621, | |
| "rewards/rejected": -5.671758651733398, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 0.9895805503606733, | |
| "grad_norm": 20.702683706379894, | |
| "learning_rate": 1.5664870077383707e-10, | |
| "logits/chosen": -1.0570118427276611, | |
| "logits/rejected": -1.0660746097564697, | |
| "logps/chosen": -655.458251953125, | |
| "logps/rejected": -789.9827880859375, | |
| "loss": 0.311, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -4.035057544708252, | |
| "rewards/margins": 1.403611183166504, | |
| "rewards/rejected": -5.438669204711914, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 0.9917178733636121, | |
| "grad_norm": 30.473984400030304, | |
| "learning_rate": 8.811777012271427e-11, | |
| "logits/chosen": -1.2154146432876587, | |
| "logits/rejected": -1.2058013677597046, | |
| "logps/chosen": -639.0227661132812, | |
| "logps/rejected": -820.2091064453125, | |
| "loss": 0.3856, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -4.121957778930664, | |
| "rewards/margins": 1.6997212171554565, | |
| "rewards/rejected": -5.821678638458252, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 0.9938551963665508, | |
| "grad_norm": 23.82578403366403, | |
| "learning_rate": 3.9164366404298435e-11, | |
| "logits/chosen": -1.0110024213790894, | |
| "logits/rejected": -1.0080432891845703, | |
| "logps/chosen": -551.5545654296875, | |
| "logps/rejected": -713.6527099609375, | |
| "loss": 0.3884, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -3.4868009090423584, | |
| "rewards/margins": 1.5603687763214111, | |
| "rewards/rejected": -5.047170162200928, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.9959925193694897, | |
| "grad_norm": 23.209923292158702, | |
| "learning_rate": 9.791228555566799e-12, | |
| "logits/chosen": -0.9966657161712646, | |
| "logits/rejected": -1.022011160850525, | |
| "logps/chosen": -511.430908203125, | |
| "logps/rejected": -595.9152221679688, | |
| "loss": 0.397, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -3.4004158973693848, | |
| "rewards/margins": 0.7674331068992615, | |
| "rewards/rejected": -4.167849063873291, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 0.9981298423724285, | |
| "grad_norm": 21.729353783070525, | |
| "learning_rate": 0.0, | |
| "logits/chosen": -1.0969784259796143, | |
| "logits/rejected": -1.0713554620742798, | |
| "logps/chosen": -821.9149780273438, | |
| "logps/rejected": -1118.57958984375, | |
| "loss": 0.3546, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -4.988923072814941, | |
| "rewards/margins": 3.1074531078338623, | |
| "rewards/rejected": -8.096376419067383, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 0.9981298423724285, | |
| "step": 467, | |
| "total_flos": 0.0, | |
| "train_loss": 0.4754098514673286, | |
| "train_runtime": 9903.4254, | |
| "train_samples_per_second": 6.046, | |
| "train_steps_per_second": 0.047 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 467, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 0.0, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |