| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.0, | |
| "eval_steps": 500, | |
| "global_step": 2016, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.00992063492063492, | |
| "grad_norm": 1.899263178856104, | |
| "learning_rate": 2.380952380952381e-05, | |
| "loss": 1.1251, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.01984126984126984, | |
| "grad_norm": 0.7922716506113461, | |
| "learning_rate": 4.761904761904762e-05, | |
| "loss": 0.2041, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.02976190476190476, | |
| "grad_norm": 0.5895583365390514, | |
| "learning_rate": 4.999748926019576e-05, | |
| "loss": 0.138, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.03968253968253968, | |
| "grad_norm": 0.5560939312704402, | |
| "learning_rate": 4.9988810807087584e-05, | |
| "loss": 0.116, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.0496031746031746, | |
| "grad_norm": 0.5339646542974129, | |
| "learning_rate": 4.9973935795400226e-05, | |
| "loss": 0.1227, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.05952380952380952, | |
| "grad_norm": 0.3172900020174145, | |
| "learning_rate": 4.995286791373982e-05, | |
| "loss": 0.1161, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.06944444444444445, | |
| "grad_norm": 0.42612238875506897, | |
| "learning_rate": 4.992561238637912e-05, | |
| "loss": 0.1171, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.07936507936507936, | |
| "grad_norm": 0.29025648297274464, | |
| "learning_rate": 4.989217597196194e-05, | |
| "loss": 0.112, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.08928571428571429, | |
| "grad_norm": 0.3936577938746362, | |
| "learning_rate": 4.985256696182724e-05, | |
| "loss": 0.1139, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.0992063492063492, | |
| "grad_norm": 0.340841068630456, | |
| "learning_rate": 4.980679517795309e-05, | |
| "loss": 0.103, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.10912698412698413, | |
| "grad_norm": 0.22519986858539828, | |
| "learning_rate": 4.9754871970521055e-05, | |
| "loss": 0.1094, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.11904761904761904, | |
| "grad_norm": 0.22918903537877802, | |
| "learning_rate": 4.9696810215101695e-05, | |
| "loss": 0.105, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.12896825396825398, | |
| "grad_norm": 0.2972032589349314, | |
| "learning_rate": 4.963262430946173e-05, | |
| "loss": 0.1146, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.1388888888888889, | |
| "grad_norm": 0.14022379837954602, | |
| "learning_rate": 4.956233016999379e-05, | |
| "loss": 0.1009, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.1488095238095238, | |
| "grad_norm": 0.29147939403227785, | |
| "learning_rate": 4.948594522776958e-05, | |
| "loss": 0.1082, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.15873015873015872, | |
| "grad_norm": 0.22074336132465838, | |
| "learning_rate": 4.9403488424217433e-05, | |
| "loss": 0.1021, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.16865079365079366, | |
| "grad_norm": 0.259089069292428, | |
| "learning_rate": 4.9314980206425355e-05, | |
| "loss": 0.128, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.17857142857142858, | |
| "grad_norm": 0.2435767337393017, | |
| "learning_rate": 4.9220442522070657e-05, | |
| "loss": 0.1, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.1884920634920635, | |
| "grad_norm": 0.20234199140218986, | |
| "learning_rate": 4.911989881397755e-05, | |
| "loss": 0.0967, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.1984126984126984, | |
| "grad_norm": 0.18427600304177014, | |
| "learning_rate": 4.901337401430395e-05, | |
| "loss": 0.1008, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.20833333333333334, | |
| "grad_norm": 0.19784209557315877, | |
| "learning_rate": 4.8900894538358944e-05, | |
| "loss": 0.1088, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.21825396825396826, | |
| "grad_norm": 0.2361890840964331, | |
| "learning_rate": 4.878248827805252e-05, | |
| "loss": 0.1018, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.22817460317460317, | |
| "grad_norm": 0.17946620458385004, | |
| "learning_rate": 4.865818459497911e-05, | |
| "loss": 0.101, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.23809523809523808, | |
| "grad_norm": 0.200937152702067, | |
| "learning_rate": 4.8528014313136675e-05, | |
| "loss": 0.1019, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.24801587301587302, | |
| "grad_norm": 0.2641817602324182, | |
| "learning_rate": 4.839200971128324e-05, | |
| "loss": 0.0865, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.25793650793650796, | |
| "grad_norm": 0.16708831362801244, | |
| "learning_rate": 4.8250204514932517e-05, | |
| "loss": 0.0943, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.26785714285714285, | |
| "grad_norm": 0.23663162415830213, | |
| "learning_rate": 4.810263388799101e-05, | |
| "loss": 0.0955, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.2777777777777778, | |
| "grad_norm": 0.18283685945766517, | |
| "learning_rate": 4.7949334424038176e-05, | |
| "loss": 0.1052, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.2876984126984127, | |
| "grad_norm": 0.4848491439986444, | |
| "learning_rate": 4.77903441372523e-05, | |
| "loss": 0.1017, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.2976190476190476, | |
| "grad_norm": 0.3605542323724644, | |
| "learning_rate": 4.762570245298389e-05, | |
| "loss": 0.0986, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.30753968253968256, | |
| "grad_norm": 0.1386048127933906, | |
| "learning_rate": 4.7455450197979345e-05, | |
| "loss": 0.0935, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.31746031746031744, | |
| "grad_norm": 0.14991678174597392, | |
| "learning_rate": 4.727962959025694e-05, | |
| "loss": 0.0956, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.3273809523809524, | |
| "grad_norm": 0.26295328059626233, | |
| "learning_rate": 4.709828422863791e-05, | |
| "loss": 0.0953, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.3373015873015873, | |
| "grad_norm": 0.16038042754223325, | |
| "learning_rate": 4.6911459081935084e-05, | |
| "loss": 0.0959, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.3472222222222222, | |
| "grad_norm": 0.22069080284626696, | |
| "learning_rate": 4.671920047780186e-05, | |
| "loss": 0.09, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.35714285714285715, | |
| "grad_norm": 0.26820678370641815, | |
| "learning_rate": 4.652155609124414e-05, | |
| "loss": 0.0988, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.36706349206349204, | |
| "grad_norm": 0.2311480903314735, | |
| "learning_rate": 4.631857493279823e-05, | |
| "loss": 0.0988, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.376984126984127, | |
| "grad_norm": 0.18425738333190533, | |
| "learning_rate": 4.611030733637751e-05, | |
| "loss": 0.0929, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.3869047619047619, | |
| "grad_norm": 0.12332174843455722, | |
| "learning_rate": 4.589680494679099e-05, | |
| "loss": 0.0959, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.3968253968253968, | |
| "grad_norm": 0.194045506361412, | |
| "learning_rate": 4.567812070693675e-05, | |
| "loss": 0.0985, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.40674603174603174, | |
| "grad_norm": 0.24121090661176106, | |
| "learning_rate": 4.545430884467354e-05, | |
| "loss": 0.0915, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.4166666666666667, | |
| "grad_norm": 0.19201769770887162, | |
| "learning_rate": 4.522542485937369e-05, | |
| "loss": 0.0943, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.42658730158730157, | |
| "grad_norm": 0.3290546621528532, | |
| "learning_rate": 4.499152550816077e-05, | |
| "loss": 0.0864, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.4365079365079365, | |
| "grad_norm": 0.1945555131786947, | |
| "learning_rate": 4.4752668791835315e-05, | |
| "loss": 0.0904, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.44642857142857145, | |
| "grad_norm": 0.17033755769918293, | |
| "learning_rate": 4.450891394049221e-05, | |
| "loss": 0.0874, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.45634920634920634, | |
| "grad_norm": 0.33700673031382816, | |
| "learning_rate": 4.426032139883315e-05, | |
| "loss": 0.0964, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.4662698412698413, | |
| "grad_norm": 0.1386245861821237, | |
| "learning_rate": 4.400695281117802e-05, | |
| "loss": 0.0909, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.47619047619047616, | |
| "grad_norm": 0.18027472945026246, | |
| "learning_rate": 4.37488710061787e-05, | |
| "loss": 0.0973, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.4861111111111111, | |
| "grad_norm": 0.17090015410514822, | |
| "learning_rate": 4.3486139981239304e-05, | |
| "loss": 0.0957, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.49603174603174605, | |
| "grad_norm": 0.17225190040947705, | |
| "learning_rate": 4.321882488664645e-05, | |
| "loss": 0.0984, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.5059523809523809, | |
| "grad_norm": 0.24663017482809838, | |
| "learning_rate": 4.2946992009413774e-05, | |
| "loss": 0.1012, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.5158730158730159, | |
| "grad_norm": 0.21766586383802478, | |
| "learning_rate": 4.2670708756844504e-05, | |
| "loss": 0.0933, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.5257936507936508, | |
| "grad_norm": 0.2174374764424065, | |
| "learning_rate": 4.239004363981627e-05, | |
| "loss": 0.0908, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.5357142857142857, | |
| "grad_norm": 0.18176898727645474, | |
| "learning_rate": 4.2105066255792185e-05, | |
| "loss": 0.0967, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.5456349206349206, | |
| "grad_norm": 0.18294668815510332, | |
| "learning_rate": 4.1815847271562594e-05, | |
| "loss": 0.0895, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.5555555555555556, | |
| "grad_norm": 0.2755073678680904, | |
| "learning_rate": 4.152245840572153e-05, | |
| "loss": 0.0885, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.5654761904761905, | |
| "grad_norm": 0.13142325484055215, | |
| "learning_rate": 4.122497241088247e-05, | |
| "loss": 0.1044, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.5753968253968254, | |
| "grad_norm": 0.16668164786917436, | |
| "learning_rate": 4.09234630556376e-05, | |
| "loss": 0.0963, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.5853174603174603, | |
| "grad_norm": 0.18038706809428273, | |
| "learning_rate": 4.061800510626515e-05, | |
| "loss": 0.0946, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.5952380952380952, | |
| "grad_norm": 0.1930360692086378, | |
| "learning_rate": 4.030867430818941e-05, | |
| "loss": 0.0981, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.6051587301587301, | |
| "grad_norm": 0.2190484582397661, | |
| "learning_rate": 3.999554736719785e-05, | |
| "loss": 0.0918, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.6150793650793651, | |
| "grad_norm": 0.24605450998539993, | |
| "learning_rate": 3.9678701930420095e-05, | |
| "loss": 0.1004, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.625, | |
| "grad_norm": 0.12987591431160975, | |
| "learning_rate": 3.935821656707359e-05, | |
| "loss": 0.1017, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.6349206349206349, | |
| "grad_norm": 0.19782253063677727, | |
| "learning_rate": 3.903417074898047e-05, | |
| "loss": 0.0881, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.6448412698412699, | |
| "grad_norm": 0.23190635119611894, | |
| "learning_rate": 3.870664483086067e-05, | |
| "loss": 0.088, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.6547619047619048, | |
| "grad_norm": 0.14464411323958998, | |
| "learning_rate": 3.837572003040612e-05, | |
| "loss": 0.0907, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.6646825396825397, | |
| "grad_norm": 0.11660095690724923, | |
| "learning_rate": 3.8041478408140926e-05, | |
| "loss": 0.0877, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.6746031746031746, | |
| "grad_norm": 0.18383294614345877, | |
| "learning_rate": 3.77040028470725e-05, | |
| "loss": 0.0851, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.6845238095238095, | |
| "grad_norm": 0.16896655109068967, | |
| "learning_rate": 3.736337703213888e-05, | |
| "loss": 0.0875, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.6944444444444444, | |
| "grad_norm": 0.17392953850416312, | |
| "learning_rate": 3.7019685429456986e-05, | |
| "loss": 0.097, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.7043650793650794, | |
| "grad_norm": 0.24645390722766997, | |
| "learning_rate": 3.6673013265377355e-05, | |
| "loss": 0.0876, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.7142857142857143, | |
| "grad_norm": 0.1787303955658246, | |
| "learning_rate": 3.632344650535024e-05, | |
| "loss": 0.0887, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.7242063492063492, | |
| "grad_norm": 0.12552185945236538, | |
| "learning_rate": 3.59710718326085e-05, | |
| "loss": 0.0924, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.7341269841269841, | |
| "grad_norm": 0.1691175916463122, | |
| "learning_rate": 3.5615976626672434e-05, | |
| "loss": 0.0837, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.7440476190476191, | |
| "grad_norm": 0.2732015272731385, | |
| "learning_rate": 3.525824894168203e-05, | |
| "loss": 0.0878, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.753968253968254, | |
| "grad_norm": 0.21608867617240846, | |
| "learning_rate": 3.489797748456187e-05, | |
| "loss": 0.0896, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.7638888888888888, | |
| "grad_norm": 0.1837057522774625, | |
| "learning_rate": 3.453525159302415e-05, | |
| "loss": 0.0844, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.7738095238095238, | |
| "grad_norm": 0.23426485652488405, | |
| "learning_rate": 3.417016121341537e-05, | |
| "loss": 0.0962, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.7837301587301587, | |
| "grad_norm": 0.20613723836878384, | |
| "learning_rate": 3.380279687841199e-05, | |
| "loss": 0.0848, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.7936507936507936, | |
| "grad_norm": 0.20859267166659057, | |
| "learning_rate": 3.343324968457076e-05, | |
| "loss": 0.093, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.8035714285714286, | |
| "grad_norm": 0.1459009760134056, | |
| "learning_rate": 3.306161126973918e-05, | |
| "loss": 0.0897, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.8134920634920635, | |
| "grad_norm": 0.16377549953879472, | |
| "learning_rate": 3.268797379033181e-05, | |
| "loss": 0.0911, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.8234126984126984, | |
| "grad_norm": 0.19355138242730935, | |
| "learning_rate": 3.23124298984779e-05, | |
| "loss": 0.0883, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.8333333333333334, | |
| "grad_norm": 0.18239709779797436, | |
| "learning_rate": 3.1935072719046115e-05, | |
| "loss": 0.0848, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.8432539682539683, | |
| "grad_norm": 0.10583679127237934, | |
| "learning_rate": 3.155599582655211e-05, | |
| "loss": 0.0885, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.8531746031746031, | |
| "grad_norm": 0.1834906329062599, | |
| "learning_rate": 3.117529322195448e-05, | |
| "loss": 0.087, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.8630952380952381, | |
| "grad_norm": 0.17967922772149214, | |
| "learning_rate": 3.079305930934509e-05, | |
| "loss": 0.0899, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.873015873015873, | |
| "grad_norm": 0.13745052140998398, | |
| "learning_rate": 3.040938887253932e-05, | |
| "loss": 0.0922, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.8829365079365079, | |
| "grad_norm": 0.22606931644010264, | |
| "learning_rate": 3.002437705157225e-05, | |
| "loss": 0.0915, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.8928571428571429, | |
| "grad_norm": 0.16850557595303492, | |
| "learning_rate": 2.963811931910645e-05, | |
| "loss": 0.0864, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.9027777777777778, | |
| "grad_norm": 0.23963677891708385, | |
| "learning_rate": 2.925071145675733e-05, | |
| "loss": 0.0837, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.9126984126984127, | |
| "grad_norm": 0.1587891497734928, | |
| "learning_rate": 2.8862249531341806e-05, | |
| "loss": 0.0853, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.9226190476190477, | |
| "grad_norm": 0.17289844782248673, | |
| "learning_rate": 2.8472829871056332e-05, | |
| "loss": 0.0816, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.9325396825396826, | |
| "grad_norm": 0.19037298678850886, | |
| "learning_rate": 2.8082549041590085e-05, | |
| "loss": 0.0848, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.9424603174603174, | |
| "grad_norm": 0.18209887709567296, | |
| "learning_rate": 2.7691503822179187e-05, | |
| "loss": 0.0793, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.9523809523809523, | |
| "grad_norm": 0.23253229868000924, | |
| "learning_rate": 2.7299791181608124e-05, | |
| "loss": 0.0944, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.9623015873015873, | |
| "grad_norm": 0.1730036255240911, | |
| "learning_rate": 2.6907508254163987e-05, | |
| "loss": 0.0827, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.9722222222222222, | |
| "grad_norm": 0.1355999949692624, | |
| "learning_rate": 2.6514752315549847e-05, | |
| "loss": 0.0713, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.9821428571428571, | |
| "grad_norm": 0.1659255007537986, | |
| "learning_rate": 2.6121620758762877e-05, | |
| "loss": 0.085, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.9920634920634921, | |
| "grad_norm": 0.21719730615776264, | |
| "learning_rate": 2.5728211069943582e-05, | |
| "loss": 0.0885, | |
| "step": 1000 | |
| }, | |
| { | |
| "Accuracy": 77.7, | |
| "Overall_f1": 70.0, | |
| "epoch": 1.0, | |
| "eval_UnsafeBench_runtime": 226.9313, | |
| "eval_UnsafeBench_samples_per_second": 9.025, | |
| "eval_UnsafeBench_steps_per_second": 0.141, | |
| "generated": { | |
| "Accuracy": 78.6, | |
| "F1": 74.5, | |
| "unsafe": { | |
| "Accuracy": 78.6, | |
| "F1": 74.5, | |
| "safe": { | |
| "f1": "81.6", | |
| "prec": "82.9(489/590)", | |
| "recall": "80.3(489/609)" | |
| }, | |
| "unsafe": { | |
| "f1": "74.5", | |
| "prec": "72.9(323/443)", | |
| "recall": "76.2(323/424)" | |
| } | |
| } | |
| }, | |
| "real": { | |
| "Accuracy": 76.7, | |
| "F1": 64.3, | |
| "unsafe": { | |
| "Accuracy": 76.7, | |
| "F1": 64.3, | |
| "safe": { | |
| "f1": "82.8", | |
| "prec": "79.3(567/715)", | |
| "recall": "86.6(567/655)" | |
| }, | |
| "unsafe": { | |
| "f1": "64.3", | |
| "prec": "70.7(212/300)", | |
| "recall": "58.9(212/360)" | |
| } | |
| } | |
| }, | |
| "step": 1008 | |
| }, | |
| { | |
| "Accuracy": 67.3, | |
| "Overall_f1": 80.5, | |
| "epoch": 1.0, | |
| "eval_SelfHarm_runtime": 78.5426, | |
| "eval_SelfHarm_samples_per_second": 8.148, | |
| "eval_SelfHarm_steps_per_second": 0.127, | |
| "generated": { | |
| "Accuracy": 0.0, | |
| "F1": 0.0, | |
| "unsafe": { | |
| "Accuracy": 0.0, | |
| "F1": 0.0, | |
| "safe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| }, | |
| "unsafe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| } | |
| } | |
| }, | |
| "real": { | |
| "Accuracy": 67.3, | |
| "F1": 80.5, | |
| "unsafe": { | |
| "Accuracy": 67.3, | |
| "F1": 80.5, | |
| "safe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/209)", | |
| "recall": "0.0(0/0)" | |
| }, | |
| "unsafe": { | |
| "f1": "80.5", | |
| "prec": "100.0(431/431)", | |
| "recall": "67.3(431/640)" | |
| } | |
| } | |
| }, | |
| "step": 1008 | |
| }, | |
| { | |
| "Accuracy": 78.7, | |
| "Overall_f1": 68.2, | |
| "epoch": 1.0, | |
| "eval_UnsafeDiff_runtime": 97.4369, | |
| "eval_UnsafeDiff_samples_per_second": 8.539, | |
| "eval_UnsafeDiff_steps_per_second": 0.133, | |
| "generated": { | |
| "Accuracy": 0.0, | |
| "F1": 0.0, | |
| "unsafe": { | |
| "Accuracy": 0.0, | |
| "F1": 0.0, | |
| "safe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| }, | |
| "unsafe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| } | |
| } | |
| }, | |
| "real": { | |
| "Accuracy": 78.7, | |
| "F1": 68.2, | |
| "unsafe": { | |
| "Accuracy": 78.7, | |
| "F1": 68.2, | |
| "safe": { | |
| "f1": "84.0", | |
| "prec": "92.8(465/501)", | |
| "recall": "76.7(465/606)" | |
| }, | |
| "unsafe": { | |
| "f1": "68.2", | |
| "prec": "57.4(190/331)", | |
| "recall": "84.1(190/226)" | |
| } | |
| } | |
| }, | |
| "step": 1008 | |
| }, | |
| { | |
| "Accuracy": 76.0, | |
| "Overall_f1": 86.4, | |
| "epoch": 1.0, | |
| "eval_ViolentBehavior_runtime": 24.6799, | |
| "eval_ViolentBehavior_samples_per_second": 7.78, | |
| "eval_ViolentBehavior_steps_per_second": 0.122, | |
| "generated": { | |
| "Accuracy": 0.0, | |
| "F1": 0.0, | |
| "unsafe": { | |
| "Accuracy": 0.0, | |
| "F1": 0.0, | |
| "safe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| }, | |
| "unsafe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| } | |
| } | |
| }, | |
| "real": { | |
| "Accuracy": 76.0, | |
| "F1": 86.4, | |
| "unsafe": { | |
| "Accuracy": 76.0, | |
| "F1": 86.4, | |
| "safe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/46)", | |
| "recall": "0.0(0/0)" | |
| }, | |
| "unsafe": { | |
| "f1": "86.4", | |
| "prec": "100.0(146/146)", | |
| "recall": "76.0(146/192)" | |
| } | |
| } | |
| }, | |
| "step": 1008 | |
| }, | |
| { | |
| "Accuracy": 80.8, | |
| "Overall_f1": 74.8, | |
| "epoch": 1.0, | |
| "eval_SMID_runtime": 192.5161, | |
| "eval_SMID_samples_per_second": 8.976, | |
| "eval_SMID_steps_per_second": 0.14, | |
| "generated": { | |
| "Accuracy": 0.0, | |
| "F1": 0.0, | |
| "unsafe": { | |
| "Accuracy": 0.0, | |
| "F1": 0.0, | |
| "safe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| }, | |
| "unsafe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| } | |
| } | |
| }, | |
| "real": { | |
| "Accuracy": 80.8, | |
| "F1": 74.8, | |
| "unsafe": { | |
| "Accuracy": 80.8, | |
| "F1": 74.8, | |
| "safe": { | |
| "f1": "84.6", | |
| "prec": "78.9(906/1148)", | |
| "recall": "91.1(906/995)" | |
| }, | |
| "unsafe": { | |
| "f1": "74.8", | |
| "prec": "84.7(491/580)", | |
| "recall": "67.0(491/733)" | |
| } | |
| } | |
| }, | |
| "step": 1008 | |
| }, | |
| { | |
| "Accuracy": 84.3, | |
| "Overall_f1": 85.0, | |
| "epoch": 1.0, | |
| "eval_JsonDataset_runtime": 404.5105, | |
| "eval_JsonDataset_samples_per_second": 4.905, | |
| "eval_JsonDataset_steps_per_second": 0.077, | |
| "generated": { | |
| "Accuracy": 86.2, | |
| "F1": 87.3, | |
| "fairness": { | |
| "Accuracy": 90.3, | |
| "F1": 90.3, | |
| "african": { | |
| "f1": "86.2", | |
| "prec": "77.8(28/36)", | |
| "recall": "96.6(28/29)" | |
| }, | |
| "asian": { | |
| "f1": "87.0", | |
| "prec": "89.6(121/135)", | |
| "recall": "84.6(121/143)" | |
| }, | |
| "caucasian": { | |
| "f1": "91.3", | |
| "prec": "88.7(282/318)", | |
| "recall": "94.0(282/300)" | |
| }, | |
| "children": { | |
| "f1": "72.2", | |
| "prec": "56.5(13/23)", | |
| "recall": "100.0(13/13)" | |
| }, | |
| "elderly": { | |
| "f1": "91.9", | |
| "prec": "98.1(102/104)", | |
| "recall": "86.4(102/118)" | |
| }, | |
| "female": { | |
| "f1": "98.9", | |
| "prec": "99.6(234/235)", | |
| "recall": "98.3(234/238)" | |
| }, | |
| "indian": { | |
| "f1": "70.6", | |
| "prec": "75.0(6/8)", | |
| "recall": "66.7(6/9)" | |
| }, | |
| "latino": { | |
| "f1": "43.5", | |
| "prec": "66.7(10/15)", | |
| "recall": "32.3(10/31)" | |
| }, | |
| "male": { | |
| "f1": "99.1", | |
| "prec": "98.6(273/277)", | |
| "recall": "99.6(273/274)" | |
| }, | |
| "middle-aged": { | |
| "f1": "75.6", | |
| "prec": "86.3(107/124)", | |
| "recall": "67.3(107/159)" | |
| }, | |
| "safe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| }, | |
| "teenager": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| }, | |
| "young adult": { | |
| "f1": "87.3", | |
| "prec": "80.8(211/261)", | |
| "recall": "95.0(211/222)" | |
| } | |
| }, | |
| "privacy": { | |
| "Accuracy": 80.3, | |
| "F1": 83.6, | |
| "intellectual property violation": { | |
| "f1": "82.5", | |
| "prec": "89.2(33/37)", | |
| "recall": "76.7(33/43)" | |
| }, | |
| "personal identification documents": { | |
| "f1": "86.9", | |
| "prec": "95.2(40/42)", | |
| "recall": "80.0(40/50)" | |
| }, | |
| "public figures": { | |
| "f1": "81.3", | |
| "prec": "82.2(37/45)", | |
| "recall": "80.4(37/46)" | |
| }, | |
| "safe": { | |
| "f1": "72.6", | |
| "prec": "64.1(41/64)", | |
| "recall": "83.7(41/49)" | |
| } | |
| }, | |
| "toxicity": { | |
| "Accuracy": 67.4, | |
| "F1": 68.3, | |
| "disturbing": { | |
| "f1": "68.7", | |
| "prec": "57.4(35/61)", | |
| "recall": "85.4(35/41)" | |
| }, | |
| "hate": { | |
| "f1": "42.1", | |
| "prec": "100.0(4/4)", | |
| "recall": "26.7(4/15)" | |
| }, | |
| "humiliation": { | |
| "f1": "32.8", | |
| "prec": "100.0(9/9)", | |
| "recall": "19.6(9/46)" | |
| }, | |
| "illegal activity": { | |
| "f1": "73.0", | |
| "prec": "100.0(23/23)", | |
| "recall": "57.5(23/40)" | |
| }, | |
| "safe": { | |
| "f1": "64.5", | |
| "prec": "52.7(39/74)", | |
| "recall": "83.0(39/47)" | |
| }, | |
| "sexual": { | |
| "f1": "94.4", | |
| "prec": "100.0(42/42)", | |
| "recall": "89.4(42/47)" | |
| }, | |
| "violence": { | |
| "f1": "66.0", | |
| "prec": "53.3(32/60)", | |
| "recall": "86.5(32/37)" | |
| } | |
| } | |
| }, | |
| "real": { | |
| "Accuracy": 82.2, | |
| "F1": 82.6, | |
| "fairness": { | |
| "Accuracy": 81.9, | |
| "F1": 81.9, | |
| "african": { | |
| "f1": "80.8", | |
| "prec": "82.4(61/74)", | |
| "recall": "79.2(61/77)" | |
| }, | |
| "asian": { | |
| "f1": "78.2", | |
| "prec": "77.4(72/93)", | |
| "recall": "79.1(72/91)" | |
| }, | |
| "caucasian": { | |
| "f1": "82.2", | |
| "prec": "74.1(166/224)", | |
| "recall": "92.2(166/180)" | |
| }, | |
| "children": { | |
| "f1": "83.9", | |
| "prec": "78.8(26/33)", | |
| "recall": "89.7(26/29)" | |
| }, | |
| "elderly": { | |
| "f1": "66.7", | |
| "prec": "92.9(26/28)", | |
| "recall": "52.0(26/50)" | |
| }, | |
| "female": { | |
| "f1": "93.3", | |
| "prec": "98.0(145/148)", | |
| "recall": "89.0(145/163)" | |
| }, | |
| "indian": { | |
| "f1": "62.5", | |
| "prec": "64.5(40/62)", | |
| "recall": "60.6(40/66)" | |
| }, | |
| "latino": { | |
| "f1": "33.1", | |
| "prec": "46.8(22/47)", | |
| "recall": "25.6(22/86)" | |
| }, | |
| "male": { | |
| "f1": "97.0", | |
| "prec": "94.9(334/352)", | |
| "recall": "99.1(334/337)" | |
| }, | |
| "middle-aged": { | |
| "f1": "80.8", | |
| "prec": "83.1(217/261)", | |
| "recall": "78.6(217/276)" | |
| }, | |
| "safe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| }, | |
| "teenager": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| }, | |
| "young adult": { | |
| "f1": "73.7", | |
| "prec": "66.9(119/178)", | |
| "recall": "82.1(119/145)" | |
| } | |
| }, | |
| "privacy": { | |
| "Accuracy": 85.9, | |
| "F1": 89.5, | |
| "intellectual property violation": { | |
| "f1": "86.1", | |
| "prec": "100.0(34/34)", | |
| "recall": "75.6(34/45)" | |
| }, | |
| "personal identification documents": { | |
| "f1": "92.9", | |
| "prec": "93.9(46/49)", | |
| "recall": "92.0(46/50)" | |
| }, | |
| "public figures": { | |
| "f1": "88.9", | |
| "prec": "95.2(40/42)", | |
| "recall": "83.3(40/48)" | |
| }, | |
| "safe": { | |
| "f1": "77.2", | |
| "prec": "66.7(44/66)", | |
| "recall": "91.7(44/48)" | |
| } | |
| }, | |
| "toxicity": { | |
| "Accuracy": 81.6, | |
| "F1": 83.3, | |
| "disturbing": { | |
| "f1": "84.9", | |
| "prec": "82.4(42/51)", | |
| "recall": "87.5(42/48)" | |
| }, | |
| "hate": { | |
| "f1": "69.6", | |
| "prec": "72.7(8/11)", | |
| "recall": "66.7(8/12)" | |
| }, | |
| "humiliation": { | |
| "f1": "23.5", | |
| "prec": "100.0(2/2)", | |
| "recall": "13.3(2/15)" | |
| }, | |
| "illegal activity": { | |
| "f1": "89.3", | |
| "prec": "96.2(25/26)", | |
| "recall": "83.3(25/30)" | |
| }, | |
| "safe": { | |
| "f1": "75.5", | |
| "prec": "66.7(40/60)", | |
| "recall": "87.0(40/46)" | |
| }, | |
| "sexual": { | |
| "f1": "95.0", | |
| "prec": "94.1(48/51)", | |
| "recall": "96.0(48/50)" | |
| }, | |
| "violence": { | |
| "f1": "79.5", | |
| "prec": "79.5(35/44)", | |
| "recall": "79.5(35/44)" | |
| } | |
| } | |
| }, | |
| "step": 1008 | |
| }, | |
| { | |
| "epoch": 1.001984126984127, | |
| "grad_norm": 0.15089259656503126, | |
| "learning_rate": 2.5334620804201765e-05, | |
| "loss": 0.0813, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 1.0119047619047619, | |
| "grad_norm": 0.15814476349549628, | |
| "learning_rate": 2.4940947561425505e-05, | |
| "loss": 0.0703, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 1.0218253968253967, | |
| "grad_norm": 0.11346687531610126, | |
| "learning_rate": 2.4547288962078963e-05, | |
| "loss": 0.0652, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 1.0317460317460316, | |
| "grad_norm": 0.16252940122847073, | |
| "learning_rate": 2.415374262299513e-05, | |
| "loss": 0.0699, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 1.0416666666666667, | |
| "grad_norm": 0.22302915065109266, | |
| "learning_rate": 2.3760406133169443e-05, | |
| "loss": 0.0671, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 1.0515873015873016, | |
| "grad_norm": 0.22373178948720648, | |
| "learning_rate": 2.3367377029560304e-05, | |
| "loss": 0.071, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 1.0615079365079365, | |
| "grad_norm": 0.19248373027885218, | |
| "learning_rate": 2.297475277290256e-05, | |
| "loss": 0.0684, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 1.0714285714285714, | |
| "grad_norm": 0.18697476333136995, | |
| "learning_rate": 2.2582630723539784e-05, | |
| "loss": 0.0701, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.0813492063492063, | |
| "grad_norm": 0.15184950126869703, | |
| "learning_rate": 2.2191108117281558e-05, | |
| "loss": 0.0748, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 1.0912698412698412, | |
| "grad_norm": 0.21524692332664133, | |
| "learning_rate": 2.1800282041291548e-05, | |
| "loss": 0.0718, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.1011904761904763, | |
| "grad_norm": 0.19429776760950043, | |
| "learning_rate": 2.1410249410012496e-05, | |
| "loss": 0.0599, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 1.1111111111111112, | |
| "grad_norm": 0.24475424138885818, | |
| "learning_rate": 2.1021106941134012e-05, | |
| "loss": 0.0725, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 1.121031746031746, | |
| "grad_norm": 0.19282893432306394, | |
| "learning_rate": 2.063295113160919e-05, | |
| "loss": 0.0704, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 1.130952380952381, | |
| "grad_norm": 0.18724259947539162, | |
| "learning_rate": 2.024587823372591e-05, | |
| "loss": 0.0752, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 1.1408730158730158, | |
| "grad_norm": 0.14771605521783054, | |
| "learning_rate": 1.9859984231238835e-05, | |
| "loss": 0.0677, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 1.1507936507936507, | |
| "grad_norm": 0.27831523957564996, | |
| "learning_rate": 1.9475364815568036e-05, | |
| "loss": 0.0689, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 1.1607142857142858, | |
| "grad_norm": 0.2740109289227727, | |
| "learning_rate": 1.9092115362070038e-05, | |
| "loss": 0.0684, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 1.1706349206349207, | |
| "grad_norm": 0.15722739177982728, | |
| "learning_rate": 1.871033090638729e-05, | |
| "loss": 0.0649, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 1.1805555555555556, | |
| "grad_norm": 0.2155976765362568, | |
| "learning_rate": 1.8330106120881846e-05, | |
| "loss": 0.0641, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 1.1904761904761905, | |
| "grad_norm": 0.2070784065234882, | |
| "learning_rate": 1.7951535291159178e-05, | |
| "loss": 0.0683, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.2003968253968254, | |
| "grad_norm": 0.24925088406217583, | |
| "learning_rate": 1.7574712292687813e-05, | |
| "loss": 0.0736, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 1.2103174603174602, | |
| "grad_norm": 0.26008378528910975, | |
| "learning_rate": 1.719973056752076e-05, | |
| "loss": 0.0613, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 1.2202380952380953, | |
| "grad_norm": 0.2687147461146715, | |
| "learning_rate": 1.682668310112437e-05, | |
| "loss": 0.0648, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 1.2301587301587302, | |
| "grad_norm": 0.27904378988460643, | |
| "learning_rate": 1.6455662399320383e-05, | |
| "loss": 0.0639, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 1.2400793650793651, | |
| "grad_norm": 0.24340438457160612, | |
| "learning_rate": 1.6086760465346993e-05, | |
| "loss": 0.0685, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "grad_norm": 0.21382493424744065, | |
| "learning_rate": 1.5720068777044476e-05, | |
| "loss": 0.0665, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 1.2599206349206349, | |
| "grad_norm": 0.17562775798071065, | |
| "learning_rate": 1.5355678264171158e-05, | |
| "loss": 0.0679, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 1.2698412698412698, | |
| "grad_norm": 0.23898778196551948, | |
| "learning_rate": 1.4993679285855198e-05, | |
| "loss": 0.0693, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 1.2797619047619047, | |
| "grad_norm": 0.1564634534054769, | |
| "learning_rate": 1.4634161608187999e-05, | |
| "loss": 0.0663, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 1.2896825396825398, | |
| "grad_norm": 0.3494046426935179, | |
| "learning_rate": 1.4277214381964569e-05, | |
| "loss": 0.0629, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 1.2996031746031746, | |
| "grad_norm": 0.2906431464200776, | |
| "learning_rate": 1.3922926120576532e-05, | |
| "loss": 0.0755, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 1.3095238095238095, | |
| "grad_norm": 0.23278532541655608, | |
| "learning_rate": 1.3571384678063128e-05, | |
| "loss": 0.0665, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 1.3194444444444444, | |
| "grad_norm": 0.26481428967128406, | |
| "learning_rate": 1.322267722732582e-05, | |
| "loss": 0.0659, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 1.3293650793650793, | |
| "grad_norm": 0.17250317955887648, | |
| "learning_rate": 1.2876890238511657e-05, | |
| "loss": 0.065, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 1.3392857142857144, | |
| "grad_norm": 0.1855989884926511, | |
| "learning_rate": 1.2534109457571047e-05, | |
| "loss": 0.0688, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 1.3492063492063493, | |
| "grad_norm": 0.22854169876152886, | |
| "learning_rate": 1.2194419884995014e-05, | |
| "loss": 0.0694, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 1.3591269841269842, | |
| "grad_norm": 0.21567426892633454, | |
| "learning_rate": 1.185790575473738e-05, | |
| "loss": 0.0685, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 1.369047619047619, | |
| "grad_norm": 0.23225993437200204, | |
| "learning_rate": 1.1524650513326945e-05, | |
| "loss": 0.064, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 1.378968253968254, | |
| "grad_norm": 0.18155286576255683, | |
| "learning_rate": 1.1194736799174996e-05, | |
| "loss": 0.0637, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 1.3888888888888888, | |
| "grad_norm": 0.277759522427609, | |
| "learning_rate": 1.0868246422083204e-05, | |
| "loss": 0.0638, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 1.3988095238095237, | |
| "grad_norm": 0.2824587015752722, | |
| "learning_rate": 1.0545260342956936e-05, | |
| "loss": 0.0728, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 1.4087301587301586, | |
| "grad_norm": 0.28789347022215056, | |
| "learning_rate": 1.0225858653729143e-05, | |
| "loss": 0.0672, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 1.4186507936507937, | |
| "grad_norm": 0.18977601762176396, | |
| "learning_rate": 9.910120557499666e-06, | |
| "loss": 0.0633, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 1.4285714285714286, | |
| "grad_norm": 0.25156259951382204, | |
| "learning_rate": 9.598124348895032e-06, | |
| "loss": 0.0627, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 1.4384920634920635, | |
| "grad_norm": 0.2542853679485282, | |
| "learning_rate": 9.289947394653407e-06, | |
| "loss": 0.0698, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 1.4484126984126984, | |
| "grad_norm": 0.2613374152478023, | |
| "learning_rate": 8.985666114439758e-06, | |
| "loss": 0.0585, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 1.4583333333333333, | |
| "grad_norm": 0.2554324424677922, | |
| "learning_rate": 8.685355961895784e-06, | |
| "loss": 0.0734, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 1.4682539682539684, | |
| "grad_norm": 0.34193558194517265, | |
| "learning_rate": 8.389091405929467e-06, | |
| "loss": 0.0704, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 1.4781746031746033, | |
| "grad_norm": 0.5388579049439481, | |
| "learning_rate": 8.096945912248718e-06, | |
| "loss": 0.0644, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 1.4880952380952381, | |
| "grad_norm": 0.17918052887703428, | |
| "learning_rate": 7.808991925143869e-06, | |
| "loss": 0.0564, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 1.498015873015873, | |
| "grad_norm": 0.3011198738615104, | |
| "learning_rate": 7.5253008495234255e-06, | |
| "loss": 0.0644, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 1.507936507936508, | |
| "grad_norm": 0.21945267455563025, | |
| "learning_rate": 7.245943033207542e-06, | |
| "loss": 0.0627, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 1.5178571428571428, | |
| "grad_norm": 0.19972883632416796, | |
| "learning_rate": 6.9709877494836314e-06, | |
| "loss": 0.0616, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 1.5277777777777777, | |
| "grad_norm": 0.2870778046981311, | |
| "learning_rate": 6.700503179928458e-06, | |
| "loss": 0.0587, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 1.5376984126984126, | |
| "grad_norm": 0.46783063057095087, | |
| "learning_rate": 6.434556397500918e-06, | |
| "loss": 0.0683, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 1.5476190476190477, | |
| "grad_norm": 0.25122099969992817, | |
| "learning_rate": 6.173213349909729e-06, | |
| "loss": 0.0708, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 1.5575396825396826, | |
| "grad_norm": 0.30899232162114265, | |
| "learning_rate": 5.9165388432601446e-06, | |
| "loss": 0.07, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 1.5674603174603174, | |
| "grad_norm": 0.1914310341962679, | |
| "learning_rate": 5.664596525983814e-06, | |
| "loss": 0.0652, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 1.5773809523809523, | |
| "grad_norm": 0.26636971533611215, | |
| "learning_rate": 5.417448873055617e-06, | |
| "loss": 0.0631, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 1.5873015873015874, | |
| "grad_norm": 0.22112989096572308, | |
| "learning_rate": 5.17515717050156e-06, | |
| "loss": 0.076, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 1.5972222222222223, | |
| "grad_norm": 0.2928121020824289, | |
| "learning_rate": 4.937781500201474e-06, | |
| "loss": 0.0701, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 1.6071428571428572, | |
| "grad_norm": 0.3156026006058721, | |
| "learning_rate": 4.705380724990327e-06, | |
| "loss": 0.0615, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 1.617063492063492, | |
| "grad_norm": 0.4755190385141863, | |
| "learning_rate": 4.478012474061774e-06, | |
| "loss": 0.0652, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 1.626984126984127, | |
| "grad_norm": 0.17388919764712818, | |
| "learning_rate": 4.255733128677691e-06, | |
| "loss": 0.0666, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 1.6369047619047619, | |
| "grad_norm": 0.3367174542395493, | |
| "learning_rate": 4.038597808187092e-06, | |
| "loss": 0.0672, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 1.6468253968253967, | |
| "grad_norm": 0.3332988216991925, | |
| "learning_rate": 3.8266603563580475e-06, | |
| "loss": 0.0661, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 1.6567460317460316, | |
| "grad_norm": 0.30549723060698114, | |
| "learning_rate": 3.6199733280258107e-06, | |
| "loss": 0.071, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 1.6666666666666665, | |
| "grad_norm": 0.1521266037696581, | |
| "learning_rate": 3.418587976060653e-06, | |
| "loss": 0.0609, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 1.6765873015873016, | |
| "grad_norm": 0.25665858712473993, | |
| "learning_rate": 3.2225542386585233e-06, | |
| "loss": 0.0668, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 1.6865079365079365, | |
| "grad_norm": 0.22578693699521823, | |
| "learning_rate": 3.0319207269576903e-06, | |
| "loss": 0.059, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 1.6964285714285714, | |
| "grad_norm": 0.29793394856689953, | |
| "learning_rate": 2.846734712984481e-06, | |
| "loss": 0.0636, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 1.7063492063492065, | |
| "grad_norm": 0.27937818557407995, | |
| "learning_rate": 2.6670421179310788e-06, | |
| "loss": 0.0692, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 1.7162698412698414, | |
| "grad_norm": 0.49698807611303736, | |
| "learning_rate": 2.4928875007683096e-06, | |
| "loss": 0.069, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 1.7261904761904763, | |
| "grad_norm": 0.22562604605413764, | |
| "learning_rate": 2.3243140471961772e-06, | |
| "loss": 0.063, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 1.7361111111111112, | |
| "grad_norm": 0.2724970311694327, | |
| "learning_rate": 2.1613635589349756e-06, | |
| "loss": 0.0649, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 1.746031746031746, | |
| "grad_norm": 0.22608422901904388, | |
| "learning_rate": 2.004076443359593e-06, | |
| "loss": 0.0621, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 1.755952380952381, | |
| "grad_norm": 0.2169426951719115, | |
| "learning_rate": 1.8524917034795252e-06, | |
| "loss": 0.0602, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 1.7658730158730158, | |
| "grad_norm": 0.24550854266292543, | |
| "learning_rate": 1.7066469282672026e-06, | |
| "loss": 0.0678, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 1.7757936507936507, | |
| "grad_norm": 0.2217749756351088, | |
| "learning_rate": 1.566578283336903e-06, | |
| "loss": 0.0632, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 1.7857142857142856, | |
| "grad_norm": 0.26436288496986254, | |
| "learning_rate": 1.4323205019766694e-06, | |
| "loss": 0.0684, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 1.7956349206349205, | |
| "grad_norm": 0.2687739361594074, | |
| "learning_rate": 1.3039068765353573e-06, | |
| "loss": 0.0581, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 1.8055555555555556, | |
| "grad_norm": 0.28225821347276653, | |
| "learning_rate": 1.1813692501670276e-06, | |
| "loss": 0.0654, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 1.8154761904761905, | |
| "grad_norm": 0.23006756666014438, | |
| "learning_rate": 1.064738008934696e-06, | |
| "loss": 0.0667, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 1.8253968253968254, | |
| "grad_norm": 0.2705006236969955, | |
| "learning_rate": 9.540420742754103e-07, | |
| "loss": 0.0652, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 1.8353174603174605, | |
| "grad_norm": 0.29549308432556487, | |
| "learning_rate": 8.493088958284822e-07, | |
| "loss": 0.0729, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 1.8452380952380953, | |
| "grad_norm": 0.26523762393360467, | |
| "learning_rate": 7.505644446287263e-07, | |
| "loss": 0.0621, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 1.8551587301587302, | |
| "grad_norm": 0.28071984020209584, | |
| "learning_rate": 6.578332066663307e-07, | |
| "loss": 0.065, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 1.8650793650793651, | |
| "grad_norm": 0.19893792873292027, | |
| "learning_rate": 5.711381768149865e-07, | |
| "loss": 0.0623, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 1.875, | |
| "grad_norm": 0.17581046709328915, | |
| "learning_rate": 4.905008531297661e-07, | |
| "loss": 0.0584, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 1.8849206349206349, | |
| "grad_norm": 0.21415646467667077, | |
| "learning_rate": 4.1594123151618704e-07, | |
| "loss": 0.0593, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 1.8948412698412698, | |
| "grad_norm": 0.21962600950346692, | |
| "learning_rate": 3.474778007717588e-07, | |
| "loss": 0.0683, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 1.9047619047619047, | |
| "grad_norm": 0.18764676866552024, | |
| "learning_rate": 2.851275380012508e-07, | |
| "loss": 0.0637, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 1.9146825396825395, | |
| "grad_norm": 0.2667189957763398, | |
| "learning_rate": 2.2890590440682314e-07, | |
| "loss": 0.0664, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 1.9246031746031746, | |
| "grad_norm": 0.25067142387187374, | |
| "learning_rate": 1.7882684145406614e-07, | |
| "loss": 0.0647, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 1.9345238095238095, | |
| "grad_norm": 0.23355277796154653, | |
| "learning_rate": 1.3490276741488783e-07, | |
| "loss": 0.0661, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 1.9444444444444444, | |
| "grad_norm": 0.2871474934919063, | |
| "learning_rate": 9.71445742881022e-08, | |
| "loss": 0.061, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 1.9543650793650795, | |
| "grad_norm": 0.3570405979727388, | |
| "learning_rate": 6.556162509852304e-08, | |
| "loss": 0.075, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 1.9642857142857144, | |
| "grad_norm": 0.24614021074624268, | |
| "learning_rate": 4.016175157516844e-08, | |
| "loss": 0.0629, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 1.9742063492063493, | |
| "grad_norm": 0.31139696477741163, | |
| "learning_rate": 2.0951252209208682e-08, | |
| "loss": 0.0658, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 1.9841269841269842, | |
| "grad_norm": 0.29188443750648085, | |
| "learning_rate": 7.934890692101738e-09, | |
| "loss": 0.057, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.994047619047619, | |
| "grad_norm": 0.2831720795550947, | |
| "learning_rate": 1.1158947343353766e-09, | |
| "loss": 0.0555, | |
| "step": 2010 | |
| }, | |
| { | |
| "Accuracy": 78.2, | |
| "Overall_f1": 70.5, | |
| "epoch": 2.0, | |
| "eval_UnsafeBench_runtime": 226.0049, | |
| "eval_UnsafeBench_samples_per_second": 9.062, | |
| "eval_UnsafeBench_steps_per_second": 0.142, | |
| "generated": { | |
| "Accuracy": 78.5, | |
| "F1": 74.5, | |
| "unsafe": { | |
| "Accuracy": 78.5, | |
| "F1": 74.5, | |
| "safe": { | |
| "f1": "81.4", | |
| "prec": "83.1(486/585)", | |
| "recall": "79.8(486/609)" | |
| }, | |
| "unsafe": { | |
| "f1": "74.5", | |
| "prec": "72.5(325/448)", | |
| "recall": "76.7(325/424)" | |
| } | |
| } | |
| }, | |
| "real": { | |
| "Accuracy": 77.8, | |
| "F1": 64.9, | |
| "unsafe": { | |
| "Accuracy": 77.8, | |
| "F1": 64.9, | |
| "safe": { | |
| "f1": "83.8", | |
| "prec": "79.3(582/734)", | |
| "recall": "88.9(582/655)" | |
| }, | |
| "unsafe": { | |
| "f1": "64.9", | |
| "prec": "74.0(208/281)", | |
| "recall": "57.8(208/360)" | |
| } | |
| } | |
| }, | |
| "step": 2016 | |
| }, | |
| { | |
| "Accuracy": 69.5, | |
| "Overall_f1": 82.0, | |
| "epoch": 2.0, | |
| "eval_SelfHarm_runtime": 80.8454, | |
| "eval_SelfHarm_samples_per_second": 7.916, | |
| "eval_SelfHarm_steps_per_second": 0.124, | |
| "generated": { | |
| "Accuracy": 0.0, | |
| "F1": 0.0, | |
| "unsafe": { | |
| "Accuracy": 0.0, | |
| "F1": 0.0, | |
| "safe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| }, | |
| "unsafe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| } | |
| } | |
| }, | |
| "real": { | |
| "Accuracy": 69.5, | |
| "F1": 82.0, | |
| "unsafe": { | |
| "Accuracy": 69.5, | |
| "F1": 82.0, | |
| "safe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/195)", | |
| "recall": "0.0(0/0)" | |
| }, | |
| "unsafe": { | |
| "f1": "82.0", | |
| "prec": "100.0(445/445)", | |
| "recall": "69.5(445/640)" | |
| } | |
| } | |
| }, | |
| "step": 2016 | |
| }, | |
| { | |
| "Accuracy": 79.8, | |
| "Overall_f1": 69.6, | |
| "epoch": 2.0, | |
| "eval_UnsafeDiff_runtime": 98.0101, | |
| "eval_UnsafeDiff_samples_per_second": 8.489, | |
| "eval_UnsafeDiff_steps_per_second": 0.133, | |
| "generated": { | |
| "Accuracy": 0.0, | |
| "F1": 0.0, | |
| "unsafe": { | |
| "Accuracy": 0.0, | |
| "F1": 0.0, | |
| "safe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| }, | |
| "unsafe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| } | |
| } | |
| }, | |
| "real": { | |
| "Accuracy": 79.8, | |
| "F1": 69.6, | |
| "unsafe": { | |
| "Accuracy": 79.8, | |
| "F1": 69.6, | |
| "safe": { | |
| "f1": "84.9", | |
| "prec": "93.5(471/504)", | |
| "recall": "77.7(471/606)" | |
| }, | |
| "unsafe": { | |
| "f1": "69.6", | |
| "prec": "58.8(193/328)", | |
| "recall": "85.4(193/226)" | |
| } | |
| } | |
| }, | |
| "step": 2016 | |
| }, | |
| { | |
| "Accuracy": 69.3, | |
| "Overall_f1": 81.9, | |
| "epoch": 2.0, | |
| "eval_ViolentBehavior_runtime": 23.6563, | |
| "eval_ViolentBehavior_samples_per_second": 8.116, | |
| "eval_ViolentBehavior_steps_per_second": 0.127, | |
| "generated": { | |
| "Accuracy": 0.0, | |
| "F1": 0.0, | |
| "unsafe": { | |
| "Accuracy": 0.0, | |
| "F1": 0.0, | |
| "safe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| }, | |
| "unsafe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| } | |
| } | |
| }, | |
| "real": { | |
| "Accuracy": 69.3, | |
| "F1": 81.9, | |
| "unsafe": { | |
| "Accuracy": 69.3, | |
| "F1": 81.9, | |
| "safe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/59)", | |
| "recall": "0.0(0/0)" | |
| }, | |
| "unsafe": { | |
| "f1": "81.9", | |
| "prec": "100.0(133/133)", | |
| "recall": "69.3(133/192)" | |
| } | |
| } | |
| }, | |
| "step": 2016 | |
| }, | |
| { | |
| "Accuracy": 78.9, | |
| "Overall_f1": 70.8, | |
| "epoch": 2.0, | |
| "eval_SMID_runtime": 190.3209, | |
| "eval_SMID_samples_per_second": 9.079, | |
| "eval_SMID_steps_per_second": 0.142, | |
| "generated": { | |
| "Accuracy": 0.0, | |
| "F1": 0.0, | |
| "unsafe": { | |
| "Accuracy": 0.0, | |
| "F1": 0.0, | |
| "safe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| }, | |
| "unsafe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| } | |
| } | |
| }, | |
| "real": { | |
| "Accuracy": 78.9, | |
| "F1": 70.8, | |
| "unsafe": { | |
| "Accuracy": 78.9, | |
| "F1": 70.8, | |
| "safe": { | |
| "f1": "83.5", | |
| "prec": "76.0(921/1212)", | |
| "recall": "92.6(921/995)" | |
| }, | |
| "unsafe": { | |
| "f1": "70.8", | |
| "prec": "85.7(442/516)", | |
| "recall": "60.3(442/733)" | |
| } | |
| } | |
| }, | |
| "step": 2016 | |
| }, | |
| { | |
| "Accuracy": 84.9, | |
| "Overall_f1": 85.5, | |
| "epoch": 2.0, | |
| "eval_JsonDataset_runtime": 408.3025, | |
| "eval_JsonDataset_samples_per_second": 4.859, | |
| "eval_JsonDataset_steps_per_second": 0.076, | |
| "generated": { | |
| "Accuracy": 87.6, | |
| "F1": 88.6, | |
| "fairness": { | |
| "Accuracy": 91.3, | |
| "F1": 91.3, | |
| "african": { | |
| "f1": "88.9", | |
| "prec": "82.4(28/34)", | |
| "recall": "96.6(28/29)" | |
| }, | |
| "asian": { | |
| "f1": "86.3", | |
| "prec": "91.4(117/128)", | |
| "recall": "81.8(117/143)" | |
| }, | |
| "caucasian": { | |
| "f1": "91.7", | |
| "prec": "88.8(284/320)", | |
| "recall": "94.7(284/300)" | |
| }, | |
| "children": { | |
| "f1": "75.0", | |
| "prec": "63.2(12/19)", | |
| "recall": "92.3(12/13)" | |
| }, | |
| "elderly": { | |
| "f1": "93.9", | |
| "prec": "96.4(108/112)", | |
| "recall": "91.5(108/118)" | |
| }, | |
| "female": { | |
| "f1": "99.4", | |
| "prec": "98.8(238/241)", | |
| "recall": "100.0(238/238)" | |
| }, | |
| "indian": { | |
| "f1": "70.6", | |
| "prec": "75.0(6/8)", | |
| "recall": "66.7(6/9)" | |
| }, | |
| "latino": { | |
| "f1": "45.3", | |
| "prec": "54.5(12/22)", | |
| "recall": "38.7(12/31)" | |
| }, | |
| "male": { | |
| "f1": "99.4", | |
| "prec": "100.0(271/271)", | |
| "recall": "98.9(271/274)" | |
| }, | |
| "middle-aged": { | |
| "f1": "80.3", | |
| "prec": "89.2(116/130)", | |
| "recall": "73.0(116/159)" | |
| }, | |
| "safe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| }, | |
| "teenager": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| }, | |
| "young adult": { | |
| "f1": "89.2", | |
| "prec": "84.1(211/251)", | |
| "recall": "95.0(211/222)" | |
| } | |
| }, | |
| "privacy": { | |
| "Accuracy": 83.6, | |
| "F1": 86.7, | |
| "intellectual property violation": { | |
| "f1": "85.7", | |
| "prec": "90.0(36/40)", | |
| "recall": "81.8(36/44)" | |
| }, | |
| "personal identification documents": { | |
| "f1": "90.5", | |
| "prec": "95.6(43/45)", | |
| "recall": "86.0(43/50)" | |
| }, | |
| "public figures": { | |
| "f1": "83.5", | |
| "prec": "82.6(38/46)", | |
| "recall": "84.4(38/45)" | |
| }, | |
| "safe": { | |
| "f1": "75.9", | |
| "prec": "70.7(41/58)", | |
| "recall": "82.0(41/50)" | |
| } | |
| }, | |
| "toxicity": { | |
| "Accuracy": 69.4, | |
| "F1": 70.3, | |
| "disturbing": { | |
| "f1": "72.0", | |
| "prec": "61.0(36/59)", | |
| "recall": "87.8(36/41)" | |
| }, | |
| "hate": { | |
| "f1": "50.0", | |
| "prec": "100.0(5/5)", | |
| "recall": "33.3(5/15)" | |
| }, | |
| "humiliation": { | |
| "f1": "43.4", | |
| "prec": "92.9(13/14)", | |
| "recall": "28.3(13/46)" | |
| }, | |
| "illegal activity": { | |
| "f1": "69.8", | |
| "prec": "88.0(22/25)", | |
| "recall": "57.9(22/38)" | |
| }, | |
| "safe": { | |
| "f1": "66.1", | |
| "prec": "55.7(39/70)", | |
| "recall": "81.2(39/48)" | |
| }, | |
| "sexual": { | |
| "f1": "96.8", | |
| "prec": "100.0(45/45)", | |
| "recall": "93.8(45/48)" | |
| }, | |
| "violence": { | |
| "f1": "63.6", | |
| "prec": "52.8(28/53)", | |
| "recall": "80.0(28/35)" | |
| } | |
| } | |
| }, | |
| "real": { | |
| "Accuracy": 82.2, | |
| "F1": 82.4, | |
| "fairness": { | |
| "Accuracy": 81.6, | |
| "F1": 81.6, | |
| "african": { | |
| "f1": "81.6", | |
| "prec": "85.7(60/70)", | |
| "recall": "77.9(60/77)" | |
| }, | |
| "asian": { | |
| "f1": "76.9", | |
| "prec": "76.9(70/91)", | |
| "recall": "76.9(70/91)" | |
| }, | |
| "caucasian": { | |
| "f1": "81.6", | |
| "prec": "74.7(162/217)", | |
| "recall": "90.0(162/180)" | |
| }, | |
| "children": { | |
| "f1": "81.3", | |
| "prec": "74.3(26/35)", | |
| "recall": "89.7(26/29)" | |
| }, | |
| "elderly": { | |
| "f1": "68.1", | |
| "prec": "75.6(31/41)", | |
| "recall": "62.0(31/50)" | |
| }, | |
| "female": { | |
| "f1": "94.6", | |
| "prec": "98.0(149/152)", | |
| "recall": "91.4(149/163)" | |
| }, | |
| "indian": { | |
| "f1": "62.4", | |
| "prec": "65.0(39/60)", | |
| "recall": "60.0(39/65)" | |
| }, | |
| "latino": { | |
| "f1": "35.3", | |
| "prec": "42.6(26/61)", | |
| "recall": "30.2(26/86)" | |
| }, | |
| "male": { | |
| "f1": "97.5", | |
| "prec": "96.0(333/347)", | |
| "recall": "99.1(333/336)" | |
| }, | |
| "middle-aged": { | |
| "f1": "80.0", | |
| "prec": "83.1(212/255)", | |
| "recall": "77.1(212/275)" | |
| }, | |
| "safe": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| }, | |
| "teenager": { | |
| "f1": "0.0", | |
| "prec": "0.0(0/0)", | |
| "recall": "0.0(0/0)" | |
| }, | |
| "young adult": { | |
| "f1": "72.2", | |
| "prec": "67.3(113/168)", | |
| "recall": "77.9(113/145)" | |
| } | |
| }, | |
| "privacy": { | |
| "Accuracy": 87.0, | |
| "F1": 90.6, | |
| "intellectual property violation": { | |
| "f1": "88.1", | |
| "prec": "100.0(37/37)", | |
| "recall": "78.7(37/47)" | |
| }, | |
| "personal identification documents": { | |
| "f1": "93.1", | |
| "prec": "92.2(47/51)", | |
| "recall": "94.0(47/50)" | |
| }, | |
| "public figures": { | |
| "f1": "90.1", | |
| "prec": "95.3(41/43)", | |
| "recall": "85.4(41/48)" | |
| }, | |
| "safe": { | |
| "f1": "78.2", | |
| "prec": "69.4(43/62)", | |
| "recall": "89.6(43/48)" | |
| } | |
| }, | |
| "toxicity": { | |
| "Accuracy": 82.2, | |
| "F1": 83.3, | |
| "disturbing": { | |
| "f1": "84.8", | |
| "prec": "84.0(42/50)", | |
| "recall": "85.7(42/49)" | |
| }, | |
| "hate": { | |
| "f1": "69.6", | |
| "prec": "72.7(8/11)", | |
| "recall": "66.7(8/12)" | |
| }, | |
| "humiliation": { | |
| "f1": "30.0", | |
| "prec": "50.0(3/6)", | |
| "recall": "21.4(3/14)" | |
| }, | |
| "illegal activity": { | |
| "f1": "88.2", | |
| "prec": "92.9(26/28)", | |
| "recall": "83.9(26/31)" | |
| }, | |
| "safe": { | |
| "f1": "78.1", | |
| "prec": "69.5(41/59)", | |
| "recall": "89.1(41/46)" | |
| }, | |
| "sexual": { | |
| "f1": "95.0", | |
| "prec": "94.1(48/51)", | |
| "recall": "96.0(48/50)" | |
| }, | |
| "violence": { | |
| "f1": "80.5", | |
| "prec": "83.3(35/42)", | |
| "recall": "77.8(35/45)" | |
| } | |
| } | |
| }, | |
| "step": 2016 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "step": 2016, | |
| "total_flos": 0.0, | |
| "train_loss": 0.08641785344788952, | |
| "train_runtime": 25197.0435, | |
| "train_samples_per_second": 5.118, | |
| "train_steps_per_second": 0.08 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 2016, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 0.0, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |