| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 5.0, | |
| "eval_steps": 5000, | |
| "global_step": 87895, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.05688605722737357, | |
| "grad_norm": 2.4062280654907227, | |
| "learning_rate": 0.0007908982308436202, | |
| "loss": 2.1918, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.11377211445474714, | |
| "grad_norm": 1.431848406791687, | |
| "learning_rate": 0.0007817964616872405, | |
| "loss": 1.4818, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.17065817168212072, | |
| "grad_norm": 1.5747077465057373, | |
| "learning_rate": 0.0007726946925308607, | |
| "loss": 1.3634, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.22754422890949427, | |
| "grad_norm": 1.4864206314086914, | |
| "learning_rate": 0.0007635929233744809, | |
| "loss": 1.2967, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.2844302861368678, | |
| "grad_norm": 1.2000905275344849, | |
| "learning_rate": 0.0007544911542181011, | |
| "loss": 1.2574, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.2844302861368678, | |
| "eval_accuracy": 0.689128, | |
| "eval_loss": 1.238457202911377, | |
| "eval_runtime": 203.0197, | |
| "eval_samples_per_second": 1231.407, | |
| "eval_steps_per_second": 4.812, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.34131634336424144, | |
| "grad_norm": 1.2910780906677246, | |
| "learning_rate": 0.0007453893850617214, | |
| "loss": 1.2181, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.398202400591615, | |
| "grad_norm": 1.1383774280548096, | |
| "learning_rate": 0.0007362876159053416, | |
| "loss": 1.1863, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.45508845781898855, | |
| "grad_norm": 1.135689616203308, | |
| "learning_rate": 0.0007271858467489618, | |
| "loss": 1.1653, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.5119745150463622, | |
| "grad_norm": 1.1965036392211914, | |
| "learning_rate": 0.0007180840775925821, | |
| "loss": 1.147, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.5688605722737357, | |
| "grad_norm": 1.0561026334762573, | |
| "learning_rate": 0.0007089823084362024, | |
| "loss": 1.1281, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.5688605722737357, | |
| "eval_accuracy": 0.715764, | |
| "eval_loss": 1.1192152500152588, | |
| "eval_runtime": 128.8781, | |
| "eval_samples_per_second": 1939.817, | |
| "eval_steps_per_second": 7.581, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.6257466295011093, | |
| "grad_norm": 0.9711835980415344, | |
| "learning_rate": 0.0006998805392798226, | |
| "loss": 1.1232, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.6826326867284829, | |
| "grad_norm": 0.8913602828979492, | |
| "learning_rate": 0.0006907787701234428, | |
| "loss": 1.0988, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.7395187439558564, | |
| "grad_norm": 1.092698097229004, | |
| "learning_rate": 0.000681677000967063, | |
| "loss": 1.0897, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.79640480118323, | |
| "grad_norm": 0.9319038391113281, | |
| "learning_rate": 0.0006725752318106833, | |
| "loss": 1.0826, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.8532908584106036, | |
| "grad_norm": 1.0223675966262817, | |
| "learning_rate": 0.0006634734626543035, | |
| "loss": 1.0698, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.8532908584106036, | |
| "eval_accuracy": 0.728676, | |
| "eval_loss": 1.0653605461120605, | |
| "eval_runtime": 128.0826, | |
| "eval_samples_per_second": 1951.866, | |
| "eval_steps_per_second": 7.628, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.9101769156379771, | |
| "grad_norm": 0.8995338678359985, | |
| "learning_rate": 0.0006543716934979237, | |
| "loss": 1.0624, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.9670629728653507, | |
| "grad_norm": 0.8418471217155457, | |
| "learning_rate": 0.0006452699243415439, | |
| "loss": 1.0538, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 1.0239490300927243, | |
| "grad_norm": 1.024624228477478, | |
| "learning_rate": 0.0006361681551851641, | |
| "loss": 1.0311, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 1.0808350873200978, | |
| "grad_norm": 0.9130891561508179, | |
| "learning_rate": 0.0006270663860287844, | |
| "loss": 0.999, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 1.1377211445474713, | |
| "grad_norm": 0.8896342515945435, | |
| "learning_rate": 0.0006179646168724045, | |
| "loss": 1.0, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 1.1377211445474713, | |
| "eval_accuracy": 0.739712, | |
| "eval_loss": 1.0235533714294434, | |
| "eval_runtime": 127.2585, | |
| "eval_samples_per_second": 1964.505, | |
| "eval_steps_per_second": 7.677, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 1.194607201774845, | |
| "grad_norm": 0.7940112948417664, | |
| "learning_rate": 0.0006088628477160248, | |
| "loss": 0.9957, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 1.2514932590022185, | |
| "grad_norm": 0.9015308618545532, | |
| "learning_rate": 0.000599761078559645, | |
| "loss": 0.9967, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 1.3083793162295922, | |
| "grad_norm": 0.9106078147888184, | |
| "learning_rate": 0.0005906593094032653, | |
| "loss": 0.9939, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 1.3652653734569657, | |
| "grad_norm": 0.9563422203063965, | |
| "learning_rate": 0.0005815575402468854, | |
| "loss": 0.9931, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 1.4221514306843392, | |
| "grad_norm": 0.7646272778511047, | |
| "learning_rate": 0.0005724557710905057, | |
| "loss": 0.9774, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 1.4221514306843392, | |
| "eval_accuracy": 0.743348, | |
| "eval_loss": 1.0054922103881836, | |
| "eval_runtime": 127.7729, | |
| "eval_samples_per_second": 1956.596, | |
| "eval_steps_per_second": 7.646, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 1.4790374879117127, | |
| "grad_norm": 0.7779045104980469, | |
| "learning_rate": 0.000563354001934126, | |
| "loss": 0.9792, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 1.5359235451390862, | |
| "grad_norm": 0.8506484627723694, | |
| "learning_rate": 0.0005542522327777463, | |
| "loss": 0.9778, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 1.59280960236646, | |
| "grad_norm": 0.8443676829338074, | |
| "learning_rate": 0.0005451504636213664, | |
| "loss": 0.9715, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 1.6496956595938337, | |
| "grad_norm": 0.9333568215370178, | |
| "learning_rate": 0.0005360486944649867, | |
| "loss": 0.9679, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 1.7065817168212072, | |
| "grad_norm": 0.9501623511314392, | |
| "learning_rate": 0.0005269469253086069, | |
| "loss": 0.9684, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 1.7065817168212072, | |
| "eval_accuracy": 0.749276, | |
| "eval_loss": 0.9812818765640259, | |
| "eval_runtime": 128.5758, | |
| "eval_samples_per_second": 1944.379, | |
| "eval_steps_per_second": 7.599, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 1.7634677740485807, | |
| "grad_norm": 0.7442188262939453, | |
| "learning_rate": 0.0005178451561522272, | |
| "loss": 0.9636, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 1.8203538312759542, | |
| "grad_norm": 0.7510819435119629, | |
| "learning_rate": 0.0005087433869958473, | |
| "loss": 0.9647, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 1.8772398885033277, | |
| "grad_norm": 0.7448764443397522, | |
| "learning_rate": 0.0004996416178394676, | |
| "loss": 0.9591, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 1.9341259457307014, | |
| "grad_norm": 0.8019358515739441, | |
| "learning_rate": 0.0004905398486830878, | |
| "loss": 0.9513, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 1.9910120029580751, | |
| "grad_norm": 0.9495121240615845, | |
| "learning_rate": 0.00048143807952670797, | |
| "loss": 0.9511, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 1.9910120029580751, | |
| "eval_accuracy": 0.755448, | |
| "eval_loss": 0.9558805227279663, | |
| "eval_runtime": 127.8711, | |
| "eval_samples_per_second": 1955.094, | |
| "eval_steps_per_second": 7.641, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 2.0478980601854486, | |
| "grad_norm": 0.8410281538963318, | |
| "learning_rate": 0.00047233631037032825, | |
| "loss": 0.9081, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 2.104784117412822, | |
| "grad_norm": 0.8246123194694519, | |
| "learning_rate": 0.00046323454121394847, | |
| "loss": 0.8964, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 2.1616701746401956, | |
| "grad_norm": 0.9567108154296875, | |
| "learning_rate": 0.0004541327720575687, | |
| "loss": 0.8952, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 2.218556231867569, | |
| "grad_norm": 0.8104901313781738, | |
| "learning_rate": 0.0004450310029011889, | |
| "loss": 0.8925, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 2.2754422890949426, | |
| "grad_norm": 0.9034276008605957, | |
| "learning_rate": 0.0004359292337448092, | |
| "loss": 0.8998, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 2.2754422890949426, | |
| "eval_accuracy": 0.755948, | |
| "eval_loss": 0.9492226839065552, | |
| "eval_runtime": 127.8812, | |
| "eval_samples_per_second": 1954.94, | |
| "eval_steps_per_second": 7.64, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 2.3323283463223166, | |
| "grad_norm": 1.3229442834854126, | |
| "learning_rate": 0.00042682746458842937, | |
| "loss": 0.8962, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 2.38921440354969, | |
| "grad_norm": 0.8582925200462341, | |
| "learning_rate": 0.00041772569543204965, | |
| "loss": 0.8976, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 2.4461004607770636, | |
| "grad_norm": 0.8881712555885315, | |
| "learning_rate": 0.0004086239262756698, | |
| "loss": 0.8898, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 2.502986518004437, | |
| "grad_norm": 0.8713961839675903, | |
| "learning_rate": 0.00039952215711929005, | |
| "loss": 0.8927, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 2.5598725752318106, | |
| "grad_norm": 0.7883007526397705, | |
| "learning_rate": 0.00039042038796291027, | |
| "loss": 0.8967, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 2.5598725752318106, | |
| "eval_accuracy": 0.760028, | |
| "eval_loss": 0.937300980091095, | |
| "eval_runtime": 130.0782, | |
| "eval_samples_per_second": 1921.921, | |
| "eval_steps_per_second": 7.511, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 2.6167586324591845, | |
| "grad_norm": 0.8600155711174011, | |
| "learning_rate": 0.00038131861880653055, | |
| "loss": 0.8927, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 2.673644689686558, | |
| "grad_norm": 0.8501909971237183, | |
| "learning_rate": 0.0003722168496501508, | |
| "loss": 0.8913, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 2.7305307469139315, | |
| "grad_norm": 0.8116582632064819, | |
| "learning_rate": 0.000363115080493771, | |
| "loss": 0.8889, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 2.787416804141305, | |
| "grad_norm": 0.8065186738967896, | |
| "learning_rate": 0.0003540133113373912, | |
| "loss": 0.8896, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 2.8443028613686785, | |
| "grad_norm": 0.9248031973838806, | |
| "learning_rate": 0.00034491154218101145, | |
| "loss": 0.8837, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 2.8443028613686785, | |
| "eval_accuracy": 0.762176, | |
| "eval_loss": 0.9251159429550171, | |
| "eval_runtime": 128.4439, | |
| "eval_samples_per_second": 1946.376, | |
| "eval_steps_per_second": 7.606, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 2.901188918596052, | |
| "grad_norm": 0.8191467523574829, | |
| "learning_rate": 0.0003358097730246317, | |
| "loss": 0.878, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 2.9580749758234255, | |
| "grad_norm": 0.7620063424110413, | |
| "learning_rate": 0.0003267080038682519, | |
| "loss": 0.8832, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 3.0149610330507994, | |
| "grad_norm": 0.8365482687950134, | |
| "learning_rate": 0.0003176062347118721, | |
| "loss": 0.8621, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 3.071847090278173, | |
| "grad_norm": 0.9817807078361511, | |
| "learning_rate": 0.00030850446555549235, | |
| "loss": 0.8224, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 3.1287331475055464, | |
| "grad_norm": 0.847806453704834, | |
| "learning_rate": 0.00029940269639911263, | |
| "loss": 0.8253, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 3.1287331475055464, | |
| "eval_accuracy": 0.76438, | |
| "eval_loss": 0.9235970973968506, | |
| "eval_runtime": 126.2531, | |
| "eval_samples_per_second": 1980.15, | |
| "eval_steps_per_second": 7.738, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 3.18561920473292, | |
| "grad_norm": 1.1729530096054077, | |
| "learning_rate": 0.00029030092724273285, | |
| "loss": 0.8225, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 3.2425052619602934, | |
| "grad_norm": 1.0548408031463623, | |
| "learning_rate": 0.0002811991580863531, | |
| "loss": 0.821, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 3.299391319187667, | |
| "grad_norm": 1.0199774503707886, | |
| "learning_rate": 0.0002720973889299733, | |
| "loss": 0.8213, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 3.356277376415041, | |
| "grad_norm": 0.9180177450180054, | |
| "learning_rate": 0.00026299561977359353, | |
| "loss": 0.8274, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 3.4131634336424144, | |
| "grad_norm": 0.9745663404464722, | |
| "learning_rate": 0.0002538938506172137, | |
| "loss": 0.8229, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 3.4131634336424144, | |
| "eval_accuracy": 0.766832, | |
| "eval_loss": 0.9138370156288147, | |
| "eval_runtime": 129.2727, | |
| "eval_samples_per_second": 1933.897, | |
| "eval_steps_per_second": 7.558, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 3.470049490869788, | |
| "grad_norm": 0.8708947896957397, | |
| "learning_rate": 0.0002447920814608339, | |
| "loss": 0.8256, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 3.5269355480971614, | |
| "grad_norm": 0.9808185696601868, | |
| "learning_rate": 0.00023569031230445418, | |
| "loss": 0.8298, | |
| "step": 62000 | |
| }, | |
| { | |
| "epoch": 3.583821605324535, | |
| "grad_norm": 0.8228833079338074, | |
| "learning_rate": 0.0002265885431480744, | |
| "loss": 0.827, | |
| "step": 63000 | |
| }, | |
| { | |
| "epoch": 3.6407076625519084, | |
| "grad_norm": 0.9581019878387451, | |
| "learning_rate": 0.00021748677399169463, | |
| "loss": 0.8275, | |
| "step": 64000 | |
| }, | |
| { | |
| "epoch": 3.697593719779282, | |
| "grad_norm": 0.8560314178466797, | |
| "learning_rate": 0.00020838500483531488, | |
| "loss": 0.8145, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 3.697593719779282, | |
| "eval_accuracy": 0.769172, | |
| "eval_loss": 0.9042648673057556, | |
| "eval_runtime": 129.2138, | |
| "eval_samples_per_second": 1934.778, | |
| "eval_steps_per_second": 7.561, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 3.754479777006656, | |
| "grad_norm": 0.8918451070785522, | |
| "learning_rate": 0.0001992832356789351, | |
| "loss": 0.819, | |
| "step": 66000 | |
| }, | |
| { | |
| "epoch": 3.8113658342340293, | |
| "grad_norm": 1.0977294445037842, | |
| "learning_rate": 0.00019018146652255533, | |
| "loss": 0.8122, | |
| "step": 67000 | |
| }, | |
| { | |
| "epoch": 3.868251891461403, | |
| "grad_norm": 0.7856444716453552, | |
| "learning_rate": 0.00018107969736617555, | |
| "loss": 0.8225, | |
| "step": 68000 | |
| }, | |
| { | |
| "epoch": 3.9251379486887763, | |
| "grad_norm": 0.9270259141921997, | |
| "learning_rate": 0.00017197792820979578, | |
| "loss": 0.8158, | |
| "step": 69000 | |
| }, | |
| { | |
| "epoch": 3.98202400591615, | |
| "grad_norm": 1.082774043083191, | |
| "learning_rate": 0.00016287615905341603, | |
| "loss": 0.8156, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 3.98202400591615, | |
| "eval_accuracy": 0.770764, | |
| "eval_loss": 0.8961142301559448, | |
| "eval_runtime": 138.0555, | |
| "eval_samples_per_second": 1810.866, | |
| "eval_steps_per_second": 7.077, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 4.038910063143524, | |
| "grad_norm": 0.909858226776123, | |
| "learning_rate": 0.00015377438989703626, | |
| "loss": 0.7785, | |
| "step": 71000 | |
| }, | |
| { | |
| "epoch": 4.095796120370897, | |
| "grad_norm": 0.931280791759491, | |
| "learning_rate": 0.00014467262074065645, | |
| "loss": 0.7637, | |
| "step": 72000 | |
| }, | |
| { | |
| "epoch": 4.152682177598271, | |
| "grad_norm": 0.94422847032547, | |
| "learning_rate": 0.0001355708515842767, | |
| "loss": 0.7612, | |
| "step": 73000 | |
| }, | |
| { | |
| "epoch": 4.209568234825644, | |
| "grad_norm": 0.9250127077102661, | |
| "learning_rate": 0.00012646908242789693, | |
| "loss": 0.7616, | |
| "step": 74000 | |
| }, | |
| { | |
| "epoch": 4.266454292053018, | |
| "grad_norm": 0.8467296957969666, | |
| "learning_rate": 0.00011736731327151716, | |
| "loss": 0.7557, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 4.266454292053018, | |
| "eval_accuracy": 0.77204, | |
| "eval_loss": 0.9022773504257202, | |
| "eval_runtime": 144.5432, | |
| "eval_samples_per_second": 1729.587, | |
| "eval_steps_per_second": 6.759, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 4.323340349280391, | |
| "grad_norm": 0.8985564708709717, | |
| "learning_rate": 0.00010826554411513738, | |
| "loss": 0.7604, | |
| "step": 76000 | |
| }, | |
| { | |
| "epoch": 4.380226406507765, | |
| "grad_norm": 0.8618564605712891, | |
| "learning_rate": 9.916377495875762e-05, | |
| "loss": 0.7632, | |
| "step": 77000 | |
| }, | |
| { | |
| "epoch": 4.437112463735138, | |
| "grad_norm": 0.9467126727104187, | |
| "learning_rate": 9.006200580237784e-05, | |
| "loss": 0.7614, | |
| "step": 78000 | |
| }, | |
| { | |
| "epoch": 4.493998520962512, | |
| "grad_norm": 1.0163730382919312, | |
| "learning_rate": 8.096023664599807e-05, | |
| "loss": 0.7575, | |
| "step": 79000 | |
| }, | |
| { | |
| "epoch": 4.550884578189885, | |
| "grad_norm": 1.1194038391113281, | |
| "learning_rate": 7.18584674896183e-05, | |
| "loss": 0.7595, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 4.550884578189885, | |
| "eval_accuracy": 0.772256, | |
| "eval_loss": 0.897346019744873, | |
| "eval_runtime": 136.9434, | |
| "eval_samples_per_second": 1825.571, | |
| "eval_steps_per_second": 7.134, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 4.607770635417259, | |
| "grad_norm": 1.0589629411697388, | |
| "learning_rate": 6.275669833323853e-05, | |
| "loss": 0.7548, | |
| "step": 81000 | |
| }, | |
| { | |
| "epoch": 4.664656692644633, | |
| "grad_norm": 0.8540852665901184, | |
| "learning_rate": 5.365492917685876e-05, | |
| "loss": 0.7601, | |
| "step": 82000 | |
| }, | |
| { | |
| "epoch": 4.721542749872007, | |
| "grad_norm": 1.127475380897522, | |
| "learning_rate": 4.455316002047898e-05, | |
| "loss": 0.7554, | |
| "step": 83000 | |
| }, | |
| { | |
| "epoch": 4.77842880709938, | |
| "grad_norm": 0.9464063048362732, | |
| "learning_rate": 3.545139086409921e-05, | |
| "loss": 0.756, | |
| "step": 84000 | |
| }, | |
| { | |
| "epoch": 4.835314864326754, | |
| "grad_norm": 0.9705914855003357, | |
| "learning_rate": 2.634962170771944e-05, | |
| "loss": 0.7581, | |
| "step": 85000 | |
| }, | |
| { | |
| "epoch": 4.835314864326754, | |
| "eval_accuracy": 0.773724, | |
| "eval_loss": 0.8925997018814087, | |
| "eval_runtime": 138.7415, | |
| "eval_samples_per_second": 1801.913, | |
| "eval_steps_per_second": 7.042, | |
| "step": 85000 | |
| }, | |
| { | |
| "epoch": 4.892200921554127, | |
| "grad_norm": 0.8879310488700867, | |
| "learning_rate": 1.7247852551339668e-05, | |
| "loss": 0.758, | |
| "step": 86000 | |
| }, | |
| { | |
| "epoch": 4.949086978781501, | |
| "grad_norm": 1.2024400234222412, | |
| "learning_rate": 8.146083394959896e-06, | |
| "loss": 0.751, | |
| "step": 87000 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "step": 87895, | |
| "total_flos": 1.93274424e+18, | |
| "train_loss": 0.9357909288237652, | |
| "train_runtime": 45635.435, | |
| "train_samples_per_second": 493.038, | |
| "train_steps_per_second": 1.926 | |
| } | |
| ], | |
| "logging_steps": 1000, | |
| "max_steps": 87895, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 5, | |
| "save_steps": 5000, | |
| "total_flos": 1.93274424e+18, | |
| "train_batch_size": 256, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |