| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.9929627023223082, |
| "eval_steps": 500, |
| "global_step": 236, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.00844475721323012, |
| "grad_norm": 2.089867158089946, |
| "learning_rate": 4.1666666666666667e-07, |
| "loss": 2.161, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.0422237860661506, |
| "grad_norm": 1.5848399731293203, |
| "learning_rate": 2.0833333333333334e-06, |
| "loss": 2.0465, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.0844475721323012, |
| "grad_norm": 0.7578527824638689, |
| "learning_rate": 4.166666666666667e-06, |
| "loss": 1.5984, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.1266713581984518, |
| "grad_norm": 0.346915259777201, |
| "learning_rate": 6.25e-06, |
| "loss": 0.9384, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.1688951442646024, |
| "grad_norm": 0.17556784814479495, |
| "learning_rate": 8.333333333333334e-06, |
| "loss": 0.7181, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.211118930330753, |
| "grad_norm": 0.13630834383254564, |
| "learning_rate": 9.999451015497595e-06, |
| "loss": 0.6024, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.2533427163969036, |
| "grad_norm": 0.11554222320406965, |
| "learning_rate": 9.980249213076085e-06, |
| "loss": 0.5901, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.2955665024630542, |
| "grad_norm": 0.12278253986244328, |
| "learning_rate": 9.933718620186745e-06, |
| "loss": 0.5266, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.3377902885292048, |
| "grad_norm": 0.09979826087347965, |
| "learning_rate": 9.860114570402055e-06, |
| "loss": 0.4992, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.3800140745953554, |
| "grad_norm": 0.09403541146186689, |
| "learning_rate": 9.759840961111098e-06, |
| "loss": 0.4552, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.422237860661506, |
| "grad_norm": 0.11355916833002154, |
| "learning_rate": 9.633448037159167e-06, |
| "loss": 0.4481, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.4644616467276566, |
| "grad_norm": 0.08073377113874591, |
| "learning_rate": 9.481629371415315e-06, |
| "loss": 0.4421, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.5066854327938072, |
| "grad_norm": 0.09267587470901371, |
| "learning_rate": 9.305218058836778e-06, |
| "loss": 0.4315, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.5489092188599578, |
| "grad_norm": 0.07970275638417536, |
| "learning_rate": 9.10518214491513e-06, |
| "loss": 0.4193, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.5911330049261084, |
| "grad_norm": 0.09176650758741302, |
| "learning_rate": 8.882619313590212e-06, |
| "loss": 0.4359, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.633356790992259, |
| "grad_norm": 0.06297292993950475, |
| "learning_rate": 8.638750863781614e-06, |
| "loss": 0.3866, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.6755805770584096, |
| "grad_norm": 0.09109073826112205, |
| "learning_rate": 8.374915007591053e-06, |
| "loss": 0.4041, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.7178043631245602, |
| "grad_norm": 0.05865190056091487, |
| "learning_rate": 8.092559526951374e-06, |
| "loss": 0.4239, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.7600281491907108, |
| "grad_norm": 0.06974442479960637, |
| "learning_rate": 7.793233829018263e-06, |
| "loss": 0.3986, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.8022519352568613, |
| "grad_norm": 0.05463034337178698, |
| "learning_rate": 7.478580443900247e-06, |
| "loss": 0.4247, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.844475721323012, |
| "grad_norm": 0.06205362965112395, |
| "learning_rate": 7.1503260113826035e-06, |
| "loss": 0.3913, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.8866995073891626, |
| "grad_norm": 0.067892476771574, |
| "learning_rate": 6.810271806104931e-06, |
| "loss": 0.401, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.9289232934553132, |
| "grad_norm": 0.05154900712833624, |
| "learning_rate": 6.46028385318488e-06, |
| "loss": 0.3938, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.9711470795214637, |
| "grad_norm": 0.05021016889651189, |
| "learning_rate": 6.10228268852786e-06, |
| "loss": 0.3675, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.9964813511611541, |
| "eval_loss": 0.36927565932273865, |
| "eval_runtime": 345.1421, |
| "eval_samples_per_second": 4.117, |
| "eval_steps_per_second": 1.031, |
| "step": 118 |
| }, |
| { |
| "epoch": 1.0133708655876144, |
| "grad_norm": 0.06461727535494917, |
| "learning_rate": 5.738232820012407e-06, |
| "loss": 0.437, |
| "step": 120 |
| }, |
| { |
| "epoch": 1.0555946516537649, |
| "grad_norm": 0.05489917224274474, |
| "learning_rate": 5.370131947382215e-06, |
| "loss": 0.3656, |
| "step": 125 |
| }, |
| { |
| "epoch": 1.0978184377199156, |
| "grad_norm": 0.06042558978858904, |
| "learning_rate": 5e-06, |
| "loss": 0.4009, |
| "step": 130 |
| }, |
| { |
| "epoch": 1.1400422237860661, |
| "grad_norm": 0.06049921311472245, |
| "learning_rate": 4.629868052617786e-06, |
| "loss": 0.3482, |
| "step": 135 |
| }, |
| { |
| "epoch": 1.1822660098522166, |
| "grad_norm": 0.053628889597754824, |
| "learning_rate": 4.261767179987595e-06, |
| "loss": 0.3484, |
| "step": 140 |
| }, |
| { |
| "epoch": 1.2244897959183674, |
| "grad_norm": 0.05427865015224349, |
| "learning_rate": 3.897717311472141e-06, |
| "loss": 0.3639, |
| "step": 145 |
| }, |
| { |
| "epoch": 1.266713581984518, |
| "grad_norm": 0.057489395038740544, |
| "learning_rate": 3.539716146815122e-06, |
| "loss": 0.3835, |
| "step": 150 |
| }, |
| { |
| "epoch": 1.3089373680506686, |
| "grad_norm": 0.05111391290442162, |
| "learning_rate": 3.1897281938950693e-06, |
| "loss": 0.3645, |
| "step": 155 |
| }, |
| { |
| "epoch": 1.3511611541168191, |
| "grad_norm": 0.05075774897048314, |
| "learning_rate": 2.8496739886173994e-06, |
| "loss": 0.3693, |
| "step": 160 |
| }, |
| { |
| "epoch": 1.3933849401829699, |
| "grad_norm": 0.04943651400266673, |
| "learning_rate": 2.5214195560997546e-06, |
| "loss": 0.3786, |
| "step": 165 |
| }, |
| { |
| "epoch": 1.4356087262491204, |
| "grad_norm": 0.044263408717175816, |
| "learning_rate": 2.2067661709817384e-06, |
| "loss": 0.3802, |
| "step": 170 |
| }, |
| { |
| "epoch": 1.477832512315271, |
| "grad_norm": 0.05037201642103367, |
| "learning_rate": 1.9074404730486264e-06, |
| "loss": 0.3548, |
| "step": 175 |
| }, |
| { |
| "epoch": 1.5200562983814216, |
| "grad_norm": 0.047573069715430114, |
| "learning_rate": 1.6250849924089485e-06, |
| "loss": 0.3737, |
| "step": 180 |
| }, |
| { |
| "epoch": 1.5622800844475722, |
| "grad_norm": 0.05285111504365142, |
| "learning_rate": 1.3612491362183887e-06, |
| "loss": 0.3602, |
| "step": 185 |
| }, |
| { |
| "epoch": 1.6045038705137227, |
| "grad_norm": 0.05156216290861093, |
| "learning_rate": 1.1173806864097885e-06, |
| "loss": 0.3609, |
| "step": 190 |
| }, |
| { |
| "epoch": 1.6467276565798734, |
| "grad_norm": 0.04756228232260825, |
| "learning_rate": 8.948178550848702e-07, |
| "loss": 0.3344, |
| "step": 195 |
| }, |
| { |
| "epoch": 1.688951442646024, |
| "grad_norm": 0.05015333465574718, |
| "learning_rate": 6.947819411632223e-07, |
| "loss": 0.326, |
| "step": 200 |
| }, |
| { |
| "epoch": 1.7311752287121744, |
| "grad_norm": 0.05391623766653661, |
| "learning_rate": 5.183706285846873e-07, |
| "loss": 0.3743, |
| "step": 205 |
| }, |
| { |
| "epoch": 1.7733990147783252, |
| "grad_norm": 0.05381208848410942, |
| "learning_rate": 3.665519628408332e-07, |
| "loss": 0.3482, |
| "step": 210 |
| }, |
| { |
| "epoch": 1.8156228008444757, |
| "grad_norm": 0.04072660534703765, |
| "learning_rate": 2.401590388889025e-07, |
| "loss": 0.3681, |
| "step": 215 |
| }, |
| { |
| "epoch": 1.8578465869106262, |
| "grad_norm": 0.05078699788677098, |
| "learning_rate": 1.3988542959794627e-07, |
| "loss": 0.3522, |
| "step": 220 |
| }, |
| { |
| "epoch": 1.900070372976777, |
| "grad_norm": 0.05235725304027692, |
| "learning_rate": 6.628137981325611e-08, |
| "loss": 0.3994, |
| "step": 225 |
| }, |
| { |
| "epoch": 1.9422941590429277, |
| "grad_norm": 0.04983097587956561, |
| "learning_rate": 1.975078692391552e-08, |
| "loss": 0.36, |
| "step": 230 |
| }, |
| { |
| "epoch": 1.984517945109078, |
| "grad_norm": 0.058131705794752526, |
| "learning_rate": 5.489845024053698e-10, |
| "loss": 0.3467, |
| "step": 235 |
| }, |
| { |
| "epoch": 1.9929627023223082, |
| "eval_loss": 0.3553777039051056, |
| "eval_runtime": 342.1435, |
| "eval_samples_per_second": 4.153, |
| "eval_steps_per_second": 1.04, |
| "step": 236 |
| }, |
| { |
| "epoch": 1.9929627023223082, |
| "step": 236, |
| "total_flos": 4.237289220525261e+16, |
| "train_loss": 0.47937014673726036, |
| "train_runtime": 10453.3911, |
| "train_samples_per_second": 1.087, |
| "train_steps_per_second": 0.023 |
| } |
| ], |
| "logging_steps": 5, |
| "max_steps": 236, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 2, |
| "save_steps": 100, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 4.237289220525261e+16, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|