|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 8842, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0565482922415743, |
|
"grad_norm": 6.099003791809082, |
|
"learning_rate": 4.7172585387921285e-05, |
|
"loss": 5.2788, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.1130965844831486, |
|
"grad_norm": 4.438333988189697, |
|
"learning_rate": 4.434517077584257e-05, |
|
"loss": 4.8237, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.16964487672472292, |
|
"grad_norm": 5.307681083679199, |
|
"learning_rate": 4.1517756163763856e-05, |
|
"loss": 4.737, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.2261931689662972, |
|
"grad_norm": 5.611371040344238, |
|
"learning_rate": 3.869034155168514e-05, |
|
"loss": 4.6918, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.2827414612078715, |
|
"grad_norm": 5.551920413970947, |
|
"learning_rate": 3.586292693960643e-05, |
|
"loss": 4.6628, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.33928975344944584, |
|
"grad_norm": 5.272426605224609, |
|
"learning_rate": 3.303551232752771e-05, |
|
"loss": 4.6474, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.39583804569102016, |
|
"grad_norm": 5.137388229370117, |
|
"learning_rate": 3.0208097715448996e-05, |
|
"loss": 4.6272, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.4523863379325944, |
|
"grad_norm": 6.018069744110107, |
|
"learning_rate": 2.7380683103370282e-05, |
|
"loss": 4.6025, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.5089346301741687, |
|
"grad_norm": 5.987523555755615, |
|
"learning_rate": 2.4553268491291564e-05, |
|
"loss": 4.5791, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.565482922415743, |
|
"grad_norm": 5.519354820251465, |
|
"learning_rate": 2.172585387921285e-05, |
|
"loss": 4.5844, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.6220312146573174, |
|
"grad_norm": 5.536919593811035, |
|
"learning_rate": 1.8898439267134132e-05, |
|
"loss": 4.5647, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.6785795068988917, |
|
"grad_norm": 5.017457962036133, |
|
"learning_rate": 1.6071024655055418e-05, |
|
"loss": 4.5574, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.735127799140466, |
|
"grad_norm": 5.590214252471924, |
|
"learning_rate": 1.3243610042976704e-05, |
|
"loss": 4.548, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.7916760913820403, |
|
"grad_norm": 4.2981181144714355, |
|
"learning_rate": 1.0416195430897988e-05, |
|
"loss": 4.5641, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.8482243836236145, |
|
"grad_norm": 5.723196983337402, |
|
"learning_rate": 7.588780818819271e-06, |
|
"loss": 4.5633, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.9047726758651888, |
|
"grad_norm": 5.700985431671143, |
|
"learning_rate": 4.761366206740556e-06, |
|
"loss": 4.5486, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.9613209681067632, |
|
"grad_norm": 5.034333229064941, |
|
"learning_rate": 1.9339515946618415e-06, |
|
"loss": 4.5307, |
|
"step": 8500 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 8842, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 5000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.3333874049690624e+17, |
|
"train_batch_size": 32, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|