|
{"current_steps": 10, "total_steps": 54, "loss": 2.3587, "learning_rate": 4.588719528532342e-05, "epoch": 0.5555555555555556, "percentage": 18.52, "elapsed_time": "0:00:54", "remaining_time": "0:03:57", "throughput": "1515.65", "total_tokens": 81920}
|
|
{"current_steps": 20, "total_steps": 54, "loss": 2.3192, "learning_rate": 3.490199415097892e-05, "epoch": 1.1111111111111112, "percentage": 37.04, "elapsed_time": "0:01:41", "remaining_time": "0:02:53", "throughput": "1607.01", "total_tokens": 163840}
|
|
{"current_steps": 10, "total_steps": 126, "loss": 2.359, "learning_rate": 4.922693215572695e-05, "epoch": 0.5555555555555556, "percentage": 7.94, "elapsed_time": "0:00:49", "remaining_time": "0:09:31", "throughput": "1661.45", "total_tokens": 81920}
|
|
{"current_steps": 20, "total_steps": 126, "loss": 2.3166, "learning_rate": 4.6955539334255716e-05, "epoch": 1.1111111111111112, "percentage": 15.87, "elapsed_time": "0:01:37", "remaining_time": "0:08:36", "throughput": "1682.36", "total_tokens": 163840}
|
|
{"current_steps": 30, "total_steps": 126, "loss": 2.2205, "learning_rate": 4.332629679574566e-05, "epoch": 1.6666666666666665, "percentage": 23.81, "elapsed_time": "0:02:25", "remaining_time": "0:07:45", "throughput": "1687.87", "total_tokens": 245760}
|
|
{"current_steps": 40, "total_steps": 126, "loss": 2.159, "learning_rate": 3.856365659664399e-05, "epoch": 2.2222222222222223, "percentage": 31.75, "elapsed_time": "0:03:13", "remaining_time": "0:06:56", "throughput": "1693.37", "total_tokens": 327680}
|
|
{"current_steps": 50, "total_steps": 126, "loss": 2.0625, "learning_rate": 3.2962166256292113e-05, "epoch": 2.7777777777777777, "percentage": 39.68, "elapsed_time": "0:04:01", "remaining_time": "0:06:06", "throughput": "1698.47", "total_tokens": 409600}
|
|
{"current_steps": 60, "total_steps": 126, "loss": 2.0294, "learning_rate": 2.686825233966061e-05, "epoch": 3.3333333333333335, "percentage": 47.62, "elapsed_time": "0:04:48", "remaining_time": "0:05:17", "throughput": "1701.34", "total_tokens": 491520}
|
|
{"current_steps": 70, "total_steps": 126, "loss": 1.9556, "learning_rate": 2.0658795558326743e-05, "epoch": 3.888888888888889, "percentage": 55.56, "elapsed_time": "0:05:37", "remaining_time": "0:04:29", "throughput": "1700.17", "total_tokens": 573440}
|
|
{"current_steps": 80, "total_steps": 126, "loss": 1.8462, "learning_rate": 1.4717822421734718e-05, "epoch": 4.444444444444445, "percentage": 63.49, "elapsed_time": "0:06:24", "remaining_time": "0:03:41", "throughput": "1702.75", "total_tokens": 655360}
|
|
{"current_steps": 90, "total_steps": 126, "loss": 1.8544, "learning_rate": 9.412754953531663e-06, "epoch": 5.0, "percentage": 71.43, "elapsed_time": "0:07:12", "remaining_time": "0:02:53", "throughput": "1702.96", "total_tokens": 737280}
|
|
{"current_steps": 100, "total_steps": 126, "loss": 1.7957, "learning_rate": 5.071687319426946e-06, "epoch": 5.555555555555555, "percentage": 79.37, "elapsed_time": "0:08:01", "remaining_time": "0:02:05", "throughput": "1702.82", "total_tokens": 819200}
|
|
{"current_steps": 110, "total_steps": 126, "loss": 1.7244, "learning_rate": 1.9630947032398067e-06, "epoch": 6.111111111111111, "percentage": 87.3, "elapsed_time": "0:08:50", "remaining_time": "0:01:17", "throughput": "1697.22", "total_tokens": 901120}
|
|
{"current_steps": 120, "total_steps": 126, "loss": 1.7368, "learning_rate": 2.7922934437178695e-07, "epoch": 6.666666666666667, "percentage": 95.24, "elapsed_time": "0:09:41", "remaining_time": "0:00:29", "throughput": "1691.34", "total_tokens": 983040}
|
|
{"current_steps": 126, "total_steps": 126, "epoch": 7.0, "percentage": 100.0, "elapsed_time": "0:10:12", "remaining_time": "0:00:00", "throughput": "1684.82", "total_tokens": 1032192}
|
|
|