nlparabic commited on
Commit
2c1903c
·
verified ·
1 Parent(s): 842f545

End of training

Browse files
all_results.json CHANGED
@@ -1,19 +1,19 @@
1
  {
2
  "epoch": 3.0,
3
- "eval_bleu": 0.2510920700175718,
4
- "eval_loss": 2.5200693607330322,
5
- "eval_rouge1": 0.45339120041307046,
6
- "eval_rouge2": 0.21369562389429725,
7
- "eval_rougeL": 0.3983722007696566,
8
- "eval_runtime": 28.0888,
9
- "eval_samples": 884,
10
- "eval_samples_per_second": 31.472,
11
- "eval_steps_per_second": 3.952,
12
- "perplexity": 12.42945875004982,
13
- "total_flos": 4151799742464000.0,
14
- "train_loss": 3.14277273785115,
15
- "train_runtime": 586.188,
16
- "train_samples": 3531,
17
- "train_samples_per_second": 18.071,
18
- "train_steps_per_second": 2.262
19
  }
 
1
  {
2
  "epoch": 3.0,
3
+ "eval_bleu": 0.22465329922096197,
4
+ "eval_loss": 0.758726954460144,
5
+ "eval_rouge1": 0.5688224315923487,
6
+ "eval_rouge2": 0.3116501353023259,
7
+ "eval_rougeL": 0.56490169958113,
8
+ "eval_runtime": 45.9889,
9
+ "eval_samples": 14212,
10
+ "eval_samples_per_second": 309.031,
11
+ "eval_steps_per_second": 38.64,
12
+ "perplexity": 2.135555829449579,
13
+ "total_flos": 1.1141034983424e+16,
14
+ "train_loss": 0.8646288613632099,
15
+ "train_runtime": 2092.8708,
16
+ "train_samples": 56851,
17
+ "train_samples_per_second": 81.492,
18
+ "train_steps_per_second": 10.187
19
  }
egy_training_log.txt CHANGED
@@ -1394,3 +1394,5 @@ INFO:root:Epoch 2.0: Train Loss = 1.148, Eval Loss = 0.8351342082023621
1394
  INFO:absl:Using default tokenizer.
1395
  INFO:root:Epoch 3.0: Train Loss = 0.7647, Eval Loss = 0.7705017924308777
1396
  INFO:absl:Using default tokenizer.
 
 
 
1394
  INFO:absl:Using default tokenizer.
1395
  INFO:root:Epoch 3.0: Train Loss = 0.7647, Eval Loss = 0.7705017924308777
1396
  INFO:absl:Using default tokenizer.
1397
+ INFO:__main__:*** Evaluate ***
1398
+ INFO:absl:Using default tokenizer.
eval_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
  "epoch": 3.0,
3
- "eval_bleu": 0.2510920700175718,
4
- "eval_loss": 2.5200693607330322,
5
- "eval_rouge1": 0.45339120041307046,
6
- "eval_rouge2": 0.21369562389429725,
7
- "eval_rougeL": 0.3983722007696566,
8
- "eval_runtime": 28.0888,
9
- "eval_samples": 884,
10
- "eval_samples_per_second": 31.472,
11
- "eval_steps_per_second": 3.952,
12
- "perplexity": 12.42945875004982
13
  }
 
1
  {
2
  "epoch": 3.0,
3
+ "eval_bleu": 0.22465329922096197,
4
+ "eval_loss": 0.758726954460144,
5
+ "eval_rouge1": 0.5688224315923487,
6
+ "eval_rouge2": 0.3116501353023259,
7
+ "eval_rougeL": 0.56490169958113,
8
+ "eval_runtime": 45.9889,
9
+ "eval_samples": 14212,
10
+ "eval_samples_per_second": 309.031,
11
+ "eval_steps_per_second": 38.64,
12
+ "perplexity": 2.135555829449579
13
  }
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 3.0,
3
- "total_flos": 4151799742464000.0,
4
- "train_loss": 3.14277273785115,
5
- "train_runtime": 586.188,
6
- "train_samples": 3531,
7
- "train_samples_per_second": 18.071,
8
- "train_steps_per_second": 2.262
9
  }
 
1
  {
2
  "epoch": 3.0,
3
+ "total_flos": 1.1141034983424e+16,
4
+ "train_loss": 0.8646288613632099,
5
+ "train_runtime": 2092.8708,
6
+ "train_samples": 56851,
7
+ "train_samples_per_second": 81.492,
8
+ "train_steps_per_second": 10.187
9
  }
train_vs_val_loss.png CHANGED
trainer_state.json CHANGED
@@ -1,82 +1,82 @@
1
  {
2
- "best_metric": 2.5200693607330322,
3
- "best_model_checkpoint": "/home/iais_marenpielka/Bouthaina/res_nw_egy/checkpoint-1326",
4
  "epoch": 3.0,
5
  "eval_steps": 500,
6
- "global_step": 1326,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 1.0,
13
- "grad_norm": 1.6992123126983643,
14
- "learning_rate": 4.4200000000000004e-05,
15
- "loss": 4.0723,
16
- "step": 442
17
  },
18
  {
19
  "epoch": 1.0,
20
- "eval_bleu": 0.22275841970876573,
21
- "eval_loss": 2.958087921142578,
22
- "eval_rouge1": 0.3912678429308313,
23
- "eval_rouge2": 0.16481568948149655,
24
- "eval_rougeL": 0.32008248461173605,
25
- "eval_runtime": 27.8657,
26
- "eval_samples_per_second": 31.724,
27
- "eval_steps_per_second": 3.983,
28
- "step": 442
29
  },
30
  {
31
  "epoch": 2.0,
32
- "grad_norm": 1.676450252532959,
33
- "learning_rate": 2.6755447941888623e-05,
34
- "loss": 2.8055,
35
- "step": 884
36
  },
37
  {
38
  "epoch": 2.0,
39
- "eval_bleu": 0.24154624358915514,
40
- "eval_loss": 2.606330394744873,
41
- "eval_rouge1": 0.43470628711649895,
42
- "eval_rouge2": 0.19710102219495218,
43
- "eval_rougeL": 0.3776869994743499,
44
- "eval_runtime": 28.1779,
45
- "eval_samples_per_second": 31.372,
46
- "eval_steps_per_second": 3.939,
47
- "step": 884
48
  },
49
  {
50
  "epoch": 3.0,
51
- "grad_norm": 1.5551155805587769,
52
  "learning_rate": 0.0,
53
- "loss": 2.5505,
54
- "step": 1326
55
  },
56
  {
57
  "epoch": 3.0,
58
- "eval_bleu": 0.2510920700175718,
59
- "eval_loss": 2.5200693607330322,
60
- "eval_rouge1": 0.45339120041307046,
61
- "eval_rouge2": 0.21369562389429725,
62
- "eval_rougeL": 0.3983722007696566,
63
- "eval_runtime": 28.2169,
64
- "eval_samples_per_second": 31.329,
65
- "eval_steps_per_second": 3.934,
66
- "step": 1326
67
  },
68
  {
69
  "epoch": 3.0,
70
- "step": 1326,
71
- "total_flos": 4151799742464000.0,
72
- "train_loss": 3.14277273785115,
73
- "train_runtime": 586.188,
74
- "train_samples_per_second": 18.071,
75
- "train_steps_per_second": 2.262
76
  }
77
  ],
78
  "logging_steps": 500,
79
- "max_steps": 1326,
80
  "num_input_tokens_seen": 0,
81
  "num_train_epochs": 3,
82
  "save_steps": 500,
@@ -101,7 +101,7 @@
101
  "attributes": {}
102
  }
103
  },
104
- "total_flos": 4151799742464000.0,
105
  "train_batch_size": 8,
106
  "trial_name": null,
107
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.758726954460144,
3
+ "best_model_checkpoint": "/home/iais_marenpielka/Bouthaina/res_nw_egy/checkpoint-21321",
4
  "epoch": 3.0,
5
  "eval_steps": 500,
6
+ "global_step": 21321,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 1.0,
13
+ "grad_norm": 2.9758260250091553,
14
+ "learning_rate": 3.413380721387061e-05,
15
+ "loss": 1.148,
16
+ "step": 7107
17
  },
18
  {
19
  "epoch": 1.0,
20
+ "eval_bleu": 0.18719204852263469,
21
+ "eval_loss": 0.8351342082023621,
22
+ "eval_rouge1": 0.5142833565402041,
23
+ "eval_rouge2": 0.2505733532313653,
24
+ "eval_rougeL": 0.5096825616972613,
25
+ "eval_runtime": 169.5172,
26
+ "eval_samples_per_second": 83.838,
27
+ "eval_steps_per_second": 10.483,
28
+ "step": 7107
29
  },
30
  {
31
  "epoch": 2.0,
32
+ "grad_norm": 3.232348680496216,
33
+ "learning_rate": 1.7066903606935305e-05,
34
+ "loss": 0.7647,
35
+ "step": 14214
36
  },
37
  {
38
  "epoch": 2.0,
39
+ "eval_bleu": 0.2145574961692684,
40
+ "eval_loss": 0.7705017924308777,
41
+ "eval_rouge1": 0.5551766890409657,
42
+ "eval_rouge2": 0.29664921512752485,
43
+ "eval_rougeL": 0.5514002756721588,
44
+ "eval_runtime": 108.2094,
45
+ "eval_samples_per_second": 131.338,
46
+ "eval_steps_per_second": 16.422,
47
+ "step": 14214
48
  },
49
  {
50
  "epoch": 3.0,
51
+ "grad_norm": 2.0039141178131104,
52
  "learning_rate": 0.0,
53
+ "loss": 0.6812,
54
+ "step": 21321
55
  },
56
  {
57
  "epoch": 3.0,
58
+ "eval_bleu": 0.22465329922096197,
59
+ "eval_loss": 0.758726954460144,
60
+ "eval_rouge1": 0.5688224315923487,
61
+ "eval_rouge2": 0.3116501353023259,
62
+ "eval_rougeL": 0.56490169958113,
63
+ "eval_runtime": 292.0707,
64
+ "eval_samples_per_second": 48.659,
65
+ "eval_steps_per_second": 6.084,
66
+ "step": 21321
67
  },
68
  {
69
  "epoch": 3.0,
70
+ "step": 21321,
71
+ "total_flos": 1.1141034983424e+16,
72
+ "train_loss": 0.8646288613632099,
73
+ "train_runtime": 2092.8708,
74
+ "train_samples_per_second": 81.492,
75
+ "train_steps_per_second": 10.187
76
  }
77
  ],
78
  "logging_steps": 500,
79
+ "max_steps": 21321,
80
  "num_input_tokens_seen": 0,
81
  "num_train_epochs": 3,
82
  "save_steps": 500,
 
101
  "attributes": {}
102
  }
103
  },
104
+ "total_flos": 1.1141034983424e+16,
105
  "train_batch_size": 8,
106
  "trial_name": null,
107
  "trial_params": null