yuriivoievidka commited on
Commit
233344d
·
verified ·
1 Parent(s): a3b7a64

Upload folder using huggingface_hub

Browse files
1_Pooling/config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "word_embedding_dimension": 768,
3
- "pooling_mode_cls_token": false,
4
- "pooling_mode_mean_tokens": true,
5
  "pooling_mode_max_tokens": false,
6
  "pooling_mode_mean_sqrt_len_tokens": false,
7
  "pooling_mode_weightedmean_tokens": false,
 
1
  {
2
  "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": true,
4
+ "pooling_mode_mean_tokens": false,
5
  "pooling_mode_max_tokens": false,
6
  "pooling_mode_mean_sqrt_len_tokens": false,
7
  "pooling_mode_weightedmean_tokens": false,
README.md CHANGED
@@ -6,7 +6,7 @@ tags:
6
  - generated_from_trainer
7
  - dataset_size:10635
8
  - loss:MultipleNegativesSymmetricRankingLoss
9
- base_model: sentence-transformers/all-mpnet-base-v2
10
  widget:
11
  - source_sentence: '12 Rules For Life: An Antidote to Chaos by Jordan B. Peterson'
12
  sentences:
@@ -40,18 +40,18 @@ pipeline_tag: sentence-similarity
40
  library_name: sentence-transformers
41
  ---
42
 
43
- # SentenceTransformer based on sentence-transformers/all-mpnet-base-v2
44
 
45
- This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-mpnet-base-v2](https://huggingface.co/sentence-transformers/all-mpnet-base-v2) on the train dataset. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
46
 
47
  ## Model Details
48
 
49
  ### Model Description
50
  - **Model Type:** Sentence Transformer
51
- - **Base model:** [sentence-transformers/all-mpnet-base-v2](https://huggingface.co/sentence-transformers/all-mpnet-base-v2) <!-- at revision 12e86a3c702fc3c50205a8db88f0ec7c0b6b94a0 -->
52
- - **Maximum Sequence Length:** 384 tokens
53
  - **Output Dimensionality:** 768 dimensions
54
- - **Similarity Function:** Cosine Similarity
55
  - **Training Dataset:**
56
  - train
57
  <!-- - **Language:** Unknown -->
@@ -67,9 +67,8 @@ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [s
67
 
68
  ```
69
  SentenceTransformer(
70
- (0): Transformer({'max_seq_length': 384, 'do_lower_case': False}) with Transformer model: MPNetModel
71
- (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
72
- (2): Normalize()
73
  )
74
  ```
75
 
@@ -202,7 +201,7 @@ You can finetune this model on your own dataset.
202
  - `per_device_train_batch_size`: 16
203
  - `per_device_eval_batch_size`: 16
204
  - `learning_rate`: 2e-05
205
- - `num_train_epochs`: 7
206
  - `warmup_ratio`: 0.1
207
 
208
  #### All Hyperparameters
@@ -225,7 +224,7 @@ You can finetune this model on your own dataset.
225
  - `adam_beta2`: 0.999
226
  - `adam_epsilon`: 1e-08
227
  - `max_grad_norm`: 1.0
228
- - `num_train_epochs`: 7
229
  - `max_steps`: -1
230
  - `lr_scheduler_type`: linear
231
  - `lr_scheduler_kwargs`: {}
@@ -327,17 +326,22 @@ You can finetune this model on your own dataset.
327
  ### Training Logs
328
  | Epoch | Step | Training Loss | train loss |
329
  |:------:|:----:|:-------------:|:----------:|
330
- | 0.6006 | 200 | 2.5755 | 2.4113 |
331
- | 1.2012 | 400 | 2.2395 | 2.3553 |
332
- | 1.8018 | 600 | 2.0813 | 2.3290 |
333
- | 2.4024 | 800 | 1.9813 | 2.3169 |
334
- | 3.0030 | 1000 | 1.9233 | 2.3081 |
335
- | 3.6036 | 1200 | 1.8338 | 2.3076 |
336
- | 4.2042 | 1400 | 1.8029 | 2.3380 |
337
- | 4.8048 | 1600 | 1.7766 | 2.3005 |
338
- | 5.4054 | 1800 | 1.722 | 2.3254 |
339
- | 6.0060 | 2000 | 1.7217 | 2.3215 |
340
- | 6.6066 | 2200 | 1.6759 | 2.3322 |
 
 
 
 
 
341
 
342
 
343
  ### Framework Versions
 
6
  - generated_from_trainer
7
  - dataset_size:10635
8
  - loss:MultipleNegativesSymmetricRankingLoss
9
+ base_model: sentence-transformers/multi-qa-mpnet-base-dot-v1
10
  widget:
11
  - source_sentence: '12 Rules For Life: An Antidote to Chaos by Jordan B. Peterson'
12
  sentences:
 
40
  library_name: sentence-transformers
41
  ---
42
 
43
+ # SentenceTransformer based on sentence-transformers/multi-qa-mpnet-base-dot-v1
44
 
45
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/multi-qa-mpnet-base-dot-v1](https://huggingface.co/sentence-transformers/multi-qa-mpnet-base-dot-v1) on the train dataset. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
46
 
47
  ## Model Details
48
 
49
  ### Model Description
50
  - **Model Type:** Sentence Transformer
51
+ - **Base model:** [sentence-transformers/multi-qa-mpnet-base-dot-v1](https://huggingface.co/sentence-transformers/multi-qa-mpnet-base-dot-v1) <!-- at revision 4633e80e17ea975bc090c97b049da26062b054d3 -->
52
+ - **Maximum Sequence Length:** 512 tokens
53
  - **Output Dimensionality:** 768 dimensions
54
+ - **Similarity Function:** Dot Product
55
  - **Training Dataset:**
56
  - train
57
  <!-- - **Language:** Unknown -->
 
67
 
68
  ```
69
  SentenceTransformer(
70
+ (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: MPNetModel
71
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
 
72
  )
73
  ```
74
 
 
201
  - `per_device_train_batch_size`: 16
202
  - `per_device_eval_batch_size`: 16
203
  - `learning_rate`: 2e-05
204
+ - `num_train_epochs`: 10
205
  - `warmup_ratio`: 0.1
206
 
207
  #### All Hyperparameters
 
224
  - `adam_beta2`: 0.999
225
  - `adam_epsilon`: 1e-08
226
  - `max_grad_norm`: 1.0
227
+ - `num_train_epochs`: 10
228
  - `max_steps`: -1
229
  - `lr_scheduler_type`: linear
230
  - `lr_scheduler_kwargs`: {}
 
326
  ### Training Logs
327
  | Epoch | Step | Training Loss | train loss |
328
  |:------:|:----:|:-------------:|:----------:|
329
+ | 0.6006 | 200 | 2.6385 | 2.4890 |
330
+ | 1.2012 | 400 | 2.3324 | 2.4199 |
331
+ | 1.8018 | 600 | 2.1772 | 2.3891 |
332
+ | 2.4024 | 800 | 2.0635 | 2.3691 |
333
+ | 3.0030 | 1000 | 1.9915 | 2.3609 |
334
+ | 3.6036 | 1200 | 1.9008 | 2.3689 |
335
+ | 4.2042 | 1400 | 1.8603 | 2.3850 |
336
+ | 4.8048 | 1600 | 1.8421 | 2.3468 |
337
+ | 5.4054 | 1800 | 1.785 | 2.3649 |
338
+ | 6.0060 | 2000 | 1.786 | 2.3783 |
339
+ | 6.6066 | 2200 | 1.7331 | 2.3782 |
340
+ | 7.2072 | 2400 | 1.7062 | 2.3826 |
341
+ | 7.8078 | 2600 | 1.6929 | 2.3926 |
342
+ | 8.4084 | 2800 | 1.6618 | 2.4069 |
343
+ | 9.0090 | 3000 | 1.6348 | 2.4155 |
344
+ | 9.6096 | 3200 | 1.6553 | 2.4060 |
345
 
346
 
347
  ### Framework Versions
config_sentence_transformers.json CHANGED
@@ -6,5 +6,5 @@
6
  },
7
  "prompts": {},
8
  "default_prompt_name": null,
9
- "similarity_fn_name": "cosine"
10
  }
 
6
  },
7
  "prompts": {},
8
  "default_prompt_name": null,
9
+ "similarity_fn_name": "dot"
10
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2ddbdb6468cca74299465849d3460bd26cc567441655d3084af35bb3acd7144e
3
  size 437967672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75eb15217feda2677c50021eb1b205602aafb32075a93777c06d32e5ed495f54
3
  size 437967672
modules.json CHANGED
@@ -10,11 +10,5 @@
10
  "name": "1",
11
  "path": "1_Pooling",
12
  "type": "sentence_transformers.models.Pooling"
13
- },
14
- {
15
- "idx": 2,
16
- "name": "2",
17
- "path": "2_Normalize",
18
- "type": "sentence_transformers.models.Normalize"
19
  }
20
  ]
 
10
  "name": "1",
11
  "path": "1_Pooling",
12
  "type": "sentence_transformers.models.Pooling"
 
 
 
 
 
 
13
  }
14
  ]
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:43f6626a50f5d4a1fe140af797071be81b4523615afa4e4b6d5795ee9ef59320
3
  size 876058170
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b3cab1a8a0c140c73e96cbcb529e091bfa6bba42ac1dc93a4ecf69ccf501fbf
3
  size 876058170
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9887d9179089333ff9b4030c7aa932e0435c5243b5cc42026e85559ac64ae3e
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e0d77b405f4c1d7f21a2c4933ee4193fa622ed4c467c356582f6f67c7022846
3
  size 14244
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:190fc72819eea0b8f2844c8816cd0625c6bce70b27c2d3b3ce154d7ea3cae54a
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:243e8de64f4f19d16558eb6f5b4a77ed28295d1f7205fc19fb8d2b20be1c6b46
3
  size 1064
sentence_bert_config.json CHANGED
@@ -1,4 +1,4 @@
1
  {
2
- "max_seq_length": 384,
3
  "do_lower_case": false
4
  }
 
1
  {
2
+ "max_seq_length": 512,
3
  "do_lower_case": false
4
  }
tokenizer.json CHANGED
@@ -2,7 +2,7 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 384,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 512,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
tokenizer_config.json CHANGED
@@ -56,8 +56,8 @@
56
  "eos_token": "</s>",
57
  "extra_special_tokens": {},
58
  "mask_token": "<mask>",
59
- "max_length": 128,
60
- "model_max_length": 384,
61
  "pad_to_multiple_of": null,
62
  "pad_token": "<pad>",
63
  "pad_token_type_id": 0,
 
56
  "eos_token": "</s>",
57
  "extra_special_tokens": {},
58
  "mask_token": "<mask>",
59
+ "max_length": 250,
60
+ "model_max_length": 512,
61
  "pad_to_multiple_of": null,
62
  "pad_token": "<pad>",
63
  "pad_token_type_id": 0,
trainer_state.json CHANGED
@@ -2,184 +2,259 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 7.0,
6
  "eval_steps": 200,
7
- "global_step": 2331,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
  "epoch": 0.6006006006006006,
14
- "grad_norm": 9.713151931762695,
15
- "learning_rate": 1.700854700854701e-05,
16
- "loss": 2.5755,
17
  "step": 200
18
  },
19
  {
20
  "epoch": 0.6006006006006006,
21
- "eval_train_loss": 2.4112653732299805,
22
- "eval_train_runtime": 16.1652,
23
- "eval_train_samples_per_second": 331.515,
24
- "eval_train_steps_per_second": 10.393,
25
  "step": 200
26
  },
27
  {
28
  "epoch": 1.2012012012012012,
29
- "grad_norm": 8.402304649353027,
30
- "learning_rate": 1.8426323319027183e-05,
31
- "loss": 2.2395,
32
  "step": 400
33
  },
34
  {
35
  "epoch": 1.2012012012012012,
36
- "eval_train_loss": 2.3553037643432617,
37
- "eval_train_runtime": 16.1973,
38
- "eval_train_samples_per_second": 330.857,
39
- "eval_train_steps_per_second": 10.372,
40
  "step": 400
41
  },
42
  {
43
  "epoch": 1.8018018018018018,
44
- "grad_norm": 8.638038635253906,
45
- "learning_rate": 1.6518836432999526e-05,
46
- "loss": 2.0813,
47
  "step": 600
48
  },
49
  {
50
  "epoch": 1.8018018018018018,
51
- "eval_train_loss": 2.3290350437164307,
52
- "eval_train_runtime": 15.8942,
53
- "eval_train_samples_per_second": 337.167,
54
- "eval_train_steps_per_second": 10.57,
55
  "step": 600
56
  },
57
  {
58
  "epoch": 2.4024024024024024,
59
- "grad_norm": 8.17331314086914,
60
- "learning_rate": 1.4611349546971865e-05,
61
- "loss": 1.9813,
62
  "step": 800
63
  },
64
  {
65
  "epoch": 2.4024024024024024,
66
- "eval_train_loss": 2.316850423812866,
67
- "eval_train_runtime": 16.4597,
68
- "eval_train_samples_per_second": 325.584,
69
- "eval_train_steps_per_second": 10.207,
70
  "step": 800
71
  },
72
  {
73
  "epoch": 3.003003003003003,
74
- "grad_norm": 8.972688674926758,
75
- "learning_rate": 1.2703862660944206e-05,
76
- "loss": 1.9233,
77
  "step": 1000
78
  },
79
  {
80
  "epoch": 3.003003003003003,
81
- "eval_train_loss": 2.3080697059631348,
82
- "eval_train_runtime": 16.059,
83
- "eval_train_samples_per_second": 333.707,
84
- "eval_train_steps_per_second": 10.461,
85
  "step": 1000
86
  },
87
  {
88
  "epoch": 3.6036036036036037,
89
- "grad_norm": 8.94318962097168,
90
- "learning_rate": 1.0796375774916547e-05,
91
- "loss": 1.8338,
92
  "step": 1200
93
  },
94
  {
95
  "epoch": 3.6036036036036037,
96
- "eval_train_loss": 2.3076283931732178,
97
- "eval_train_runtime": 16.1704,
98
- "eval_train_samples_per_second": 331.408,
99
- "eval_train_steps_per_second": 10.389,
100
  "step": 1200
101
  },
102
  {
103
  "epoch": 4.2042042042042045,
104
- "grad_norm": 10.612234115600586,
105
- "learning_rate": 8.888888888888888e-06,
106
- "loss": 1.8029,
107
  "step": 1400
108
  },
109
  {
110
  "epoch": 4.2042042042042045,
111
- "eval_train_loss": 2.337951183319092,
112
- "eval_train_runtime": 16.2105,
113
- "eval_train_samples_per_second": 330.588,
114
- "eval_train_steps_per_second": 10.364,
115
  "step": 1400
116
  },
117
  {
118
  "epoch": 4.804804804804805,
119
- "grad_norm": 8.080140113830566,
120
- "learning_rate": 6.981402002861231e-06,
121
- "loss": 1.7766,
122
  "step": 1600
123
  },
124
  {
125
  "epoch": 4.804804804804805,
126
- "eval_train_loss": 2.300466775894165,
127
- "eval_train_runtime": 16.2606,
128
- "eval_train_samples_per_second": 329.569,
129
- "eval_train_steps_per_second": 10.332,
130
  "step": 1600
131
  },
132
  {
133
  "epoch": 5.405405405405405,
134
- "grad_norm": 8.161681175231934,
135
- "learning_rate": 5.073915116833572e-06,
136
- "loss": 1.722,
137
  "step": 1800
138
  },
139
  {
140
  "epoch": 5.405405405405405,
141
- "eval_train_loss": 2.325410842895508,
142
- "eval_train_runtime": 16.2864,
143
- "eval_train_samples_per_second": 329.047,
144
- "eval_train_steps_per_second": 10.315,
145
  "step": 1800
146
  },
147
  {
148
  "epoch": 6.006006006006006,
149
- "grad_norm": 9.505444526672363,
150
- "learning_rate": 3.1664282308059137e-06,
151
- "loss": 1.7217,
152
  "step": 2000
153
  },
154
  {
155
  "epoch": 6.006006006006006,
156
- "eval_train_loss": 2.3215274810791016,
157
- "eval_train_runtime": 15.9019,
158
- "eval_train_samples_per_second": 337.003,
159
- "eval_train_steps_per_second": 10.565,
160
  "step": 2000
161
  },
162
  {
163
  "epoch": 6.606606606606607,
164
- "grad_norm": 11.631622314453125,
165
- "learning_rate": 1.2589413447782547e-06,
166
- "loss": 1.6759,
167
  "step": 2200
168
  },
169
  {
170
  "epoch": 6.606606606606607,
171
- "eval_train_loss": 2.3322482109069824,
172
- "eval_train_runtime": 15.9374,
173
- "eval_train_samples_per_second": 336.253,
174
- "eval_train_steps_per_second": 10.541,
175
  "step": 2200
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
176
  }
177
  ],
178
  "logging_steps": 200,
179
- "max_steps": 2331,
180
  "num_input_tokens_seen": 0,
181
- "num_train_epochs": 7,
182
- "save_steps": 3000,
183
  "stateful_callbacks": {
184
  "TrainerControl": {
185
  "args": {
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 10.0,
6
  "eval_steps": 200,
7
+ "global_step": 3330,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
  "epoch": 0.6006006006006006,
14
+ "grad_norm": 10.51630973815918,
15
+ "learning_rate": 1.1951951951951951e-05,
16
+ "loss": 2.6385,
17
  "step": 200
18
  },
19
  {
20
  "epoch": 0.6006006006006006,
21
+ "eval_train_loss": 2.488961696624756,
22
+ "eval_train_runtime": 15.8109,
23
+ "eval_train_samples_per_second": 338.943,
24
+ "eval_train_steps_per_second": 10.626,
25
  "step": 200
26
  },
27
  {
28
  "epoch": 1.2012012012012012,
29
+ "grad_norm": 9.947551727294922,
30
+ "learning_rate": 1.9559559559559563e-05,
31
+ "loss": 2.3324,
32
  "step": 400
33
  },
34
  {
35
  "epoch": 1.2012012012012012,
36
+ "eval_train_loss": 2.419891119003296,
37
+ "eval_train_runtime": 15.4172,
38
+ "eval_train_samples_per_second": 347.6,
39
+ "eval_train_steps_per_second": 10.897,
40
  "step": 400
41
  },
42
  {
43
  "epoch": 1.8018018018018018,
44
+ "grad_norm": 9.27906322479248,
45
+ "learning_rate": 1.8224891558224895e-05,
46
+ "loss": 2.1772,
47
  "step": 600
48
  },
49
  {
50
  "epoch": 1.8018018018018018,
51
+ "eval_train_loss": 2.389059543609619,
52
+ "eval_train_runtime": 15.629,
53
+ "eval_train_samples_per_second": 342.889,
54
+ "eval_train_steps_per_second": 10.749,
55
  "step": 600
56
  },
57
  {
58
  "epoch": 2.4024024024024024,
59
+ "grad_norm": 8.641522407531738,
60
+ "learning_rate": 1.6890223556890223e-05,
61
+ "loss": 2.0635,
62
  "step": 800
63
  },
64
  {
65
  "epoch": 2.4024024024024024,
66
+ "eval_train_loss": 2.369140148162842,
67
+ "eval_train_runtime": 15.6802,
68
+ "eval_train_samples_per_second": 341.768,
69
+ "eval_train_steps_per_second": 10.714,
70
  "step": 800
71
  },
72
  {
73
  "epoch": 3.003003003003003,
74
+ "grad_norm": 9.14696216583252,
75
+ "learning_rate": 1.555555555555556e-05,
76
+ "loss": 1.9915,
77
  "step": 1000
78
  },
79
  {
80
  "epoch": 3.003003003003003,
81
+ "eval_train_loss": 2.360917806625366,
82
+ "eval_train_runtime": 15.1166,
83
+ "eval_train_samples_per_second": 354.511,
84
+ "eval_train_steps_per_second": 11.114,
85
  "step": 1000
86
  },
87
  {
88
  "epoch": 3.6036036036036037,
89
+ "grad_norm": 8.530783653259277,
90
+ "learning_rate": 1.4220887554220888e-05,
91
+ "loss": 1.9008,
92
  "step": 1200
93
  },
94
  {
95
  "epoch": 3.6036036036036037,
96
+ "eval_train_loss": 2.3688642978668213,
97
+ "eval_train_runtime": 15.4199,
98
+ "eval_train_samples_per_second": 347.539,
99
+ "eval_train_steps_per_second": 10.895,
100
  "step": 1200
101
  },
102
  {
103
  "epoch": 4.2042042042042045,
104
+ "grad_norm": 10.91430377960205,
105
+ "learning_rate": 1.288621955288622e-05,
106
+ "loss": 1.8603,
107
  "step": 1400
108
  },
109
  {
110
  "epoch": 4.2042042042042045,
111
+ "eval_train_loss": 2.3849704265594482,
112
+ "eval_train_runtime": 15.4842,
113
+ "eval_train_samples_per_second": 346.094,
114
+ "eval_train_steps_per_second": 10.85,
115
  "step": 1400
116
  },
117
  {
118
  "epoch": 4.804804804804805,
119
+ "grad_norm": 7.67103910446167,
120
+ "learning_rate": 1.1551551551551552e-05,
121
+ "loss": 1.8421,
122
  "step": 1600
123
  },
124
  {
125
  "epoch": 4.804804804804805,
126
+ "eval_train_loss": 2.3467628955841064,
127
+ "eval_train_runtime": 15.716,
128
+ "eval_train_samples_per_second": 340.99,
129
+ "eval_train_steps_per_second": 10.69,
130
  "step": 1600
131
  },
132
  {
133
  "epoch": 5.405405405405405,
134
+ "grad_norm": 10.279878616333008,
135
+ "learning_rate": 1.0216883550216886e-05,
136
+ "loss": 1.785,
137
  "step": 1800
138
  },
139
  {
140
  "epoch": 5.405405405405405,
141
+ "eval_train_loss": 2.3648931980133057,
142
+ "eval_train_runtime": 15.8675,
143
+ "eval_train_samples_per_second": 337.735,
144
+ "eval_train_steps_per_second": 10.588,
145
  "step": 1800
146
  },
147
  {
148
  "epoch": 6.006006006006006,
149
+ "grad_norm": 9.513252258300781,
150
+ "learning_rate": 8.882215548882216e-06,
151
+ "loss": 1.786,
152
  "step": 2000
153
  },
154
  {
155
  "epoch": 6.006006006006006,
156
+ "eval_train_loss": 2.378257989883423,
157
+ "eval_train_runtime": 16.2336,
158
+ "eval_train_samples_per_second": 330.118,
159
+ "eval_train_steps_per_second": 10.349,
160
  "step": 2000
161
  },
162
  {
163
  "epoch": 6.606606606606607,
164
+ "grad_norm": 11.29470157623291,
165
+ "learning_rate": 7.547547547547548e-06,
166
+ "loss": 1.7331,
167
  "step": 2200
168
  },
169
  {
170
  "epoch": 6.606606606606607,
171
+ "eval_train_loss": 2.3782169818878174,
172
+ "eval_train_runtime": 15.9451,
173
+ "eval_train_samples_per_second": 336.091,
174
+ "eval_train_steps_per_second": 10.536,
175
  "step": 2200
176
+ },
177
+ {
178
+ "epoch": 7.207207207207207,
179
+ "grad_norm": 10.022029876708984,
180
+ "learning_rate": 6.21287954621288e-06,
181
+ "loss": 1.7062,
182
+ "step": 2400
183
+ },
184
+ {
185
+ "epoch": 7.207207207207207,
186
+ "eval_train_loss": 2.3825714588165283,
187
+ "eval_train_runtime": 15.9203,
188
+ "eval_train_samples_per_second": 336.615,
189
+ "eval_train_steps_per_second": 10.553,
190
+ "step": 2400
191
+ },
192
+ {
193
+ "epoch": 7.807807807807808,
194
+ "grad_norm": 10.572230339050293,
195
+ "learning_rate": 4.878211544878212e-06,
196
+ "loss": 1.6929,
197
+ "step": 2600
198
+ },
199
+ {
200
+ "epoch": 7.807807807807808,
201
+ "eval_train_loss": 2.3926050662994385,
202
+ "eval_train_runtime": 15.9657,
203
+ "eval_train_samples_per_second": 335.657,
204
+ "eval_train_steps_per_second": 10.523,
205
+ "step": 2600
206
+ },
207
+ {
208
+ "epoch": 8.408408408408409,
209
+ "grad_norm": 10.39648723602295,
210
+ "learning_rate": 3.5435435435435437e-06,
211
+ "loss": 1.6618,
212
+ "step": 2800
213
+ },
214
+ {
215
+ "epoch": 8.408408408408409,
216
+ "eval_train_loss": 2.4069101810455322,
217
+ "eval_train_runtime": 15.7461,
218
+ "eval_train_samples_per_second": 340.337,
219
+ "eval_train_steps_per_second": 10.669,
220
+ "step": 2800
221
+ },
222
+ {
223
+ "epoch": 9.00900900900901,
224
+ "grad_norm": 12.46285343170166,
225
+ "learning_rate": 2.2088755422088755e-06,
226
+ "loss": 1.6348,
227
+ "step": 3000
228
+ },
229
+ {
230
+ "epoch": 9.00900900900901,
231
+ "eval_train_loss": 2.415452480316162,
232
+ "eval_train_runtime": 15.6461,
233
+ "eval_train_samples_per_second": 342.513,
234
+ "eval_train_steps_per_second": 10.737,
235
+ "step": 3000
236
+ },
237
+ {
238
+ "epoch": 9.60960960960961,
239
+ "grad_norm": 11.425995826721191,
240
+ "learning_rate": 8.742075408742076e-07,
241
+ "loss": 1.6553,
242
+ "step": 3200
243
+ },
244
+ {
245
+ "epoch": 9.60960960960961,
246
+ "eval_train_loss": 2.40596079826355,
247
+ "eval_train_runtime": 15.9222,
248
+ "eval_train_samples_per_second": 336.573,
249
+ "eval_train_steps_per_second": 10.551,
250
+ "step": 3200
251
  }
252
  ],
253
  "logging_steps": 200,
254
+ "max_steps": 3330,
255
  "num_input_tokens_seen": 0,
256
+ "num_train_epochs": 10,
257
+ "save_steps": 2000,
258
  "stateful_callbacks": {
259
  "TrainerControl": {
260
  "args": {
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:20f9cc1bfbf387326ed07dacc8a11b82a6ff607c0cb073f258fc1350a90ff02a
3
- size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4190153994cc09d055ab5654484b8b69d39cb01e17665d63b0fc073bb9a0ba91
3
+ size 5560