PEFT
Safetensors
Generated from Trainer
cpapa79 commited on
Commit
3b06f2e
·
verified ·
1 Parent(s): 396499d

Training in progress, step 18444, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:86f6c98c0f897bcc3ac3b4414b05848ac7776ea94e028266f9686dfe3ccc2140
3
  size 18899856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fdda39341f708250fb5998f447d13d70f26a8944dbc0fc19b0f6972737238302
3
  size 18899856
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a84f196d23f22e0c4eba236ffee8cec2d96782b391895a0215af71b3e1bfbbe
3
  size 37911546
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d214aa30c3dc0ff733d67ad7ab92c45d9fc115f184e0c43db66917659dad3221
3
  size 37911546
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:13d0bbbc8564cdc936edb07b992feb927ee48638b36c16ae513b6b501eb6a4d7
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1810cf1d4372b4023939ddac1cf1b6ce0e3f798979b5a4c9c646ef01643ac09c
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:82ab3f4c71b15efcf419ae699bcc09c12d9d69fda628dcac52fdb8059760dd34
3
  size 1000
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:112b0e3f3850ccaa762f9eb894c6c56a98cff86a4eab4f8d883ff8a96dbc8748
3
  size 1000
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 3.9973975276512688,
5
  "eval_steps": 2048,
6
- "global_step": 18432,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -129211,6 +129211,90 @@
129211
  "eval_test_samples_per_second": 14.815,
129212
  "eval_test_steps_per_second": 0.926,
129213
  "step": 18432
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
129214
  }
129215
  ],
129216
  "logging_steps": 1,
@@ -129225,12 +129309,12 @@
129225
  "should_evaluate": false,
129226
  "should_log": false,
129227
  "should_save": true,
129228
- "should_training_stop": false
129229
  },
129230
  "attributes": {}
129231
  }
129232
  },
129233
- "total_flos": 4.448916451273212e+18,
129234
  "train_batch_size": 16,
129235
  "trial_name": null,
129236
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 4.0,
5
  "eval_steps": 2048,
6
+ "global_step": 18444,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
129211
  "eval_test_samples_per_second": 14.815,
129212
  "eval_test_steps_per_second": 0.926,
129213
  "step": 18432
129214
+ },
129215
+ {
129216
+ "epoch": 3.9976144003469964,
129217
+ "grad_norm": 0.7556371092796326,
129218
+ "learning_rate": 3.0003218233616098e-06,
129219
+ "loss": 2.1035,
129220
+ "step": 18433
129221
+ },
129222
+ {
129223
+ "epoch": 3.997831273042724,
129224
+ "grad_norm": 0.8098835349082947,
129225
+ "learning_rate": 3.000265969737023e-06,
129226
+ "loss": 2.1821,
129227
+ "step": 18434
129228
+ },
129229
+ {
129230
+ "epoch": 3.9980481457384514,
129231
+ "grad_norm": 0.8024788498878479,
129232
+ "learning_rate": 3.000215435499207e-06,
129233
+ "loss": 2.1821,
129234
+ "step": 18435
129235
+ },
129236
+ {
129237
+ "epoch": 3.998265018434179,
129238
+ "grad_norm": 0.6993958353996277,
129239
+ "learning_rate": 3.000170220649976e-06,
129240
+ "loss": 1.8316,
129241
+ "step": 18436
129242
+ },
129243
+ {
129244
+ "epoch": 3.9984818911299067,
129245
+ "grad_norm": 0.8931436538696289,
129246
+ "learning_rate": 3.000130325190978e-06,
129247
+ "loss": 1.9669,
129248
+ "step": 18437
129249
+ },
129250
+ {
129251
+ "epoch": 3.9986987638256344,
129252
+ "grad_norm": 0.8958826065063477,
129253
+ "learning_rate": 3.0000957491236143e-06,
129254
+ "loss": 2.0962,
129255
+ "step": 18438
129256
+ },
129257
+ {
129258
+ "epoch": 3.998915636521362,
129259
+ "grad_norm": 0.7808493375778198,
129260
+ "learning_rate": 3.000066492449139e-06,
129261
+ "loss": 2.0824,
129262
+ "step": 18439
129263
+ },
129264
+ {
129265
+ "epoch": 3.9991325092170893,
129266
+ "grad_norm": 0.6934779286384583,
129267
+ "learning_rate": 3.00004255516859e-06,
129268
+ "loss": 2.0578,
129269
+ "step": 18440
129270
+ },
129271
+ {
129272
+ "epoch": 3.9993493819128174,
129273
+ "grad_norm": 0.8631987571716309,
129274
+ "learning_rate": 3.0000239372828404e-06,
129275
+ "loss": 2.2236,
129276
+ "step": 18441
129277
+ },
129278
+ {
129279
+ "epoch": 3.9995662546085446,
129280
+ "grad_norm": 0.6597968935966492,
129281
+ "learning_rate": 3.0000106387925348e-06,
129282
+ "loss": 2.0531,
129283
+ "step": 18442
129284
+ },
129285
+ {
129286
+ "epoch": 3.9997831273042723,
129287
+ "grad_norm": 0.7519582509994507,
129288
+ "learning_rate": 3.00000265969815e-06,
129289
+ "loss": 2.1216,
129290
+ "step": 18443
129291
+ },
129292
+ {
129293
+ "epoch": 4.0,
129294
+ "grad_norm": 1.018497347831726,
129295
+ "learning_rate": 2.9999999999999997e-06,
129296
+ "loss": 2.0909,
129297
+ "step": 18444
129298
  }
129299
  ],
129300
  "logging_steps": 1,
 
129309
  "should_evaluate": false,
129310
  "should_log": false,
129311
  "should_save": true,
129312
+ "should_training_stop": true
129313
  },
129314
  "attributes": {}
129315
  }
129316
  },
129317
+ "total_flos": 4.451693457275093e+18,
129318
  "train_batch_size": 16,
129319
  "trial_name": null,
129320
  "trial_params": null