ben81828 commited on
Commit
7d16876
·
verified ·
1 Parent(s): 92ef1ed

Training in progress, step 11650, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:76c6beb62b007fa74c1f9fde13741ad45368bd6d4b83ec8b972cd4fb0a8c646d
3
  size 29034840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:571e80588f8cd6752186a69103bb988fd53e5a9da6532e4c08ae6f01c4edd4fc
3
  size 29034840
last-checkpoint/global_step11650/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2180ba83222f11e29fffc8a3c275861ad0e249b031a848ae27e5a69fe782b43
3
+ size 43429616
last-checkpoint/global_step11650/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4098ce9867c4422f3717286eb58db800288e0f0c1b7bca46c8b6972354ca5ed
3
+ size 43429616
last-checkpoint/global_step11650/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c705673d8857ac5aed5d2146bd25f3d98d899208feae5fd54f4169977ddd03f1
3
+ size 43429616
last-checkpoint/global_step11650/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f57cf0bd2897e74ccd1869b30a0f419fdfb219cd3e6c17f0f9927f2be9382f4
3
+ size 43429616
last-checkpoint/global_step11650/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:848f867d3b3ddd4219efdfab81a3a352d213e762b9d8691effb0e7edbfd2a599
3
+ size 637299
last-checkpoint/global_step11650/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb8c29838a5b3e8a27da1fc01f55efe7330ccc7d763647344b21f9c83858c6d2
3
+ size 637171
last-checkpoint/global_step11650/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61fc58258866099faad38d795eabc41d6472e47ebdf574e098a1b8c99575718b
3
+ size 637171
last-checkpoint/global_step11650/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:820b15695d0698819be1dfbf872b626365c7b9cd65ebdac09f644af6632bd548
3
+ size 637171
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step11600
 
1
+ global_step11650
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:04a608f7c753833a4201b8a81eec67084f1800198ebebdb3ca5a7c449c2250cb
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:140dcfa0ccc3e7e4e75a20ea5574fb88aaa6f66403d175a54cdb44ba5a484202
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4c7f426827da46cf917c77ad26eb67616a08f771ca15120e70865759017becdd
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51df7ea8967ee21aaf5ecebba23dea00d0f4e9d07c0ee4b09665074da06c4271
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:70ed1e4b671ca40cd0fee33c20c9ff21bf8698f4ec8ae49767b7184e98cc821b
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:703ce5f91b22ed45270a54df2495ee6ed74c9f8706cd08d9901cf4e4b2776ec9
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:526863b8e330a9651e5ae6e145b01d0e4a71c0dd8df04575f1bcd0c28be3580e
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e5302c66fc2f45ec32af6d0bdf0705fc6e7006f56e1c4176aa2683dd8a44a17
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e31fcb959dc156eb9fb669240684e5da71a7affef70d5a591a69f5f5f3ea816b
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73873c5dc0739a1c9b404225926fb3c60340455b686793a1155923d4d88f997b
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.04092838987708092,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_detect_classify_augmented/lora/sft/checkpoint-9050",
4
- "epoch": 1.421181587087685,
5
  "eval_steps": 50,
6
- "global_step": 11600,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -20655,11 +20655,100 @@
20655
  "eval_steps_per_second": 0.769,
20656
  "num_input_tokens_seen": 77549560,
20657
  "step": 11600
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20658
  }
20659
  ],
20660
  "logging_steps": 5,
20661
  "max_steps": 16324,
20662
- "num_input_tokens_seen": 77549560,
20663
  "num_train_epochs": 2,
20664
  "save_steps": 50,
20665
  "stateful_callbacks": {
@@ -20674,7 +20763,7 @@
20674
  "attributes": {}
20675
  }
20676
  },
20677
- "total_flos": 4864183102078976.0,
20678
  "train_batch_size": 1,
20679
  "trial_name": null,
20680
  "trial_params": null
 
1
  {
2
  "best_metric": 0.04092838987708092,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_detect_classify_augmented/lora/sft/checkpoint-9050",
4
+ "epoch": 1.4273069737527182,
5
  "eval_steps": 50,
6
+ "global_step": 11650,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
20655
  "eval_steps_per_second": 0.769,
20656
  "num_input_tokens_seen": 77549560,
20657
  "step": 11600
20658
+ },
20659
+ {
20660
+ "epoch": 1.4217941257541882,
20661
+ "grad_norm": 1.482456260147129,
20662
+ "learning_rate": 2.116166636439865e-05,
20663
+ "loss": 0.2383,
20664
+ "num_input_tokens_seen": 77582480,
20665
+ "step": 11605
20666
+ },
20667
+ {
20668
+ "epoch": 1.4224066644206916,
20669
+ "grad_norm": 0.7904057956562802,
20670
+ "learning_rate": 2.1120306382769223e-05,
20671
+ "loss": 0.2478,
20672
+ "num_input_tokens_seen": 77616096,
20673
+ "step": 11610
20674
+ },
20675
+ {
20676
+ "epoch": 1.4230192030871949,
20677
+ "grad_norm": 1.2212879423399834,
20678
+ "learning_rate": 2.1078976034215076e-05,
20679
+ "loss": 0.2231,
20680
+ "num_input_tokens_seen": 77649232,
20681
+ "step": 11615
20682
+ },
20683
+ {
20684
+ "epoch": 1.4236317417536983,
20685
+ "grad_norm": 1.17414440380428,
20686
+ "learning_rate": 2.1037675361144764e-05,
20687
+ "loss": 0.2292,
20688
+ "num_input_tokens_seen": 77683000,
20689
+ "step": 11620
20690
+ },
20691
+ {
20692
+ "epoch": 1.4242442804202016,
20693
+ "grad_norm": 1.633390357694746,
20694
+ "learning_rate": 2.0996404405936388e-05,
20695
+ "loss": 0.2083,
20696
+ "num_input_tokens_seen": 77716136,
20697
+ "step": 11625
20698
+ },
20699
+ {
20700
+ "epoch": 1.4248568190867048,
20701
+ "grad_norm": 0.9236835374178289,
20702
+ "learning_rate": 2.0955163210937507e-05,
20703
+ "loss": 0.204,
20704
+ "num_input_tokens_seen": 77750064,
20705
+ "step": 11630
20706
+ },
20707
+ {
20708
+ "epoch": 1.425469357753208,
20709
+ "grad_norm": 1.358055893320477,
20710
+ "learning_rate": 2.091395181846516e-05,
20711
+ "loss": 0.2225,
20712
+ "num_input_tokens_seen": 77783792,
20713
+ "step": 11635
20714
+ },
20715
+ {
20716
+ "epoch": 1.4260818964197115,
20717
+ "grad_norm": 1.2925316714831874,
20718
+ "learning_rate": 2.0872770270805834e-05,
20719
+ "loss": 0.207,
20720
+ "num_input_tokens_seen": 77817472,
20721
+ "step": 11640
20722
+ },
20723
+ {
20724
+ "epoch": 1.4266944350862147,
20725
+ "grad_norm": 1.0474518928956342,
20726
+ "learning_rate": 2.0831618610215374e-05,
20727
+ "loss": 0.2444,
20728
+ "num_input_tokens_seen": 77850496,
20729
+ "step": 11645
20730
+ },
20731
+ {
20732
+ "epoch": 1.4273069737527182,
20733
+ "grad_norm": 1.4801887733220611,
20734
+ "learning_rate": 2.079049687891898e-05,
20735
+ "loss": 0.23,
20736
+ "num_input_tokens_seen": 77884056,
20737
+ "step": 11650
20738
+ },
20739
+ {
20740
+ "epoch": 1.4273069737527182,
20741
+ "eval_loss": 0.17718727886676788,
20742
+ "eval_runtime": 19.8826,
20743
+ "eval_samples_per_second": 3.018,
20744
+ "eval_steps_per_second": 0.754,
20745
+ "num_input_tokens_seen": 77884056,
20746
+ "step": 11650
20747
  }
20748
  ],
20749
  "logging_steps": 5,
20750
  "max_steps": 16324,
20751
+ "num_input_tokens_seen": 77884056,
20752
  "num_train_epochs": 2,
20753
  "save_steps": 50,
20754
  "stateful_callbacks": {
 
20763
  "attributes": {}
20764
  }
20765
  },
20766
+ "total_flos": 4885167416672256.0,
20767
  "train_batch_size": 1,
20768
  "trial_name": null,
20769
  "trial_params": null