rak-r05 commited on
Commit
891ec02
·
verified ·
1 Parent(s): 2eea9bc

Training in progress, step 150, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2ff045ac320d83e6646c6f6ab6832b60324d8abccf00a476635b8ad51c195138
3
  size 9450352
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab26f18e6314da78474ca938c1257abcb0d2dc24bf52ff5f89f7936a1b5b755c
3
  size 9450352
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c9ea2cbf525af9c054129774c924e498aa89301d6a601575f3647cf19519500
3
  size 4906490
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a0468144609fbba619a69627d5a8d8d1502da18071b17d1a4de9a30f4be4020
3
  size 4906490
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fa46734679de1685e01f36d430a9f749b2182c205f22d5f38bcbf4bc21f64491
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f557330525e75b8ec8859b7784404c001117d2a2896a23772c13160438b441ea
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d35a71cd96ab153cf58aa68810dc4405b5e1a7fbfef5501d233ae84e34aa51c4
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:285055c203c0e51e433ff14d6fa6e0c364698ec978202e506a844a8372901f4a
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.003874519933385447,
5
  "eval_steps": 38,
6
- "global_step": 114,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -837,6 +837,258 @@
837
  "eval_samples_per_second": 25.008,
838
  "eval_steps_per_second": 12.505,
839
  "step": 114
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
840
  }
841
  ],
842
  "logging_steps": 1,
@@ -851,12 +1103,12 @@
851
  "should_evaluate": false,
852
  "should_log": false,
853
  "should_save": true,
854
- "should_training_stop": false
855
  },
856
  "attributes": {}
857
  }
858
  },
859
- "total_flos": 510358352560128.0,
860
  "train_batch_size": 2,
861
  "trial_name": null,
862
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.00509805254392822,
5
  "eval_steps": 38,
6
+ "global_step": 150,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
837
  "eval_samples_per_second": 25.008,
838
  "eval_steps_per_second": 12.505,
839
  "step": 114
840
+ },
841
+ {
842
+ "epoch": 0.0039085069503449685,
843
+ "grad_norm": 7067.63134765625,
844
+ "learning_rate": 5.857864376269051e-05,
845
+ "loss": 8.026,
846
+ "step": 115
847
+ },
848
+ {
849
+ "epoch": 0.00394249396730449,
850
+ "grad_norm": 10362.53515625,
851
+ "learning_rate": 5.544102723452171e-05,
852
+ "loss": 7.5678,
853
+ "step": 116
854
+ },
855
+ {
856
+ "epoch": 0.0039764809842640116,
857
+ "grad_norm": 3495.585693359375,
858
+ "learning_rate": 5.237620050253189e-05,
859
+ "loss": 7.1326,
860
+ "step": 117
861
+ },
862
+ {
863
+ "epoch": 0.004010468001223532,
864
+ "grad_norm": 10234.3955078125,
865
+ "learning_rate": 4.938570679927783e-05,
866
+ "loss": 8.7901,
867
+ "step": 118
868
+ },
869
+ {
870
+ "epoch": 0.004044455018183054,
871
+ "grad_norm": 4103.04052734375,
872
+ "learning_rate": 4.647105192839778e-05,
873
+ "loss": 8.4194,
874
+ "step": 119
875
+ },
876
+ {
877
+ "epoch": 0.004078442035142575,
878
+ "grad_norm": 6749.89208984375,
879
+ "learning_rate": 4.363370350639404e-05,
880
+ "loss": 7.9044,
881
+ "step": 120
882
+ },
883
+ {
884
+ "epoch": 0.004112429052102097,
885
+ "grad_norm": 3126.854248046875,
886
+ "learning_rate": 4.087509022364382e-05,
887
+ "loss": 7.5001,
888
+ "step": 121
889
+ },
890
+ {
891
+ "epoch": 0.004146416069061618,
892
+ "grad_norm": 4431.7529296875,
893
+ "learning_rate": 3.819660112501053e-05,
894
+ "loss": 9.3944,
895
+ "step": 122
896
+ },
897
+ {
898
+ "epoch": 0.00418040308602114,
899
+ "grad_norm": 5783.607421875,
900
+ "learning_rate": 3.5599584910418035e-05,
901
+ "loss": 9.3595,
902
+ "step": 123
903
+ },
904
+ {
905
+ "epoch": 0.0042143901029806615,
906
+ "grad_norm": 4225.36279296875,
907
+ "learning_rate": 3.3085349255739474e-05,
908
+ "loss": 8.6062,
909
+ "step": 124
910
+ },
911
+ {
912
+ "epoch": 0.004248377119940183,
913
+ "grad_norm": 5815.947265625,
914
+ "learning_rate": 3.0655160154343174e-05,
915
+ "loss": 6.9025,
916
+ "step": 125
917
+ },
918
+ {
919
+ "epoch": 0.0042823641368997046,
920
+ "grad_norm": 4482.81884765625,
921
+ "learning_rate": 2.831024127962678e-05,
922
+ "loss": 8.1797,
923
+ "step": 126
924
+ },
925
+ {
926
+ "epoch": 0.004316351153859226,
927
+ "grad_norm": 6426.4365234375,
928
+ "learning_rate": 2.6051773368860934e-05,
929
+ "loss": 9.5757,
930
+ "step": 127
931
+ },
932
+ {
933
+ "epoch": 0.004350338170818748,
934
+ "grad_norm": 4510.48681640625,
935
+ "learning_rate": 2.38808936286524e-05,
936
+ "loss": 7.8607,
937
+ "step": 128
938
+ },
939
+ {
940
+ "epoch": 0.004384325187778268,
941
+ "grad_norm": 5353.71875,
942
+ "learning_rate": 2.1798695162326442e-05,
943
+ "loss": 9.2137,
944
+ "step": 129
945
+ },
946
+ {
947
+ "epoch": 0.00441831220473779,
948
+ "grad_norm": 14885.669921875,
949
+ "learning_rate": 1.9806226419516192e-05,
950
+ "loss": 8.0031,
951
+ "step": 130
952
+ },
953
+ {
954
+ "epoch": 0.004452299221697311,
955
+ "grad_norm": 3923.73486328125,
956
+ "learning_rate": 1.790449066823683e-05,
957
+ "loss": 8.4159,
958
+ "step": 131
959
+ },
960
+ {
961
+ "epoch": 0.004486286238656833,
962
+ "grad_norm": 2896.96337890625,
963
+ "learning_rate": 1.6094445489709885e-05,
964
+ "loss": 8.781,
965
+ "step": 132
966
+ },
967
+ {
968
+ "epoch": 0.0045202732556163545,
969
+ "grad_norm": 4708.7001953125,
970
+ "learning_rate": 1.4377002296192233e-05,
971
+ "loss": 7.7557,
972
+ "step": 133
973
+ },
974
+ {
975
+ "epoch": 0.004554260272575876,
976
+ "grad_norm": 3181.123779296875,
977
+ "learning_rate": 1.275302587205256e-05,
978
+ "loss": 8.2043,
979
+ "step": 134
980
+ },
981
+ {
982
+ "epoch": 0.0045882472895353976,
983
+ "grad_norm": 4846.9462890625,
984
+ "learning_rate": 1.1223333938326485e-05,
985
+ "loss": 7.4767,
986
+ "step": 135
987
+ },
988
+ {
989
+ "epoch": 0.004622234306494919,
990
+ "grad_norm": 6198.0390625,
991
+ "learning_rate": 9.788696740969295e-06,
992
+ "loss": 9.4116,
993
+ "step": 136
994
+ },
995
+ {
996
+ "epoch": 0.004656221323454441,
997
+ "grad_norm": 3074.818115234375,
998
+ "learning_rate": 8.44983666301391e-06,
999
+ "loss": 8.0024,
1000
+ "step": 137
1001
+ },
1002
+ {
1003
+ "epoch": 0.004690208340413962,
1004
+ "grad_norm": 5073.27734375,
1005
+ "learning_rate": 7.2074278608293525e-06,
1006
+ "loss": 10.4057,
1007
+ "step": 138
1008
+ },
1009
+ {
1010
+ "epoch": 0.004724195357373484,
1011
+ "grad_norm": 4744.4404296875,
1012
+ "learning_rate": 6.062095924662625e-06,
1013
+ "loss": 8.419,
1014
+ "step": 139
1015
+ },
1016
+ {
1017
+ "epoch": 0.004758182374333004,
1018
+ "grad_norm": 3285.78076171875,
1019
+ "learning_rate": 5.0144175636352765e-06,
1020
+ "loss": 7.7747,
1021
+ "step": 140
1022
+ },
1023
+ {
1024
+ "epoch": 0.004792169391292526,
1025
+ "grad_norm": 3387.646484375,
1026
+ "learning_rate": 4.064920315352904e-06,
1027
+ "loss": 9.9511,
1028
+ "step": 141
1029
+ },
1030
+ {
1031
+ "epoch": 0.0048261564082520475,
1032
+ "grad_norm": 4019.52587890625,
1033
+ "learning_rate": 3.2140822802740668e-06,
1034
+ "loss": 8.495,
1035
+ "step": 142
1036
+ },
1037
+ {
1038
+ "epoch": 0.004860143425211569,
1039
+ "grad_norm": 4878.927734375,
1040
+ "learning_rate": 2.462331880972468e-06,
1041
+ "loss": 7.6559,
1042
+ "step": 143
1043
+ },
1044
+ {
1045
+ "epoch": 0.0048941304421710905,
1046
+ "grad_norm": 8631.8037109375,
1047
+ "learning_rate": 1.81004764641306e-06,
1048
+ "loss": 9.9675,
1049
+ "step": 144
1050
+ },
1051
+ {
1052
+ "epoch": 0.004928117459130612,
1053
+ "grad_norm": 4712.03515625,
1054
+ "learning_rate": 1.2575580213514792e-06,
1055
+ "loss": 9.007,
1056
+ "step": 145
1057
+ },
1058
+ {
1059
+ "epoch": 0.004962104476090134,
1060
+ "grad_norm": 4180.6337890625,
1061
+ "learning_rate": 8.051412009521864e-07,
1062
+ "loss": 7.0792,
1063
+ "step": 146
1064
+ },
1065
+ {
1066
+ "epoch": 0.004996091493049655,
1067
+ "grad_norm": 3335.519287109375,
1068
+ "learning_rate": 4.530249907087836e-07,
1069
+ "loss": 7.036,
1070
+ "step": 147
1071
+ },
1072
+ {
1073
+ "epoch": 0.005030078510009177,
1074
+ "grad_norm": 3107.85595703125,
1075
+ "learning_rate": 2.0138669173708213e-07,
1076
+ "loss": 8.3826,
1077
+ "step": 148
1078
+ },
1079
+ {
1080
+ "epoch": 0.005064065526968698,
1081
+ "grad_norm": 3581.00927734375,
1082
+ "learning_rate": 5.035301149869387e-08,
1083
+ "loss": 7.9239,
1084
+ "step": 149
1085
+ },
1086
+ {
1087
+ "epoch": 0.00509805254392822,
1088
+ "grad_norm": 4885.18798828125,
1089
+ "learning_rate": 0.0,
1090
+ "loss": 8.5003,
1091
+ "step": 150
1092
  }
1093
  ],
1094
  "logging_steps": 1,
 
1103
  "should_evaluate": false,
1104
  "should_log": false,
1105
  "should_save": true,
1106
+ "should_training_stop": true
1107
  },
1108
  "attributes": {}
1109
  }
1110
  },
1111
+ "total_flos": 672217071353856.0,
1112
  "train_batch_size": 2,
1113
  "trial_name": null,
1114
  "trial_params": null