End of training
Browse files- adapter_config.json +5 -5
- adapter_model.safetensors +1 -1
- all_results.json +4 -4
- train_results.json +4 -4
- trainer_state.json +53 -53
- training_args.bin +1 -1
adapter_config.json
CHANGED
|
@@ -24,13 +24,13 @@
|
|
| 24 |
"rank_pattern": {},
|
| 25 |
"revision": null,
|
| 26 |
"target_modules": [
|
| 27 |
-
"v_proj",
|
| 28 |
"gate_proj",
|
| 29 |
-
"up_proj",
|
| 30 |
-
"q_proj",
|
| 31 |
-
"k_proj",
|
| 32 |
"down_proj",
|
| 33 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
| 34 |
],
|
| 35 |
"task_type": "CAUSAL_LM",
|
| 36 |
"trainable_token_indices": null,
|
|
|
|
| 24 |
"rank_pattern": {},
|
| 25 |
"revision": null,
|
| 26 |
"target_modules": [
|
|
|
|
| 27 |
"gate_proj",
|
|
|
|
|
|
|
|
|
|
| 28 |
"down_proj",
|
| 29 |
+
"v_proj",
|
| 30 |
+
"k_proj",
|
| 31 |
+
"q_proj",
|
| 32 |
+
"o_proj",
|
| 33 |
+
"up_proj"
|
| 34 |
],
|
| 35 |
"task_type": "CAUSAL_LM",
|
| 36 |
"trainable_token_indices": null,
|
adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 536991984
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7ca363c1b5edad2ba4196a98e67ea59bd0227a147540d4eca1842fecc859fd50
|
| 3 |
size 536991984
|
all_results.json
CHANGED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"total_flos": 0.0,
|
| 3 |
-
"train_loss":
|
| 4 |
-
"train_runtime":
|
| 5 |
"train_samples": 9,
|
| 6 |
-
"train_samples_per_second": 0.
|
| 7 |
-
"train_steps_per_second": 0.
|
| 8 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"total_flos": 0.0,
|
| 3 |
+
"train_loss": 8.153455610226957e-07,
|
| 4 |
+
"train_runtime": 2152.5231,
|
| 5 |
"train_samples": 9,
|
| 6 |
+
"train_samples_per_second": 0.037,
|
| 7 |
+
"train_steps_per_second": 0.005
|
| 8 |
}
|
train_results.json
CHANGED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"total_flos": 0.0,
|
| 3 |
-
"train_loss":
|
| 4 |
-
"train_runtime":
|
| 5 |
"train_samples": 9,
|
| 6 |
-
"train_samples_per_second": 0.
|
| 7 |
-
"train_steps_per_second": 0.
|
| 8 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"total_flos": 0.0,
|
| 3 |
+
"train_loss": 8.153455610226957e-07,
|
| 4 |
+
"train_runtime": 2152.5231,
|
| 5 |
"train_samples": 9,
|
| 6 |
+
"train_samples_per_second": 0.037,
|
| 7 |
+
"train_steps_per_second": 0.005
|
| 8 |
}
|
trainer_state.json
CHANGED
|
@@ -10,108 +10,108 @@
|
|
| 10 |
"is_world_process_zero": true,
|
| 11 |
"log_history": [
|
| 12 |
{
|
| 13 |
-
"completion_length":
|
| 14 |
"epoch": 0.8888888888888888,
|
| 15 |
-
"grad_norm": 0.
|
| 16 |
-
"kl": 0.
|
| 17 |
"learning_rate": 5e-07,
|
| 18 |
-
"loss":
|
| 19 |
-
"reward":
|
| 20 |
-
"reward_std": 0.
|
| 21 |
-
"rewards/concensus_correctness_reward_func":
|
| 22 |
"rewards/consensus_reward_func": 2.0,
|
| 23 |
"rewards/cumulative_reward_2": 0.0,
|
| 24 |
-
"rewards/final_correctness_reward_func":
|
| 25 |
-
"rewards/question_recreation_reward_func": 0.
|
| 26 |
"rewards/soft_format_reward_func": 0.0,
|
| 27 |
"rewards/strict_format_reward_func": 0.0,
|
| 28 |
-
"rewards/xmlcount_reward_func": -0.
|
| 29 |
"step": 2
|
| 30 |
},
|
| 31 |
{
|
| 32 |
-
"completion_length":
|
| 33 |
"epoch": 1.4444444444444444,
|
| 34 |
-
"grad_norm": 0.
|
| 35 |
-
"kl": 0.
|
| 36 |
"learning_rate": 4.415111107797445e-07,
|
| 37 |
-
"loss": 0.0,
|
| 38 |
-
"reward":
|
| 39 |
-
"reward_std": 0.
|
| 40 |
-
"rewards/concensus_correctness_reward_func":
|
| 41 |
"rewards/consensus_reward_func": 2.0,
|
| 42 |
"rewards/cumulative_reward_2": 0.0,
|
| 43 |
-
"rewards/final_correctness_reward_func": 0
|
| 44 |
-
"rewards/question_recreation_reward_func": 0.
|
| 45 |
"rewards/soft_format_reward_func": 0.0,
|
| 46 |
"rewards/strict_format_reward_func": 0.0,
|
| 47 |
-
"rewards/xmlcount_reward_func": -0.
|
| 48 |
"step": 4
|
| 49 |
},
|
| 50 |
{
|
| 51 |
-
"completion_length":
|
| 52 |
"epoch": 2.0,
|
| 53 |
-
"grad_norm": 0.
|
| 54 |
-
"kl": 0.
|
| 55 |
"learning_rate": 2.934120444167326e-07,
|
| 56 |
"loss": 0.0,
|
| 57 |
-
"reward":
|
| 58 |
-
"reward_std": 0.
|
| 59 |
-
"rewards/concensus_correctness_reward_func": 0.
|
| 60 |
"rewards/consensus_reward_func": 2.0,
|
| 61 |
"rewards/cumulative_reward_2": 0.0,
|
| 62 |
-
"rewards/final_correctness_reward_func":
|
| 63 |
-
"rewards/question_recreation_reward_func": 0
|
| 64 |
"rewards/soft_format_reward_func": 0.0,
|
| 65 |
"rewards/strict_format_reward_func": 0.0,
|
| 66 |
-
"rewards/xmlcount_reward_func":
|
| 67 |
"step": 6
|
| 68 |
},
|
| 69 |
{
|
| 70 |
-
"completion_length":
|
| 71 |
"epoch": 2.888888888888889,
|
| 72 |
-
"grad_norm": 0.
|
| 73 |
-
"kl": 0.
|
| 74 |
"learning_rate": 1.2500000000000005e-07,
|
| 75 |
"loss": 0.0,
|
| 76 |
-
"reward":
|
| 77 |
-
"reward_std": 0.
|
| 78 |
-
"rewards/concensus_correctness_reward_func":
|
| 79 |
-
"rewards/consensus_reward_func":
|
| 80 |
"rewards/cumulative_reward_2": 0.0,
|
| 81 |
-
"rewards/final_correctness_reward_func":
|
| 82 |
-
"rewards/question_recreation_reward_func": 0.
|
| 83 |
"rewards/soft_format_reward_func": 0.0,
|
| 84 |
"rewards/strict_format_reward_func": 0.0,
|
| 85 |
-
"rewards/xmlcount_reward_func": -0.
|
| 86 |
"step": 8
|
| 87 |
},
|
| 88 |
{
|
| 89 |
-
"completion_length":
|
| 90 |
"epoch": 3.4444444444444446,
|
| 91 |
-
"grad_norm": 0.
|
| 92 |
-
"kl": 0.
|
| 93 |
"learning_rate": 1.507684480352292e-08,
|
| 94 |
"loss": 0.0,
|
| 95 |
-
"reward":
|
| 96 |
-
"reward_std": 0.
|
| 97 |
-
"rewards/concensus_correctness_reward_func": 0.
|
| 98 |
"rewards/consensus_reward_func": 2.0,
|
| 99 |
"rewards/cumulative_reward_2": 0.0,
|
| 100 |
-
"rewards/final_correctness_reward_func":
|
| 101 |
-
"rewards/question_recreation_reward_func": 0.
|
| 102 |
"rewards/soft_format_reward_func": 0.0,
|
| 103 |
-
"rewards/strict_format_reward_func": 0.
|
| 104 |
-
"rewards/xmlcount_reward_func": -0.
|
| 105 |
"step": 10
|
| 106 |
},
|
| 107 |
{
|
| 108 |
"epoch": 3.4444444444444446,
|
| 109 |
"step": 10,
|
| 110 |
"total_flos": 0.0,
|
| 111 |
-
"train_loss":
|
| 112 |
-
"train_runtime":
|
| 113 |
-
"train_samples_per_second": 0.
|
| 114 |
-
"train_steps_per_second": 0.
|
| 115 |
}
|
| 116 |
],
|
| 117 |
"logging_steps": 2,
|
|
|
|
| 10 |
"is_world_process_zero": true,
|
| 11 |
"log_history": [
|
| 12 |
{
|
| 13 |
+
"completion_length": 480.125,
|
| 14 |
"epoch": 0.8888888888888888,
|
| 15 |
+
"grad_norm": 0.17998529970645905,
|
| 16 |
+
"kl": 0.000938359007705003,
|
| 17 |
"learning_rate": 5e-07,
|
| 18 |
+
"loss": 0.0,
|
| 19 |
+
"reward": 4.501521214842796,
|
| 20 |
+
"reward_std": 0.16148228733800352,
|
| 21 |
+
"rewards/concensus_correctness_reward_func": 0.6350000090897083,
|
| 22 |
"rewards/consensus_reward_func": 2.0,
|
| 23 |
"rewards/cumulative_reward_2": 0.0,
|
| 24 |
+
"rewards/final_correctness_reward_func": 1.25,
|
| 25 |
+
"rewards/question_recreation_reward_func": 0.9901463091373444,
|
| 26 |
"rewards/soft_format_reward_func": 0.0,
|
| 27 |
"rewards/strict_format_reward_func": 0.0,
|
| 28 |
+
"rewards/xmlcount_reward_func": -0.373624999076128,
|
| 29 |
"step": 2
|
| 30 |
},
|
| 31 |
{
|
| 32 |
+
"completion_length": 547.1,
|
| 33 |
"epoch": 1.4444444444444444,
|
| 34 |
+
"grad_norm": 0.13144199550151825,
|
| 35 |
+
"kl": 0.0008038122265134007,
|
| 36 |
"learning_rate": 4.415111107797445e-07,
|
| 37 |
+
"loss": -0.0,
|
| 38 |
+
"reward": 6.170431995391846,
|
| 39 |
+
"reward_std": 0.15375217869877816,
|
| 40 |
+
"rewards/concensus_correctness_reward_func": 1.6456000626087188,
|
| 41 |
"rewards/consensus_reward_func": 2.0,
|
| 42 |
"rewards/cumulative_reward_2": 0.0,
|
| 43 |
+
"rewards/final_correctness_reward_func": 2.0,
|
| 44 |
+
"rewards/question_recreation_reward_func": 0.9895318984985352,
|
| 45 |
"rewards/soft_format_reward_func": 0.0,
|
| 46 |
"rewards/strict_format_reward_func": 0.0,
|
| 47 |
+
"rewards/xmlcount_reward_func": -0.4646999955177307,
|
| 48 |
"step": 4
|
| 49 |
},
|
| 50 |
{
|
| 51 |
+
"completion_length": 445.6,
|
| 52 |
"epoch": 2.0,
|
| 53 |
+
"grad_norm": 0.0919833555817604,
|
| 54 |
+
"kl": 0.0012037293054163456,
|
| 55 |
"learning_rate": 2.934120444167326e-07,
|
| 56 |
"loss": 0.0,
|
| 57 |
+
"reward": 4.09589991569519,
|
| 58 |
+
"reward_std": 0.12713775224983692,
|
| 59 |
+
"rewards/concensus_correctness_reward_func": 0.22239999771118163,
|
| 60 |
"rewards/consensus_reward_func": 2.0,
|
| 61 |
"rewards/cumulative_reward_2": 0.0,
|
| 62 |
+
"rewards/final_correctness_reward_func": 0.8,
|
| 63 |
+
"rewards/question_recreation_reward_func": 1.0,
|
| 64 |
"rewards/soft_format_reward_func": 0.0,
|
| 65 |
"rewards/strict_format_reward_func": 0.0,
|
| 66 |
+
"rewards/xmlcount_reward_func": 0.07349999248981476,
|
| 67 |
"step": 6
|
| 68 |
},
|
| 69 |
{
|
| 70 |
+
"completion_length": 483.75,
|
| 71 |
"epoch": 2.888888888888889,
|
| 72 |
+
"grad_norm": 0.25922873616218567,
|
| 73 |
+
"kl": 0.000980880402494222,
|
| 74 |
"learning_rate": 1.2500000000000005e-07,
|
| 75 |
"loss": 0.0,
|
| 76 |
+
"reward": 4.644306495785713,
|
| 77 |
+
"reward_std": 0.6483020558953285,
|
| 78 |
+
"rewards/concensus_correctness_reward_func": 0.7606250215321779,
|
| 79 |
+
"rewards/consensus_reward_func": 1.75,
|
| 80 |
"rewards/cumulative_reward_2": 0.0,
|
| 81 |
+
"rewards/final_correctness_reward_func": 1.5,
|
| 82 |
+
"rewards/question_recreation_reward_func": 0.9843065515160561,
|
| 83 |
"rewards/soft_format_reward_func": 0.0,
|
| 84 |
"rewards/strict_format_reward_func": 0.0,
|
| 85 |
+
"rewards/xmlcount_reward_func": -0.35062500834465027,
|
| 86 |
"step": 8
|
| 87 |
},
|
| 88 |
{
|
| 89 |
+
"completion_length": 547.4,
|
| 90 |
"epoch": 3.4444444444444446,
|
| 91 |
+
"grad_norm": 0.3769736886024475,
|
| 92 |
+
"kl": 0.0010651831165887416,
|
| 93 |
"learning_rate": 1.507684480352292e-08,
|
| 94 |
"loss": 0.0,
|
| 95 |
+
"reward": 3.190308856964111,
|
| 96 |
+
"reward_std": 0.7591730035841465,
|
| 97 |
+
"rewards/concensus_correctness_reward_func": -0.0031999826431274416,
|
| 98 |
"rewards/consensus_reward_func": 2.0,
|
| 99 |
"rewards/cumulative_reward_2": 0.0,
|
| 100 |
+
"rewards/final_correctness_reward_func": 0.4,
|
| 101 |
+
"rewards/question_recreation_reward_func": 0.983308744430542,
|
| 102 |
"rewards/soft_format_reward_func": 0.0,
|
| 103 |
+
"rewards/strict_format_reward_func": 0.05,
|
| 104 |
+
"rewards/xmlcount_reward_func": -0.23979999423027037,
|
| 105 |
"step": 10
|
| 106 |
},
|
| 107 |
{
|
| 108 |
"epoch": 3.4444444444444446,
|
| 109 |
"step": 10,
|
| 110 |
"total_flos": 0.0,
|
| 111 |
+
"train_loss": 8.153455610226957e-07,
|
| 112 |
+
"train_runtime": 2152.5231,
|
| 113 |
+
"train_samples_per_second": 0.037,
|
| 114 |
+
"train_steps_per_second": 0.005
|
| 115 |
}
|
| 116 |
],
|
| 117 |
"logging_steps": 2,
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 6072
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:671653c2fa91511c2714ef5e63574c1c3d0154148fa5cf944a86fa0f99b4ac34
|
| 3 |
size 6072
|