adhi29 commited on
Commit
c9e7a06
1 Parent(s): e14bebe

adhi29/openhermes-mistral-dpo-gptq

Browse files
README.md CHANGED
@@ -17,15 +17,15 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [TheBloke/OpenHermes-2-Mistral-7B-GPTQ](https://huggingface.co/TheBloke/OpenHermes-2-Mistral-7B-GPTQ) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.6522
21
- - Rewards/chosen: 0.0746
22
- - Rewards/rejected: 0.0188
23
- - Rewards/accuracies: 0.5625
24
- - Rewards/margins: 0.0558
25
- - Logps/rejected: -85.2534
26
- - Logps/chosen: -70.1958
27
- - Logits/rejected: -2.6062
28
- - Logits/chosen: -2.5855
29
 
30
  ## Model description
31
 
@@ -51,18 +51,17 @@ The following hyperparameters were used during training:
51
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
  - lr_scheduler_type: linear
53
  - lr_scheduler_warmup_steps: 2
54
- - training_steps: 50
55
  - mixed_precision_training: Native AMP
56
 
57
  ### Training results
58
 
59
  | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
60
  |:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
61
- | 0.6707 | 0.01 | 10 | 0.6784 | 0.0243 | 0.0096 | 0.375 | 0.0147 | -85.3454 | -70.6989 | -2.6043 | -2.5836 |
62
- | 0.6955 | 0.01 | 20 | 0.6748 | 0.0239 | 0.0210 | 0.3125 | 0.0029 | -85.2314 | -70.7031 | -2.6048 | -2.5849 |
63
- | 0.7423 | 0.01 | 30 | 0.6740 | 0.0542 | 0.0384 | 0.5625 | 0.0158 | -85.0572 | -70.4001 | -2.6053 | -2.5858 |
64
- | 0.7023 | 0.02 | 40 | 0.6620 | 0.0731 | 0.0323 | 0.5625 | 0.0407 | -85.1185 | -70.2116 | -2.6061 | -2.5863 |
65
- | 0.685 | 0.03 | 50 | 0.6522 | 0.0746 | 0.0188 | 0.5625 | 0.0558 | -85.2534 | -70.1958 | -2.6062 | -2.5855 |
66
 
67
 
68
  ### Framework versions
 
17
 
18
  This model is a fine-tuned version of [TheBloke/OpenHermes-2-Mistral-7B-GPTQ](https://huggingface.co/TheBloke/OpenHermes-2-Mistral-7B-GPTQ) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.9866
21
+ - Rewards/chosen: 0.2715
22
+ - Rewards/rejected: 0.5084
23
+ - Rewards/accuracies: 0.625
24
+ - Rewards/margins: -0.2369
25
+ - Logps/rejected: -217.0748
26
+ - Logps/chosen: -192.7873
27
+ - Logits/rejected: -2.1497
28
+ - Logits/chosen: -2.0212
29
 
30
  ## Model description
31
 
 
51
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
  - lr_scheduler_type: linear
53
  - lr_scheduler_warmup_steps: 2
54
+ - training_steps: 40
55
  - mixed_precision_training: Native AMP
56
 
57
  ### Training results
58
 
59
  | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
60
  |:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
61
+ | 0.6752 | 0.01 | 10 | 0.7338 | 0.0443 | 0.0693 | 0.875 | -0.0250 | -221.4665 | -195.0593 | -2.1454 | -2.0106 |
62
+ | 0.71 | 0.01 | 20 | 0.7099 | 0.0825 | 0.0676 | 0.875 | 0.0149 | -221.4828 | -194.6768 | -2.1435 | -2.0127 |
63
+ | 0.6938 | 0.01 | 30 | 0.8421 | 0.1926 | 0.3222 | 0.625 | -0.1296 | -218.9368 | -193.5758 | -2.1482 | -2.0177 |
64
+ | 0.6923 | 0.02 | 40 | 0.9866 | 0.2715 | 0.5084 | 0.625 | -0.2369 | -217.0748 | -192.7873 | -2.1497 | -2.0212 |
 
65
 
66
 
67
  ### Framework versions
adapter_config.json CHANGED
@@ -19,8 +19,8 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "q_proj",
23
- "v_proj"
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "v_proj",
23
+ "q_proj"
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed2bd803830f05aed17991050c376ef747a8829b0d969d4141a8810e47adaa2b
3
  size 13648432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c98da1863b9dadf52f7f528d729e4c15130e1da81c318696bd8b3a409507ab5b
3
  size 13648432
runs/Jan09_03-58-46_e1dc5e887cb9/events.out.tfevents.1704772817.e1dc5e887cb9.264.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2acafb1886624c72c035a78bb9c445d90a30c3e9e02c835d76e23f8eea7c2ab6
3
+ size 11244
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:accf2b53d9c6289fc381d015c0585ad7bcbb44c69f6389d73604727ffdd55f6c
3
  size 4155
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b99c58734a5e5e9083464ec5d565517cbaf9a7e4545a53ff8e9a45bab558ed9
3
  size 4155