Triangle104 commited on
Commit
c399670
·
verified ·
1 Parent(s): 31dd7a0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +68 -0
README.md CHANGED
@@ -19,6 +19,74 @@ base_model: crestf411/Q2.5-32B-Slush
19
  This model was converted to GGUF format from [`crestf411/Q2.5-32B-Slush`](https://huggingface.co/crestf411/Q2.5-32B-Slush) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
20
  Refer to the [original model card](https://huggingface.co/crestf411/Q2.5-32B-Slush) for more details on the model.
21
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
  ## Use with llama.cpp
23
  Install llama.cpp through brew (works on Mac and Linux)
24
 
 
19
  This model was converted to GGUF format from [`crestf411/Q2.5-32B-Slush`](https://huggingface.co/crestf411/Q2.5-32B-Slush) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
20
  Refer to the [original model card](https://huggingface.co/crestf411/Q2.5-32B-Slush) for more details on the model.
21
 
22
+ ---
23
+ Model details:
24
+ -
25
+ Slush is a two-stage model trained with high LoRA dropout, where stage 1 is a pretraining continuation on the base model, aimed at boosting the model's creativity and writing capabilities. This is then merged into the instruction tune model, and stage 2 is a fine tuning step on top of this to further enhance its roleplaying capabilities and/or to repair any damage caused in the stage 1 merge.
26
+
27
+ This is still early stage. As always, feedback is welcome, and begone if you demand perfection.
28
+
29
+ The second stage, like the Sunfall series, follows the Silly Tavern preset (ChatML), so ymmv in particular if you use some other tool and/or preset.
30
+ Parameter suggestions
31
+
32
+ I did all my testing with temp 1, min-p 0.1, DRY 0.8, but enabled XTC as context grew and/or the model started saying "the same stuff".
33
+
34
+ Qwen 2.5 32B Instruct (vanilla) has a strong tendency to start speaking for the user, especially in narrator scenarios. I was unable to properly train this out of the model completely, so you may want to add e.g. "\nYou" as a stopping string, and enable "trim incomplete sentences", especially if you have banned sentences.
35
+
36
+ The model has a tendency to add an unnecesary final paragraph to responses during roleplay, sort of like a "summary" of how the character is feeling. Keeping it is OK, but it may snowball quickly. Hoping to address this but not sure how.
37
+ Training details
38
+
39
+ Stage 1 (continued pretraining)
40
+ Target: Qwen/Qwen2.5-32B (resulting LoRA merged into Qwen/Qwen2.5-32B-Instruct)
41
+ LoRA dropout 0.5 (motivation)
42
+ LoRA rank 32, alpha 64 (motivation)
43
+ LR cosine 4e-6
44
+ LoRA+ with LR Ratio: 15
45
+ Context size: 8192
46
+ Gradient accumulation steps: 4
47
+ Epochs: 1
48
+ Stage 2 (fine tune)
49
+ Target: Stage 1 model
50
+ LoRA dropout 0.5
51
+ LoRA rank 32, alpha 64
52
+ LR cosine 5e-6 (min 5e-7)
53
+ LoRA+ with LR Ratio: 15
54
+ Context size: 16384
55
+ Gradient accumulation steps: 4
56
+ Epochs: 1
57
+
58
+ Merge Details
59
+ Merge Method
60
+
61
+ This model was merged using the TIES merge method.
62
+ Configuration
63
+
64
+ The following YAML configuration was used to produce this model:
65
+
66
+ models:
67
+ - model: stage1-model
68
+ parameters:
69
+ weight: 1
70
+ density: 1
71
+ - model: stage2-model
72
+ parameters:
73
+ weight: 1
74
+ density: 1
75
+ - model: Qwen/Qwen2.5-32B-Instruct
76
+ parameters:
77
+ weight: 0.9
78
+ density: 0.9
79
+ merge_method: ties
80
+ base_model: Qwen/Qwen2.5-32B
81
+ parameters:
82
+ weight: 0.9
83
+ density: 0.9
84
+ normalize: true
85
+ int8_mask: true
86
+ tokenizer_source: Qwen/Qwen2.5-32B-Instruct
87
+ dtype: bfloat16
88
+
89
+ ---
90
  ## Use with llama.cpp
91
  Install llama.cpp through brew (works on Mac and Linux)
92