zerofata commited on
Commit
2bf7260
·
verified ·
1 Parent(s): 3713f5b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +43 -13
README.md CHANGED
@@ -4,32 +4,62 @@ base_model:
4
  - EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1
5
  - LatitudeGames/Wayfarer-Large-70B-Llama-3.3
6
  - crestf411/L3.1-nemotron-sunfall-v0.7.0
 
 
 
 
7
  library_name: transformers
8
  tags:
9
  - mergekit
10
  - merge
11
-
12
  ---
13
- # merge_v7b_full
 
 
 
 
 
 
14
 
15
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
16
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  ## Merge Details
18
  ### Merge Method
19
 
20
  This model was merged using the [SCE](https://arxiv.org/abs/2408.07990) merge method using ./merge_v6_base_E as a base.
21
 
22
- ### Models Merged
23
-
24
- The following models were included in the merge:
25
- * [Sao10K/L3.1-70B-Hanami-x1](https://huggingface.co/Sao10K/L3.1-70B-Hanami-x1)
26
- * [EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1](https://huggingface.co/EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1)
27
- * [LatitudeGames/Wayfarer-Large-70B-Llama-3.3](https://huggingface.co/LatitudeGames/Wayfarer-Large-70B-Llama-3.3)
28
- * [crestf411/L3.1-nemotron-sunfall-v0.7.0](https://huggingface.co/crestf411/L3.1-nemotron-sunfall-v0.7.0)
29
-
30
- ### Configuration
 
 
 
 
 
 
31
 
32
- The following YAML configuration was used to produce this model:
33
 
34
  ```yaml
35
  models:
@@ -44,4 +74,4 @@ out_dtype: bfloat16
44
  dype: float32
45
  tokenizer:
46
  source: union
47
- ```
 
4
  - EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1
5
  - LatitudeGames/Wayfarer-Large-70B-Llama-3.3
6
  - crestf411/L3.1-nemotron-sunfall-v0.7.0
7
+ - tokyotech-llm/Llama-3.1-Swallow-70B-Instruct-v0.3
8
+ - SicariusSicariiStuff/Negative_LLAMA_70B
9
+ - nbeerbower/llama3.1-kartoffeldes-70B
10
+ - deepseek-ai/DeepSeek-R1-Distill-Llama-70B
11
  library_name: transformers
12
  tags:
13
  - mergekit
14
  - merge
 
15
  ---
16
+ # Genetic Lemonade Unleashed
17
+
18
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/65b19c6c638328850e12d38c/P8HgQAzAjEWE67u9sSKJz.png)
19
+
20
+ This model is the final result of a few dozen different merges during my process of learning how to merge. I think it's pretty good.
21
+
22
+ Designed for primarily for RP.
23
 
24
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
25
 
26
+ ## SillyTavern Settings
27
+
28
+ ### Sampler Settings
29
+ - Temp: 1
30
+ - MinP: 0.02-0.05
31
+ - Dry: 0.8, 1.75, 4
32
+
33
+ ### Instruct
34
+
35
+ [Llam@ception](https://huggingface.co/Konnect1221/The-Inception-Presets-Methception-LLamaception-Qwenception/tree/main/Llam%40ception)
36
+
37
+ Any system prompt will work, but their Context / Instruct template I recommend.
38
+
39
+ Alternatively you can use Llama-3-Instruct-Names but you will need to uncheck "System same as user".
40
+
41
  ## Merge Details
42
  ### Merge Method
43
 
44
  This model was merged using the [SCE](https://arxiv.org/abs/2408.07990) merge method using ./merge_v6_base_E as a base.
45
 
46
+ ### merge_v6_base_E
47
+ ```yaml
48
+ models:
49
+ - model: deepseek-ai/DeepSeek-R1-Distill-Llama-70B
50
+ - model: nbeerbower/llama3.1-kartoffeldes-70B
51
+ - model: tokyotech-llm/Llama-3.1-Swallow-70B-Instruct-v0.3
52
+ - model: SicariusSicariiStuff/Negative_LLAMA_70B
53
+ select_topk: .15
54
+ merge_method: sce
55
+ base_model: meta-llama/Llama-3.3-70B-Instruct
56
+ out_dtype: bfloat16
57
+ dype: float32
58
+ tokenizer:
59
+ source: base
60
+ ```
61
 
62
+ ### Genetic Lemonade Unleashed
63
 
64
  ```yaml
65
  models:
 
74
  dype: float32
75
  tokenizer:
76
  source: union
77
+ ```