zerofata's picture
Update README.md
345a848 verified
---
base_model:
- zerofata/L3.3-GeneticLemonade-Final-70B
library_name: transformers
tags:
- mergekit
- merge
---
# Genetic Lemonade Final
![image/png](https://cdn-uploads.huggingface.co/production/uploads/65b19c6c638328850e12d38c/0-pxxp_xcaIecQFcaIzFI.png)
Inspired to learn how to merge by the Nevoria series from [SteelSkull](https://huggingface.co/Steelskull).
This model is the final result of the Genetic Lemonade series.
This model is the second result of the Genetic Lemonade series.
## Model Comparison
Designed for RP and creative writing, all three models are focused around striking a balance between writing style, creativity and intelligence. The basic differences between the models are below.
| Version | Strength | Weakness |
|---------|----------------|----|
| Unleashed | Well balanced | Somewhat censored |
| **Final** | Fully uncensored | Least intelligent |
| Sunset | Well balanced, most intelligent | GPTisms / weakest writing style |
## SillyTavern Settings
[Llam@ception](https://huggingface.co/Konnect1221/The-Inception-Presets-Methception-LLamaception-Qwenception/tree/main/Llam%40ception) recommended for sane defaults if unsure, import them to SillyTavern and they're plug n play.
### Sampler Settings
- Temp: 0.9-1.0
- MinP: 0.03-0.05
- Dry: 0.8, 1.75, 4
Temperature last, neutralize other samplers. This model natively strikes a balance of creativity & intelligence.
### Instruct
Llama-3-Instruct-Names but you will need to uncheck "System same as user".
## Quants
### GGUF
- [Static quants by mradermacher](https://huggingface.co/mradermacher/L3.3-GeneticLemonade-Final-70B-GGUF)
- [iMatrix quants by mradermacher](https://huggingface.co/mradermacher/L3.3-GeneticLemonade-Final-70B-i1-GGUF)
### EXL2
- [4bpw](https://huggingface.co/zerofata/L3.3-GeneticLemonade-Final-70B-4bpw-h6-exl2)
- [4.5bpw](https://huggingface.co/zerofata/L3.3-GeneticLemonade-Final-70B-4.5bpw-h6-exl2)
- [6bpw](https://huggingface.co/zerofata/L3.3-GeneticLemonade-Final-70B-6bpw-h8-exl2)
## Merge Details
### Merge Method
This model was merged using the [SCE](https://arxiv.org/abs/2408.07990) merge method.
The base aims to build a strong general purpose model using high performing models that are trained on various datasets from different languages / cultures. This is to reduce the chance of the same datasets appearing multiple times to build natural creativity into L3.3
The second merge aims to impart specific RP / creative writing knowledge, again focusing on trying to find high performing models that use or likely use different datasets.
### Base_6_v2
```yaml
models:
- model: OpenBuddy/openbuddy-llama3.3-70b-v24.1-131k
- model: nbeerbower/llama3.1-kartoffeldes-70B
- model: tokyotech-llm/Llama-3.1-Swallow-70B-Instruct-v0.3
- model: SicariusSicariiStuff/Negative_LLAMA_70B
select_topk: .15
merge_method: sce
base_model: meta-llama/Llama-3.3-70B-Instruct
out_dtype: bfloat16
dype: float32
tokenizer:
source: base
```
### Genetic Lemonade Final
```yaml
models:
- model: EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1
- model: LatitudeGames/Wayfarer-Large-70B-Llama-3.3
- model: crestf411/L3.1-nemotron-sunfall-v0.7.0
- model: SicariusSicariiStuff/Negative_LLAMA_70B
- model: Sao10K/L3.3-70B-Euryale-v2.3
merge_method: sce
base_model: ./Base_6_v2
select_topk: 0.15
out_dtype: bfloat16
dype: float32
tokenizer:
source: union
```