ArtusDev's picture
Create README.md
c532001 verified
---
base_model:
- yvvki/Erotophobia-24B-v2.0
base_model_relation: quantized
quantized_by: ArtusDev
library_name: transformers
tags:
- mergekit
- merge
---
# Erotophobia-24-v2.0
![Model Banner](banner.png)
My second merge! Yayay!
*Technically my third but we'll ignore the first failed model :(*
This model is just headache to do, it really is! It wasted me $20 trying to fit darkc0de/BlackXorDolphTronGOAT because it's a `float32` dtype!
I didn't know I can downcast the model into `dfloat16` first until at the very end there.
But, this model has the downcast in the [df16](df16) directory for your pleasure.
> Still testing and awaiting GGUF. Please kindly give your patience. Thank you <3
## Philosophy
### Mind
Fusion of darkc0de's fusions merge, and Acree's DeepSeek distill abliterated.
I think this is a good base. The Xortron is the top performing at UGI leaderboard and Arcee has the DeepSeek distill and updated world information.
### Heart
Karcher with all 2503 base, using updated Dan's and Eurydice, and reintroduce Pantheon.
I feel like this will improve the roleplay. All has the models have unique characteristics of their own, and hopefully the Karcher mean can find a nice center.
### Soul
Stock on Sleep's Omega Gaslight and Broken Tutu, since both based on Cydonia 2.1 and have the BlackSheep model in their merge.
I'm horny...
## Merge Details
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
### Merge Method
This model was merged using multi-stage model merging method:
- **Mind** was merged using the [Arcee Fusion](https://www.arcee.ai/blog/meet-mergekit-v0-1-arcee-fusion-expanded-model-support-multi-gpu-acceleration) merge method using huihui-ai/Arcee-Blitz-abliterated as a base.
- **Heart** was merged using the [Karcher Mean](https://en.wikipedia.org/wiki/Karcher_mean) merge method.
- **Soul** was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using TheDrummer/Cydonia-24B-v2.1 as a base.
Fially, this model was merged using the [DELLA](https://arxiv.org/abs/2406.11617) merge method using **Mind** as a base.
The partial merge results (**Mind**, **Heart**, and **Soul**) are available inside the [intermediates](intermediates) directory.
### Models Merged
The following models were included in the merge:
* **Mind**
- darkc0de/BlackXorDolphTronGOAT
* **Soul**
- PocketDoc/Dans-PersonalityEngine-V1.3.0-24b
- aixonlab/Eurydice-24b-v3.5
- Gryphe/Pantheon-RP-1.8-24b-Small-3.1
* **Heart**
- ReadyArt/Broken-Tutu-24B
- ReadyArt/Omega-Darker-Gaslight_The-Final-Forgotten-Fever-Dream-24B
### Configuration
The following YAML configuration was used to produce this model (using `mergekit-multi`):
```yaml
name: Mind
merge_method: arcee_fusion
dtype: bfloat16
tokenizer:
source: union
chat_template: auto
base_model: huihui-ai/Arcee-Blitz-abliterated
models:
- model: darkc0de/BlackXorDolphTronGOAT
---
name: Heart
merge_method: karcher
tokenizer:
source: union
chat_template: auto
parameters:
max_iter: 1000
models:
- model: PocketDoc/Dans-PersonalityEngine-V1.3.0-24b
- model: aixonlab/Eurydice-24b-v3.5
- model: Gryphe/Pantheon-RP-1.8-24b-Small-3.1
---
name: Soul
merge_method: model_stock
tokenizer:
source: union
chat_template: auto
base_model: TheDrummer/Cydonia-24B-v2.1
models:
- model: ReadyArt/Broken-Tutu-24B
- model: ReadyArt/Omega-Darker-Gaslight_The-Final-Forgotten-Fever-Dream-24B
---
merge_method: della
tokenizer:
source: union
chat_template: auto
base_model: Mind
models:
- model: Mind
- model: Heart
parameters:
weight: 0.6
- model: Soul
parameters:
weight: 0.4
parameters:
density: 0.7
epsilon: 0.2
lamda: 1.1
```