ArtusDev's picture
Create README.md
c532001 verified
metadata
base_model:
  - yvvki/Erotophobia-24B-v2.0
base_model_relation: quantized
quantized_by: ArtusDev
library_name: transformers
tags:
  - mergekit
  - merge

Erotophobia-24-v2.0

Model Banner

My second merge! Yayay!
Technically my third but we'll ignore the first failed model :(

This model is just headache to do, it really is! It wasted me $20 trying to fit darkc0de/BlackXorDolphTronGOAT because it's a float32 dtype! I didn't know I can downcast the model into dfloat16 first until at the very end there. But, this model has the downcast in the df16 directory for your pleasure.

Still testing and awaiting GGUF. Please kindly give your patience. Thank you <3

Philosophy

Mind

Fusion of darkc0de's fusions merge, and Acree's DeepSeek distill abliterated.

I think this is a good base. The Xortron is the top performing at UGI leaderboard and Arcee has the DeepSeek distill and updated world information.

Heart

Karcher with all 2503 base, using updated Dan's and Eurydice, and reintroduce Pantheon.

I feel like this will improve the roleplay. All has the models have unique characteristics of their own, and hopefully the Karcher mean can find a nice center.

Soul

Stock on Sleep's Omega Gaslight and Broken Tutu, since both based on Cydonia 2.1 and have the BlackSheep model in their merge.

I'm horny...

Merge Details

This is a merge of pre-trained language models created using mergekit.

Merge Method

This model was merged using multi-stage model merging method:

  • Mind was merged using the Arcee Fusion merge method using huihui-ai/Arcee-Blitz-abliterated as a base.
  • Heart was merged using the Karcher Mean merge method.
  • Soul was merged using the Model Stock merge method using TheDrummer/Cydonia-24B-v2.1 as a base.

Fially, this model was merged using the DELLA merge method using Mind as a base.

The partial merge results (Mind, Heart, and Soul) are available inside the intermediates directory.

Models Merged

The following models were included in the merge:

  • Mind
    • darkc0de/BlackXorDolphTronGOAT
  • Soul
    • PocketDoc/Dans-PersonalityEngine-V1.3.0-24b
    • aixonlab/Eurydice-24b-v3.5
    • Gryphe/Pantheon-RP-1.8-24b-Small-3.1
  • Heart
    • ReadyArt/Broken-Tutu-24B
    • ReadyArt/Omega-Darker-Gaslight_The-Final-Forgotten-Fever-Dream-24B

Configuration

The following YAML configuration was used to produce this model (using mergekit-multi):

name: Mind
merge_method: arcee_fusion
dtype: bfloat16
tokenizer:
  source: union
chat_template: auto
base_model: huihui-ai/Arcee-Blitz-abliterated
models:
  - model: darkc0de/BlackXorDolphTronGOAT
---
name: Heart
merge_method: karcher
tokenizer:
  source: union
chat_template: auto
parameters:
  max_iter: 1000
models:
  - model: PocketDoc/Dans-PersonalityEngine-V1.3.0-24b
  - model: aixonlab/Eurydice-24b-v3.5
  - model: Gryphe/Pantheon-RP-1.8-24b-Small-3.1
---
name: Soul
merge_method: model_stock
tokenizer:
  source: union
chat_template: auto
base_model: TheDrummer/Cydonia-24B-v2.1
models:
  - model: ReadyArt/Broken-Tutu-24B
  - model: ReadyArt/Omega-Darker-Gaslight_The-Final-Forgotten-Fever-Dream-24B
---
merge_method: della
tokenizer:
  source: union
chat_template: auto
base_model: Mind
models:
  - model: Mind
  - model: Heart
    parameters:
      weight: 0.6
  - model: Soul
    parameters:
      weight: 0.4
parameters:
  density: 0.7
  epsilon: 0.2
  lamda: 1.1