metadata
base_model: []
tags:
- mergekit
- merge
Psyonic-Cetacean-20B-V2
This is a merge of pre-trained language models created using mergekit.
Merge Details
Merge Method
This model was merged using the linear merge method on two stack-merged models.
The first is jebcarter/psyonic-cetacean-20B (Orca first, reproduced so I didn't have to download that model on top of the components). The second is the same recipe with the models reversed.
Since jebcarter suggested this recipe, credit goes to him.
Models Merged
The following models were included in the merge:
- microsoft/Orca-2-13b
- KoboldAI/LLaMA2-13B-Psyfighter2
Configuration
The following YAML configuration was used to produce this model:
models:
- model: microsoft/Orca-2-13b
parameters:
weight: 1.0
merge_method: task_arithmetic
base_model: TheBloke/Llama-2-13B-fp16
dtype: float16
name: FlatOrca2
---
slices:
- sources:
- model: FlatOrca2
layer_range: [0, 16]
- sources:
- model: KoboldAI/LLaMA2-13B-Psyfighter2
layer_range: [8, 24]
- sources:
- model: FlatOrca2
layer_range: [17, 32]
- sources:
- model: KoboldAI/LLaMA2-13B-Psyfighter2
layer_range: [25, 40]
merge_method: passthrough
dtype: float16
name: Psycet
---
slices:
- sources:
- model: KoboldAI/LLaMA2-13B-Psyfighter2
layer_range: [0, 16]
- sources:
- model: FlatOrca2
layer_range: [8, 24]
- sources:
- model: KoboldAI/LLaMA2-13B-Psyfighter2
layer_range: [17, 32]
- sources:
- model: FlatOrca2
layer_range: [25, 40]
merge_method: passthrough
dtype: float16
name: Psycet-Reverse
---
models:
- model: Psycet
parameters:
weight: 0.5
- model: Psycet-Reverse
parameters:
weight: 0.5
merge_method: linear
dtype: float16