Scripturient is a culmination of my ongoing experiments with merging specialized curated models. Designed to keep creativity high, without sacrificing stability.
As for samplers, the model doesn't need samplers to reign it in much at all. My recommendation is:
Temp: 1
Min P: 0.01
That being said, it can handle even higher temperatures and Nsigma works well too.
Because of the nature of this sort of 'Hyper Multi Model Merge', my recommendation is not to run this on anything lower than a Q5 quant.
If you enjoy my work, please consider supporting me, It helps me make more models like this! Support on KO-FI <3
I want to say a special thank you to everyone at the BeaverAI community who supports me, be that with testing, feedback, advice or donations! Special shoutouts to (forgive me if I left someone out!): @Artus | @Geechan | @Kromeurus | @NarpasSword | @Thana Alt | @FrenzyBiscuit | @Saintonan | @Lightning_missile | @Inasity | @Amp | @madison ๐ฆ @ IQ3_XS | @zerofata
Configuration
The following YAML configuration was used to produce this model:
models:
- model: TareksLab/Diamond-DL-V1-LLaMa-70B
parameters:
weight: 0.10
density: 0.7
epsilon: 0.20
- model: TareksLab/Citrine-MS-V3-LLaMa-70B
parameters:
weight: [0.5, 0.2, 0.1, 0.1, 0.1]
density: 0.7
epsilon: 0.20
- model: TareksLab/Amethyst-SCE-V4-LLaMa-70B
parameters:
weight: [0.2, 0.4, 0.2, 0.1, 0.1]
density: 0.7
epsilon: 0.20
- model: TareksLab/Ruby-D-V3-LLaMa-70B
parameters:
weight: [0.1, 0.2, 0.4, 0.2, 0.1]
density: 0.7
epsilon: 0.20
- model: TareksLab/Carnelian-SCE-V4-LLaMa-70B
parameters:
weight: [0.1, 0.1, 0.2, 0.4, 0.2]
density: 0.7
epsilon: 0.20
- model: TareksLab/Emerald-SCE-V3-LLaMa-70B
parameters:
weight: [0.1, 0.1, 0.1, 0.2, 0.5]
density: 0.7
epsilon: 0.20
merge_method: della_linear
base_model: TareksLab/Diamond-DL-V1-LLaMa-70B
parameters:
lambda: 1.1
normalize: false
dtype: float32
out_dtype: bfloat16
chat_template: llama3
tokenizer:
source: TareksLab/Ruby-D-V3-LLaMa-70B
pad_to_multiple_of: 8
- Downloads last month
- 68
Model tree for narpas/Scripturient-V1.3-LLaMa-70B-6.0bpw-h8-exl2
Base model
Tarek07/Scripturient-V1.3-LLaMa-70B