about

V3.0 changes (relatively minor update) :

  • DarkHorse replace DoppelGangerR1 to add a bit of Negative Llama at the expense of a bit of Fallen Llama R1.
  • A bit of Fallen Llama is recovered by using Smartricks instead of Smartracks as a base.
  • Priestess is upgraded with Lumitron Lorablated.
  • Tess is merged with Hitachi FLDx2 in the perplexity dropper model.

Electra R1 and GutenbergDoppel are kept as they were.

If you have V2 already, this model is quite similar, and the difference might not be worth a download.


merge

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the Model Stock merge method using Nexesenex/Llama_3.x_70b_SmarTricks_V1.01 as a base.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

merge_method: model_stock
models:
  - model: Nexesenex/Llama_3.1_70b_FLDx2-Tess3_abliterated_fusion_norm
    parameters:
      weight: 1.0
  - model: nbeerbower/Llama3.1-Gutenberg-Doppel-70B
    parameters:
      weight: 1.0
  - model: Nexesenex/Llama_3.1_70b_HighPriestess_R1_V1
    parameters:
      weight: 1.0
  - model: Steelskull/L3.3-Electra-R1-70b
    parameters:
      weight: 1.0
  - model: Nexesenex/Llama_3.3_70b_DarkHorse
    parameters:
      weight: 1.0
base_model: Nexesenex/Llama_3.x_70b_SmarTricks_V1.01
dtype: bfloat16
out_dtype: bfloat16
parameters:
  int8_mask: true
  normalize: true
  rescale: false
chat_template: auto
tokenizer:
  source: union
Downloads last month
3
Safetensors
Model size
70.6B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Nexesenex/Llama_3.x_70b_Hexagon_Purple_V3