Calme-Ties-78B
Calme-Ties-78B is a 78-billion-parameter model merged using the TIES methodology, based on the Qwen2 architecture. It integrates two sub-base models: calme-3.2-instruct-78B by MaziyarPanahi and CalmeRys-78B-Orpo-v0.1 by dfurman, which serves as the base model. The merging process assigns equal weight and density to both models, with additional parameters enabling normalization and int8 masking. The model operates using the bfloat16 data type.
Model | Model Name | Model Link |
---|---|---|
Base Model | CalmeRys-78B-Orpo-v0.1 | CalmeRys-78B-Orpo-v0.1 |
Model 1 | calme-3.2-instruct-78B | calme-3.2-instruct-78B |
Model 2 | CalmeRys-78B-Orpo-v0.1 | CalmeRys-78B-Orpo-v0.1 |
Merged Models
This is a merge of pre-trained language models created using mergekit.
Merge Method
This model was merged using the TIES merge method using dfurman/CalmeRys-78B-Orpo-v0.1 as a base.
Models Merged
The following models were included in the merge:
Configuration
The following YAML configuration was used to produce this model:
models:
- model: MaziyarPanahi/calme-3.2-instruct-78b
parameters:
weight: 1
density: 1
merge_method: ties
base_model: dfurman/CalmeRys-78B-Orpo-v0.1
parameters:
weight: 1
density: 1
normalize: true
int8_mask: true
dtype: bfloat16
- Downloads last month
- 34
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the model is not deployed on the HF Inference API.
Model tree for prithivMLmods/Calme-Ties-78B
Merge model
this model