base_model: mistralai/Mistral-7B-v0.1 | |
models: | |
- model: EstherXC/mixtral_7b_protein_pretrain | |
parameters: | |
weight: 0.3 | |
- model: wanglab/mixtral_7b_dna_pretrain #dnagpt/llama-dna | |
parameters: | |
weight: 0.3 | |
merge_method: task_arithmetic | |
dtype: float16 | |
tokenizer_source: "base" |