Llama-3.2-Kapusta-JapanChibi-3B-v1

ใ‚„ใ‚ใฆใใ ใ•ใ„ใ€็งใฏๅฐใ•ใใฆๅฝนใซ็ซ‹ใกใพใ™

I love this model, but I don't understand Japanese, although it is also good in other languages.

Kapusta-JapanChibi-Logo256.png

This is an interesting merge of 3 cool models, created using mergekit. Enjoy exploring :)

Merge Details

Method

This model was merged using the model_stock method.

Models

The following models were included in the merge:

Configuration

The following YAML configurations was used to produce this model:

# Llama-3.2-Kapusta-JapanChibi-3B-v1
models:
  - model: AELLM/Llama-3.2-Chibi-3B
  - model: AXCXEPT/EZO-Llama-3.2-3B-Instruct-dpoE
merge_method: model_stock
base_model: Khetterman/Llama-3.2-Kapusta-3B-v8
dtype: bfloat16

My thanks to the authors of the original models, your work is incredible. Have a good time ๐Ÿ–ค

Downloads last month
3
Safetensors
Model size
3.61B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Khetterman/Llama-3.2-Kapusta-JapanChibi-3B-v1