This model is the quantized version of NexusFlow's NexusRaven V-2. The quantization technique used is Activation-Aware Weight Quantization. The model is suitable for high-degree of Function Calling. The functions may be Simple Functions, Compound Functions or Nested Functions. The model hasn't been fine-tuned yet.

Model creator: Nexusflow

Original model: NexusRaven V2 13B

Downloads last month
5
Safetensors
Model size
2.03B params
Tensor type
I32
·
FP16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Dataset used to train NaiveAttention/NexusRaven-V2-13B-awq