Model Details

This is meta-llama/Meta-Llama-3-8B quantized and serialized with AutoAWQ in 4-bit.

Details here:

Fine-tune Llama 3 on Your Computer

Downloads last month
7
Safetensors
Model size
1.98B params
Tensor type
FP16
·
I32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support