Description

This is a 4-bit GPTQ quantized model of failspy/Meta-Llama-3-8B-Instruct-abliterated-v3 It is quantized with wikitext2 dataset. Its file size is 5.73 GB, and can fit into a 8GB VRAM GPU.

Downloads last month
6
Safetensors
Model size
1.99B params
Tensor type
F16
·
I32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support