GGUF
Collection
DASLab support for GGUF
โข
2 items
โข
Updated
This repository contains advanced quantized versions of Llama 3.1 8B Instruct using GPTQ quantization and GPTQ+EvoPress optimization from the DASLab GGUF Toolkit.
Our GPTQ-based quantization methods achieve superior quality-compression tradeoffs compared to standard quantization:
Compatible with llama.cpp and all GGUF-supporting inference engines. No special setup required.
Full documentation, evaluation results, and toolkit source: https://github.com/IST-DASLab/gguf-toolkit
2-bit
6-bit
Base model
meta-llama/Llama-3.1-8B