Quantized gguf file from https://huggingface.co/huihui-ai/Qwen2.5-VL-7B-Instruct-abliterated

Using "--leave-output-tensor" in quantizing to keep output layer precision at FP16.

LM Studio is recommended to deploy it.

Runtime environment should be upgraded to >= 1.29.0(beta).

imatrix.dat is from mradermacher/Qwen2.5-VL-7B-Instruct-abliterated-i1-GGUF

Downloads last month
1,232
GGUF
Model size
7.62B params
Architecture
qwen2vl
Hardware compatibility
Log In to view the estimation

3-bit

4-bit

5-bit

6-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Misaka27260/Qwen2.5-VL-7B-Instruct-abliterated-GGUF