CorticalStack/mistral-7b-openhermes-gptq

CorticalStack/mistral-7b-openhermes-gptq is an GPTQ quantised version of CorticalStack/mistral-7b-openhermes-sft.

GPTQ models are currently supported on Linux (NVidia/AMD) and Windows (NVidia only). MacOS users: please use GGUF models.

These GPTQ models are known to work in the following inference servers/webuis.

Downloads last month
19
Safetensors
Model size
1.2B params
Tensor type
I32
·
F16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support