This repository contains a quantized versions of https://huggingface.co/lapa-llm/lapa-v0.1.2-instruct

There are the following quants:

  • Q4_K_S
  • Q4_K_M
  • Q5_K_S
  • Q8_0
Downloads last month
773
GGUF
Model size
12B params
Architecture
gemma3
Hardware compatibility
Log In to view the estimation

4-bit

5-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Collection including lapa-llm/lapa-v0.1.2-instruct-GGUF