These are non-imatrix FP16 here. i1-GGUF here.
The Q2_K is pretty good like 7B models.
Chat template
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
16-bit
Base model