Oganesson-TinyLlama-1.2B-GGUF

Oganesson-TinyLlama-1.2B is a lightweight and efficient language model built on the LLaMA 3.2 1.2B architecture. Fine-tuned for general-purpose inference, mathematical reasoning, and code generation, it’s ideal for edge devices, personal assistants, and educational applications requiring a compact yet capable model.

Model File

File Name Size Format
Oganesson-TinyLlama-1.2B.BF16.gguf 2.48 GB BF16
Oganesson-TinyLlama-1.2B.F16.gguf 2.48 GB F16
Oganesson-TinyLlama-1.2B.F32.gguf 4.95 GB F32
Oganesson-TinyLlama-1.2B.Q4_K_M.gguf 808 MB Q4_K_M
.gitattributes 1.8 kB -
README.md 212 B -
config.json 31 B JSON

Quants Usage

(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)

Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

image.png

Downloads last month
67
GGUF
Model size
1.24B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

4-bit

16-bit

32-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for prithivMLmods/Oganesson-TinyLlama-1.2B-GGUF

Quantized
(1)
this model

Collection including prithivMLmods/Oganesson-TinyLlama-1.2B-GGUF