Flerovium-Llama-3B-GGUF

Flerovium-Llama-3B is a compact, general-purpose language model based on the powerful llama 3.2 (llama) architecture. It is fine-tuned for a broad range of tasks including mathematical reasoning, code generation, and natural language understanding, making it a versatile choice for developers, students, and researchers seeking reliable performance in a lightweight model.

Model File

File Name Size Format
Flerovium-Llama-3B.BF16.gguf 6.43 GB BF16
Flerovium-Llama-3B.F16.gguf 6.43 GB F16
Flerovium-Llama-3B.Q4_K_M.gguf 2.02 GB Q4_K_M
Flerovium-Llama-3B.Q5_K_M.gguf 2.32 GB Q5_K_M
.gitattributes 1.78 kB -
README.md 927 B -
config.json 31 B JSON

Quants Usage

(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)

Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

image.png

Downloads last month
10
GGUF
Model size
3.21B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

4-bit

5-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for prithivMLmods/Flerovium-Llama-3B-GGUF

Quantized
(1)
this model

Collection including prithivMLmods/Flerovium-Llama-3B-GGUF