hugging-quants/Meta-Llama-3.1-8B-Instruct-GPTQ-INT4 Text Generation • 2B • Updated Aug 7, 2024 • 13.8k • 39
DavidAU/Maximizing-Model-Performance-All-Quants-Types-And-Full-Precision-by-Samplers_Parameters Updated Jul 27 • 136
DavidAU/L3.1-Dark-Reasoning-LewdPlay-evo-Hermes-R1-Uncensored-8B Text Generation • 8B • Updated Jul 28 • 180 • 24
hugging-quants/Meta-Llama-3.1-405B-Instruct-AWQ-INT4 Text Generation • 59B • Updated Sep 13, 2024 • 781 • 36
hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4 Text Generation • 2B • Updated Aug 7, 2024 • 97.8k • 76
hugging-quants/Meta-Llama-3.1-70B-Instruct-AWQ-INT4 Text Generation • 11B • Updated Aug 7, 2024 • 324k • 105
hugging-quants/Meta-Llama-3.1-405B-Instruct-GPTQ-INT4 Text Generation • 59B • Updated Aug 7, 2024 • 1.62k • 16
hugging-quants/Meta-Llama-3.1-405B-Instruct-BNB-NF4 Text Generation • 214B • Updated Sep 16, 2024 • 12 • 5
hugging-quants/Meta-Llama-3.1-8B-Instruct-BNB-NF4 Text Generation • 5B • Updated Aug 8, 2024 • 955 • 8
ModelCloud/Meta-Llama-3.1-8B-Instruct-gptq-4bit Text Generation • 2B • Updated Jul 29, 2024 • 605 • 4
ModelCloud/Meta-Llama-3.1-70B-Instruct-gptq-4bit Text Generation • 11B • Updated Jul 27, 2024 • 20 • 4
hugging-quants/Meta-Llama-3.1-70B-Instruct-GPTQ-INT4 Text Generation • 11B • Updated Aug 7, 2024 • 3k • 23
sunnyyy/openbuddy-llama3.1-8b-v22.1-131k-Q4_K_M-GGUF Text Generation • 8B • Updated Jul 25, 2024 • 10
azhiboedova/Meta-Llama-3.1-8B-Instruct-AQLM-2Bit-1x16 Text Generation • 2B • Updated Aug 28, 2024 • 2 • 13
hugging-quants/Meta-Llama-3.1-405B-BNB-NF4-BF16 Text Generation • 111B • Updated Sep 16, 2024 • 16 • 2