Qwen3-8B-CK-Pro-f32-GGUF

The CognitiveKernel/Qwen3-8B-CK-Pro model is a fine-tuned variant of the Qwen3-8B base language model, trained using self-collected trajectories from queries as detailed in the Cognitive Kernel-Pro research. It is designed as a deep research agent and foundation model, achieving strong performance with Pass@1/3 scores of 32.7%/38.2% on the full GAIA dev set and 40.3%/49.3% on the text-only subset. This model builds upon the strengths of Qwen3-8B, which supports advanced reasoning, instruction-following, and multilingual capabilities, specifically optimized for research agent tasks through the Cognitive Kernel-Pro framework. It is not currently deployed by any inference provider on Hugging Face. The model leverages the underlying Qwen3-8B base and its finetuned versions to deliver enhanced agent capabilities for complex question-answering and information synthesis scenarios.

Execute using Ollama

run ->

ollama run hf.co/prithivMLmods/Qwen3-8B-CK-Pro-f32-GGUF:Q2_K

Model Files

File Name Quant Type File Size
Qwen3-8B-CK-Pro.BF16.gguf BF16 16.4 GB
Qwen3-8B-CK-Pro.F16.gguf F16 16.4 GB
Qwen3-8B-CK-Pro.F32.gguf F32 32.8 GB
Qwen3-8B-CK-Pro.Q2_K.gguf Q2_K 3.28 GB

Quants Usage

(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)

Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

image.png

Downloads last month
2,067
GGUF
Model size
8.19B params
Architecture
qwen3
Hardware compatibility
Log In to view the estimation

2-bit

16-bit

32-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for prithivMLmods/Qwen3-8B-CK-Pro-f32-GGUF

Base model

Qwen/Qwen3-8B-Base
Finetuned
Qwen/Qwen3-8B
Quantized
(1)
this model