This model was pulled directly from the ollama model library and then renamed as a GGUF file. This is useful for solutions that require offline inference for Ollama and can not pull directly from the Ollama Library. Specifically, this GGUF will work with Ollama versions (0.11.4) and above. I have found that other GGUF providers for GPT-OSS are still not 100% with Ollama's implementation of the model. This is most likely not observed with lama.cpp, but just wanted an easy way to consume this with the Tanzu AI solutions stack.

Downloads last month
38
GGUF
Model size
20.9B params
Architecture
gptoss
Hardware compatibility
Log In to view the estimation

We're not able to determine the quantization variants.

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for tehkuhnz/gpt-oss-20b

Base model

openai/gpt-oss-20b
Quantized
(112)
this model