This model was pulled directly from the ollama model library and then renamed as a GGUF file. This is useful for solutions that require offline inference for Ollama and can not pull directly from the Ollama Library. Specifically, this GGUF will work with Ollama versions (0.11.4) and above. I have found that other GGUF providers for GPT-OSS are still not 100% with Ollama's implementation of the model. This is most likely not observed with lama.cpp, but just wanted an easy way to consume this with the Tanzu AI solutions stack.
- Downloads last month
- 38
Hardware compatibility
Log In
to view the estimation
We're not able to determine the quantization variants.
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for tehkuhnz/gpt-oss-20b
Base model
openai/gpt-oss-20b