The models are converted from Qwen/Qwen3-30B-A3B-Instruct-2507

Before using these models, please set up the generation config properly.

  • temperature = 0.7
  • top_p = 0.8
  • top_k = 20
  • min_p = 0.0
  • output token length: 16,384 tokens

Best Practice: https://huggingface.co/Qwen/Qwen3-30B-A3B-Instruct-2507#best-practices

Downloads last month
1,260
GGUF
Model size
30.5B params
Architecture
qwen3moe
Hardware compatibility
Log In to view the estimation

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support