IQ4_NL is generating gibberish on llama.cpp
#2
by
netroy
- opened
@netroy
I tried on CUDA again via ./llama.cpp/llama-cli -hf unsloth/Qwen3-30B-A3B-Thinking-2507-GGUF:IQ4_NL -ngl 99 --jinja
and it works fine - see screenshot below:
Please try redownloading the model weights as well