Transformers
GGUF
code

Q4 quants

#1
by DrRos - opened

Hi, guys, can you, please also upload a Q4 quants? It seems model cannot be converted to gguf using latest llama.cpp (I'm getting NotImplementedError: BPE pre-tokenizer was not recognized - update get_vocab_base_pre() error while running ./convert_hf_to_gguf.py --outfile ../mellum-non-quant.gguf --verbose ../Mellum-4b-sft-python/

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment