Text Generation
Transformers
Safetensors
English
mistral
text-generation-inference

Running in ollama

#1
by whsinth - opened

Hi, I'm trying to get this model to run in Ollama. It seems that it is the only model out there that is training on public domain code.

I try putting it in llama.cpp's convert_hf_to_gguf.py and I get "NotImplementedError: BPE pre-tokenizer was not recognized - update get_vocab_base_pre()". The BPE pre-tokenizer "48b3395982eda712c0f3aaa4056c791b9a2db4ee474467631e77bb46dab2b235" is also not recognized.

Would it be possible to ship the model in gguf format?

Sign up or log in to comment