This model is the original Mistral-ai 7B v0.1 model converted to the OpenNMT-py format. By original, it means with interleaved rotary (option: rotary_interleave=True)

You need to install OpenNMT-py, instructions are here: https://github.com/OpenNMT/OpenNMT-py

Running inference:

Create a text input file with prompts (ex: "Show me some attractions in Boston")
then run:
onmt_translate --config mistral-inference.yaml --src input.txt --output output.txt

Running MMLU evaluation:

If you git clone the OpenNMT-py repo then you can run:
python eval_llm/MMLU/run_mmlu_opennmt.py --config mistral-inference.yaml
For this use case make sure you use max_length=1 in the config file

Finetuning:

Read this tuto: https://forum.opennmt.net/t/finetuning-llama-7b-13b-or-mosaicml-mpt-7b-reproduce-vicuna-alpaca/5272/56
onmt_train --config mistral-finetuning.yaml
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model's library.