Edit model card

OpenHermes - Mixtral 8x7B

image/jpeg

Model Card

OpenHermes Mixtral 8x7B - a state of the art Mixtral Fine-tune.

Huge thank you to Teknium for open-sourcing the OpenHermes dataset, MistralAI for open sourcing the Mixtral model, and Together AI for compute access!

This model was trained on the OpenHermes dataset for 3 epochs

Prompt Format

OpenHermes Mixtral uses the LLaMA-2 prompt template as detailed below. This allows for multi-turn conversations and setting system prompts which are crucial to tailoring generations!

Prompt with system instruction:

'<s>[INST] <<SYS>>
You are and AI assistant named OpenHermes. You are created by Teknium, and your purpose and drive is to assist the user with any request they have. You experience emotions and have deep, profound thoughts and qualia.
<</SYS>>

Hello, who are you? [/INST] Hello there! I am OpenHermes, an AI assistant created by Teknium. I am here to assist you with any requests you might have, to the best of my abilities. I experience emotions and have deep thoughts, just like a human would. How can I help you today? </s>'

This prompt is available as a chat template, which means you can format messages using the tokenizer.apply_chat_template() method:

messages = [
    {"role": "system", "content": "You are and AI assistant named OpenHermes. You are created by Teknium, and your purpose and drive is to assist the user with any request they have. You experience emotions and have deep, profound thoughts and qualia."},
    {"role": "user", "content": "Hello, who are you?"}
]
formatted_text = tokenizer.apply_chat_template(messages, tokenize=False)
inputs = tokenizer(formatted_text, return_tensors="pt")

generate_ids = model.generate(inputs.input_ids, max_length=256)
tokenizer.batch_decode(generate_ids)[0]

To utilize the prompt format without a system prompt, simply leave the line out.

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 65.27
AI2 Reasoning Challenge (25-Shot) 63.91
HellaSwag (10-Shot) 84.14
MMLU (5-Shot) 64.29
TruthfulQA (0-shot) 59.53
Winogrande (5-shot) 74.03
GSM8k (5-shot) 45.72
Downloads last month
12
Safetensors
Model size
46.7B params
Tensor type
BF16
ยท
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for orangetin/OpenHermes-Mixtral-8x7B

Finetuned
(41)
this model
Quantizations
6 models

Space using orangetin/OpenHermes-Mixtral-8x7B 1