--- license: other language: - en - fr - es - hi - zh - code base_model: microsoft/Orca-2-13b datasets: - HuggingFaceH4/no_robots - mlabonne/guanaco-llama2-1k - OpenAssistant/oasst_top1_2023-08-25 - totally-not-an-llm/EverythingLM-data-V3 --- The "microsoft/Orca-2-13b" model fully fine-tuned on HuggingFaceH4/no_robots, totally-not-an-llm/EverythingLM-data-V3, mlabonne/guanaco-llama2-1k, and OpenAssistant/oasst_top1_2023-08-25. This model achieved a test loss of 0.18. Make sure to comply with the microsoft research license. Please read it before using this model. This model was trained on the following chat template: "<|USER|> message <|ASSISTANT|> message"