mistral_7b_sft_lora

This is a merged Mistral-7B model fine-tuned with a LoRA adapter for SFT or DPO tasks.

Base model: mistralai/Mistral-7B-v0.1
LoRA adapter: LoRA

Usage

from transformers import AutoTokenizer, AutoModelForCausalLM

model = AutoModelForCausalLM.from_pretrained("meixiang123/mistral_7b_sft_lora")
tokenizer = AutoTokenizer.from_pretrained("meixiang123/mistral_7b_sft_lora")
Downloads last month
78
Safetensors
Model size
7.24B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for meixiang123/mistral_7b_sft_lora

Finetuned
(965)
this model