Mistral-7B-v0.1-4bit-mlx

A 4-bit LoRAโ€‘fineโ€‘tuned Mistral-7B model in Apple MLX format, created via the MLX Examples LoRA tutorial.


๐Ÿš€ Overview

This model was created by following the MLX LoRA tutorial:

  • LoRA fineโ€‘tuning applied on Mistralโ€‘7B using lowโ€‘rank adapters
  • Model fusion performed with convert.py to combine base weights and LoRA adapters
  • Quantized to 4-bit for memory efficiency on Apple Silicon devices
  • Packaged in MLX format, ready for downstream use via mlx-lm

The pipeline is fully automated based on mlx-examples/lora/convert.py, with no manual weight editing.


๐Ÿ“ฆ Model Packaging

File Description
weights.npz Fused weights: base + LoRA adapters
config.json Model configuration & quantization metadata
tokenizer.model SentencePiece tokenizer for Mistral-7B

๐Ÿ’ก Usage

Inference with mlx-lm

pip install mlx-lm
Downloads last month
3
Safetensors
Model size
1.13B params
Tensor type
F16
ยท
U32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Hinova/mistral-7B-v0.1-4bit-mlx

Finetuned
(965)
this model