Mistral-7B-v0.1-4bit-mlx
A 4-bit LoRAโfineโtuned Mistral-7B model in Apple MLX format, created via the MLX Examples LoRA tutorial.
๐ Overview
This model was created by following the MLX LoRA tutorial:
- LoRA fineโtuning applied on Mistralโ7B using lowโrank adapters
- Model fusion performed with
convert.py
to combine base weights and LoRA adapters - Quantized to 4-bit for memory efficiency on Apple Silicon devices
- Packaged in MLX format, ready for downstream use via
mlx-lm
The pipeline is fully automated based on mlx-examples/lora/convert.py
, with no manual weight editing.
๐ฆ Model Packaging
File | Description |
---|---|
weights.npz |
Fused weights: base + LoRA adapters |
config.json |
Model configuration & quantization metadata |
tokenizer.model |
SentencePiece tokenizer for Mistral-7B |
๐ก Usage
Inference with mlx-lm
pip install mlx-lm
- Downloads last month
- 3
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for Hinova/mistral-7B-v0.1-4bit-mlx
Base model
mistralai/Mistral-7B-v0.1