EXL2 quants of Mistral-7B-instruct-v0.3
v0.3's vocabulary is compatible with Mistral-Large-123B, so this works as a draft model for Mistral-Large.
2.80 bits per weight
3.00 bits per weight
4.00 bits per weight
4.50 bits per weight
5.00 bits per weight
6.00 bits per weight
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support