dgomes03 commited on
Commit
41f890f
·
verified ·
1 Parent(s): 703b5ee

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -10,3 +10,5 @@ tags:
10
  - mlx
11
  pipeline_tag: text-generation
12
  ---
 
 
 
10
  - mlx
11
  pipeline_tag: text-generation
12
  ---
13
+ Mistral-7B-Instruct-v0.3 quantized with mixed precision:
14
+ This is a Mistral-7B-Instruct model where the embedding layer and output (head) layer are quantized to 6-bit precision, while the rest of the model uses 4-bit quantization. This mixed-precision approach aims to balance model size and inference speed with improved representational fidelity in critical layers.