huggingkot commited on
Commit
66adcf5
·
1 Parent(s): 24b6561
Files changed (2) hide show
  1. .gitattributes +1 -0
  2. README.md +17 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ *.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ base_model:
4
+ - NeverSleep/Lumimaid-v0.2-8B
5
+ ---
6
+
7
+ This is a converted weight from [Lumimaid-v0.2-8B](https://huggingface.co/NeverSleep/Lumimaid-v0.2-8B) model in [unsloth 4-bit dynamic quant](https://archive.is/EFz7P) using this [collab notebook](https://colab.research.google.com/drive/1P23C66j3ga49kBRnDNlmRce7R_l_-L5l?usp=sharing).
8
+
9
+ ## About this Conversion
10
+
11
+ This conversion uses **Unsloth** to load the model in **4-bit** format and force-save it in the same **4-bit** format.
12
+
13
+ ### How 4-bit Quantization Works
14
+ - The actual **4-bit quantization** is handled by **BitsAndBytes (bnb)**, which works under **Torch** via **AutoGPTQ** or **BitsAndBytes**.
15
+ - **Unsloth** acts as a wrapper, simplifying and optimizing the process for better efficiency.
16
+
17
+ This allows for reduced memory usage and faster inference while keeping the model compact.