huggingkot
commited on
Commit
·
66adcf5
1
Parent(s):
24b6561
add files
Browse files- .gitattributes +1 -0
- README.md +17 -0
.gitattributes
CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
*.json filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
---
|
3 |
+
base_model:
|
4 |
+
- NeverSleep/Lumimaid-v0.2-8B
|
5 |
+
---
|
6 |
+
|
7 |
+
This is a converted weight from [Lumimaid-v0.2-8B](https://huggingface.co/NeverSleep/Lumimaid-v0.2-8B) model in [unsloth 4-bit dynamic quant](https://archive.is/EFz7P) using this [collab notebook](https://colab.research.google.com/drive/1P23C66j3ga49kBRnDNlmRce7R_l_-L5l?usp=sharing).
|
8 |
+
|
9 |
+
## About this Conversion
|
10 |
+
|
11 |
+
This conversion uses **Unsloth** to load the model in **4-bit** format and force-save it in the same **4-bit** format.
|
12 |
+
|
13 |
+
### How 4-bit Quantization Works
|
14 |
+
- The actual **4-bit quantization** is handled by **BitsAndBytes (bnb)**, which works under **Torch** via **AutoGPTQ** or **BitsAndBytes**.
|
15 |
+
- **Unsloth** acts as a wrapper, simplifying and optimizing the process for better efficiency.
|
16 |
+
|
17 |
+
This allows for reduced memory usage and faster inference while keeping the model compact.
|