Update README.md
Browse files
README.md
CHANGED
@@ -15,3 +15,34 @@ tags:
|
|
15 |
base_model:
|
16 |
- Undi95/ReMM-v2.2-L2-13B
|
17 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
base_model:
|
16 |
- Undi95/ReMM-v2.2-L2-13B
|
17 |
---
|
18 |
+
|
19 |
+
# exl2 quants for ReMM V2.2
|
20 |
+
|
21 |
+
This repository includes the quantized models for the [ReMM V2.2](https://huggingface.co/Undi95/ReMM-v2.2-L2-13B) model by [Undi](https://huggingface.co/Undi95). ReMM is a model merge attempting to recreate [MythoMax](https://huggingface.co/Gryphe/MythoMax-L2-13b) using the [SLERP](https://github.com/Undi95/LLM-SLERP-MergeTest) merging method and newer models.
|
22 |
+
|
23 |
+
## Current models
|
24 |
+
|
25 |
+
| exl2 Quant | Model Branch | Model Size | Minimum VRAM (4096 Context) | BPW |
|
26 |
+
|-|-|-|-|-|
|
27 |
+
| 3-Bit | main | N/A | N/A | 3.72 |
|
28 |
+
| 4-Bit | 4bit | N/A | N/A | N/A |
|
29 |
+
| 5-Bit | [Orang Baik's Repo](https://huggingface.co/R136a1/ReMM-v2.2-L2-13B-exl2) | 8.96 GB | 16GB GPU | 5.33 |
|
30 |
+
|
31 |
+
### Note
|
32 |
+
|
33 |
+
TODO or delete
|
34 |
+
|
35 |
+
## Where to use
|
36 |
+
|
37 |
+
There are a couple places you can use an exl2 model, here are a few:
|
38 |
+
|
39 |
+
- [oobabooga's Text Gen Webui](https://github.com/oobabooga/text-generation-webui)
|
40 |
+
- When using the downloader, make sure to format like this: Anthonyg5005/ReMM-v2.2-L2-13B-4bit-exl2**\:QuantBranch**
|
41 |
+
- With 5-Bit download: [R136a1/ReMM-v2.2-L2-13B-exl2](https://huggingface.co/R136a1/ReMM-v2.2-L2-13B-exl2)
|
42 |
+
- [tabbyAPI](https://github.com/theroyallab/tabbyAPI)
|
43 |
+
- [ExUI](https://github.com/turboderp/exui)
|
44 |
+
- [KoboldAI](https://github.com/henk717/KoboldAI) (Clone repo, don't use snapshot)
|
45 |
+
|
46 |
+
## WARNING
|
47 |
+
|
48 |
+
Model cannot be used commercially due to the Alpaca dataset license. Only use this model for research purposes or personal use.
|