Anthonyg5005 commited on
Commit
51656b0
·
verified ·
1 Parent(s): a77d6da

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +31 -0
README.md CHANGED
@@ -15,3 +15,34 @@ tags:
15
  base_model:
16
  - Undi95/ReMM-v2.2-L2-13B
17
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  base_model:
16
  - Undi95/ReMM-v2.2-L2-13B
17
  ---
18
+
19
+ # exl2 quants for ReMM V2.2
20
+
21
+ This repository includes the quantized models for the [ReMM V2.2](https://huggingface.co/Undi95/ReMM-v2.2-L2-13B) model by [Undi](https://huggingface.co/Undi95). ReMM is a model merge attempting to recreate [MythoMax](https://huggingface.co/Gryphe/MythoMax-L2-13b) using the [SLERP](https://github.com/Undi95/LLM-SLERP-MergeTest) merging method and newer models.
22
+
23
+ ## Current models
24
+
25
+ | exl2 Quant | Model Branch | Model Size | Minimum VRAM (4096 Context) | BPW |
26
+ |-|-|-|-|-|
27
+ | 3-Bit | main | N/A | N/A | 3.72 |
28
+ | 4-Bit | 4bit | N/A | N/A | N/A |
29
+ | 5-Bit | [Orang Baik's Repo](https://huggingface.co/R136a1/ReMM-v2.2-L2-13B-exl2) | 8.96 GB | 16GB GPU | 5.33 |
30
+
31
+ ### Note
32
+
33
+ TODO or delete
34
+
35
+ ## Where to use
36
+
37
+ There are a couple places you can use an exl2 model, here are a few:
38
+
39
+ - [oobabooga's Text Gen Webui](https://github.com/oobabooga/text-generation-webui)
40
+ - When using the downloader, make sure to format like this: Anthonyg5005/ReMM-v2.2-L2-13B-4bit-exl2**\:QuantBranch**
41
+ - With 5-Bit download: [R136a1/ReMM-v2.2-L2-13B-exl2](https://huggingface.co/R136a1/ReMM-v2.2-L2-13B-exl2)
42
+ - [tabbyAPI](https://github.com/theroyallab/tabbyAPI)
43
+ - [ExUI](https://github.com/turboderp/exui)
44
+ - [KoboldAI](https://github.com/henk717/KoboldAI) (Clone repo, don't use snapshot)
45
+
46
+ ## WARNING
47
+
48
+ Model cannot be used commercially due to the Alpaca dataset license. Only use this model for research purposes or personal use.