Deaquay commited on
Commit
3b21efe
·
verified ·
1 Parent(s): 5f1e54e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -6
README.md CHANGED
@@ -6,27 +6,29 @@ tags:
6
  - llama.cpp
7
  - quantized
8
  - iq4_xs
9
- - Imatrix
 
 
10
  - saturated-labs
11
  - roleplay
12
  - storytelling
13
  - language-model
14
- quantization: IQ4_XS
15
  model_type: llama
16
  language:
17
  - en
18
  ---
19
 
20
- # 🦖 T-Rex-mini — GGUF I1 IQ4_XS (Quantized - Imatrix)
21
 
22
- This is a **quantized GGUF version** of [`saturated-labs/T-Rex-mini`](https://huggingface.co/saturated-labs/T-Rex-mini), converted using [llama.cpp](https://github.com/ggerganov/llama.cpp) and quantized to the `IQ4_XS` format with Imatrix.
23
 
24
  ## 🔧 Quantization Details
25
 
26
  - **Original Model**: `saturated-labs/T-Rex-mini`
27
  - **Format**: GGUF (`.gguf`)
28
- - **Quantization Type**: `IQ4_XS`
29
  - **Tool Used**: [`llama.cpp`](https://github.com/ggerganov/llama.cpp)
30
  - **Command**:
31
  ```bash
32
- ./llama-quantize.exe --imatrix imatrix.dat trex-mini-f16.gguf trex-mini-iq4_xs.gguf iq4_xs
 
6
  - llama.cpp
7
  - quantized
8
  - iq4_xs
9
+ - q5_k_m
10
+ - q6_k
11
+ - imatrix
12
  - saturated-labs
13
  - roleplay
14
  - storytelling
15
  - language-model
16
+ quantization: IQ4_XS, Q5_K_M, Q6_K
17
  model_type: llama
18
  language:
19
  - en
20
  ---
21
 
22
+ # 🦖 T-Rex-mini — GGUF I1 (Quantized - Imatrix)
23
 
24
+ This is a **quantized GGUF version** of [`saturated-labs/T-Rex-mini`](https://huggingface.co/saturated-labs/T-Rex-mini), converted using [llama.cpp](https://github.com/ggerganov/llama.cpp) and quantized with imatrix.
25
 
26
  ## 🔧 Quantization Details
27
 
28
  - **Original Model**: `saturated-labs/T-Rex-mini`
29
  - **Format**: GGUF (`.gguf`)
30
+ - **Quantization Type**: `IQ4_XS, Q5_K_M, Q6_K`
31
  - **Tool Used**: [`llama.cpp`](https://github.com/ggerganov/llama.cpp)
32
  - **Command**:
33
  ```bash
34
+ ./llama-quantize.exe --imatrix imatrix.dat trex-mini-f16.gguf trex-mini-QX_X_X.gguf QX_X_X