Fentible commited on
Commit
f6a2c69
·
verified ·
1 Parent(s): c473064

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +97 -97
README.md CHANGED
@@ -1,98 +1,98 @@
1
- ---
2
- base_model:
3
- - sam-paech/Darkest-muse-v1
4
- - TheDrummer/Tiger-Gemma-9B-v3
5
- library_name: transformers
6
- tags:
7
- - mergekit
8
- - merge
9
-
10
- ---
11
- # Darkest-Tiger-Gemma-v1-9B-GGUF
12
-
13
- This is a **creative, uncensored** merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
14
-
15
- ![image/png](https://i.imgur.com/Jrk1SdF.jpeg)
16
-
17
- ---
18
-
19
- This model was converted to GGUF format from [`Fentible/Darkest-Tiger-Gemma-v1-9B`](https://huggingface.co/Fentible/Darkest-Tiger-Gemma-v1-9B) using llama.cpp via the Fentible's [GGUF-repo-suite](https://huggingface.co/spaces/fentible/gguf-repo-suite).
20
-
21
- **GGUF Repo Suite** is based on a refactored fork of ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space, updated for offline use with windows and support for lower IQ quants.
22
-
23
- **imatrix.dat** generated using bartowski's [calibration_datav3.txt](https://gist.github.com/bartowski1182/eb213dccb3571f863da82e99418f81e8)
24
-
25
- Refer to the [original model card](https://huggingface.co/Fentible/Darkest-Tiger-Gemma-v1-9B) for more details on the model.
26
-
27
- ---
28
-
29
- ## Merge Details
30
- ### Merge Method
31
-
32
- This model was merged using the [SLERP](https://en.wikipedia.org/wiki/Slerp) merge method.
33
-
34
- ### Models Merged
35
-
36
- The following models were included in the merge:
37
- * [sam-paech/Darkest-muse-v1](https://huggingface.co/sam-paech/Darkest-muse-v1)
38
- * [TheDrummer/Tiger-Gemma-9B-v3](https://huggingface.co/TheDrummer/Tiger-Gemma-9B-v3)
39
-
40
- ### Configuration
41
-
42
- The following YAML configuration was used to produce this model:
43
-
44
- ```yaml
45
- base_model: TheDrummer/Tiger-Gemma-9B-v3
46
- merge_method: slerp
47
- dtype: bfloat16
48
- slices:
49
- - sources:
50
- - model: sam-paech/Darkest-muse-v1
51
- layer_range: [0, 42]
52
- - model: TheDrummer/Tiger-Gemma-9B-v3
53
- layer_range: [0, 42]
54
- parameters:
55
- t: 0.5
56
- ```
57
-
58
- ---
59
-
60
- ## Use with llama.cpp
61
- Install llama.cpp through brew (works on Mac and Linux)
62
-
63
- ```bash
64
- brew install llama.cpp
65
-
66
- ```
67
- Invoke the llama.cpp server or the CLI.
68
-
69
- ### CLI:
70
- ```bash
71
- llama-cli --hf-repo Fentible/gemma-3-4b-it-abliterated-IQ1_S-GGUF --hf-file gemma-3-4b-it-abliterated-IQ1_S-imat.gguf -p "The meaning to life and the universe is"
72
- ```
73
-
74
- ### Server:
75
- ```bash
76
- llama-server --hf-repo Fentible/gemma-3-4b-it-abliterated-IQ1_S-GGUF --hf-file gemma-3-4b-it-abliterated-IQ1_S-imat.gguf -c 2048
77
- ```
78
-
79
- Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
80
-
81
- Step 1: Clone llama.cpp from GitHub.
82
- ```
83
- git clone https://github.com/ggerganov/llama.cpp
84
- ```
85
-
86
- Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
87
- ```
88
- cd llama.cpp && LLAMA_CURL=1 make
89
- ```
90
-
91
- Step 3: Run inference through the main binary.
92
- ```
93
- ./llama-cli --hf-repo Fentible/gemma-3-4b-it-abliterated-IQ1_S-GGUF --hf-file gemma-3-4b-it-abliterated-IQ1_S-imat.gguf -p "The meaning to life and the universe is"
94
- ```
95
- or
96
- ```
97
- ./llama-server --hf-repo Fentible/gemma-3-4b-it-abliterated-IQ1_S-GGUF --hf-file gemma-3-4b-it-abliterated-IQ1_S-imat.gguf -c 2048
98
  ```
 
1
+ ---
2
+ base_model:
3
+ - sam-paech/Darkest-muse-v1
4
+ - TheDrummer/Tiger-Gemma-9B-v3
5
+ library_name: transformers
6
+ tags:
7
+ - mergekit
8
+ - merge
9
+
10
+ ---
11
+ # Darkest-Tiger-Gemma-9B-v1-GGUF
12
+
13
+ This is a **creative, uncensored** merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
14
+
15
+ ![image/png](https://i.imgur.com/Jrk1SdF.jpeg)
16
+
17
+ ---
18
+
19
+ This model was converted to GGUF format from [`Fentible/Darkest-Tiger-Gemma-9B-v1`](https://huggingface.co/Fentible/Darkest-Tiger-Gemma-9B-v1) using llama.cpp via the Fentible's [GGUF-repo-suite](https://huggingface.co/spaces/fentible/gguf-repo-suite).
20
+
21
+ **GGUF Repo Suite** is based on a refactored fork of ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space, updated for offline use with windows and support for lower IQ quants.
22
+
23
+ **imatrix.dat** generated using bartowski's [calibration_datav3.txt](https://gist.github.com/bartowski1182/eb213dccb3571f863da82e99418f81e8)
24
+
25
+ Refer to the [original model card](https://huggingface.co/Fentible/Darkest-Tiger-Gemma-9B-v1) for more details on the model.
26
+
27
+ ---
28
+
29
+ ## Merge Details
30
+ ### Merge Method
31
+
32
+ This model was merged using the [SLERP](https://en.wikipedia.org/wiki/Slerp) merge method.
33
+
34
+ ### Models Merged
35
+
36
+ The following models were included in the merge:
37
+ * [sam-paech/Darkest-muse-v1](https://huggingface.co/sam-paech/Darkest-muse-v1)
38
+ * [TheDrummer/Tiger-Gemma-9B-v3](https://huggingface.co/TheDrummer/Tiger-Gemma-9B-v3)
39
+
40
+ ### Configuration
41
+
42
+ The following YAML configuration was used to produce this model:
43
+
44
+ ```yaml
45
+ base_model: TheDrummer/Tiger-Gemma-9B-v3
46
+ merge_method: slerp
47
+ dtype: bfloat16
48
+ slices:
49
+ - sources:
50
+ - model: sam-paech/Darkest-muse-v1
51
+ layer_range: [0, 42]
52
+ - model: TheDrummer/Tiger-Gemma-9B-v3
53
+ layer_range: [0, 42]
54
+ parameters:
55
+ t: 0.5
56
+ ```
57
+
58
+ ---
59
+
60
+ ## Use with llama.cpp
61
+ Install llama.cpp through brew (works on Mac and Linux)
62
+
63
+ ```bash
64
+ brew install llama.cpp
65
+
66
+ ```
67
+ Invoke the llama.cpp server or the CLI.
68
+
69
+ ### CLI:
70
+ ```bash
71
+ llama-cli --hf-repo Fentible/Darkest-Tiger-Gemma-9B-v1 --hf-file Darkest-Tiger-Gemma-9B-v1-IQ4_XS.gguf -p "The meaning to life and the universe is"
72
+ ```
73
+
74
+ ### Server:
75
+ ```bash
76
+ llama-server --hf-repo Fentible/Darkest-Tiger-Gemma-9B-v1 --hf-file Darkest-Tiger-Gemma-9B-v1-IQ4_XS.gguf -c 2048
77
+ ```
78
+
79
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
80
+
81
+ Step 1: Clone llama.cpp from GitHub.
82
+ ```
83
+ git clone https://github.com/ggerganov/llama.cpp
84
+ ```
85
+
86
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
87
+ ```
88
+ cd llama.cpp && LLAMA_CURL=1 make
89
+ ```
90
+
91
+ Step 3: Run inference through the main binary.
92
+ ```
93
+ ./llama-cli --hf-repo Fentible/Darkest-Tiger-Gemma-9B-v1 --hf-file Darkest-Tiger-Gemma-9B-v1-IQ4_XS.gguf -p "The meaning to life and the universe is"
94
+ ```
95
+ or
96
+ ```
97
+ ./llama-server --hf-repo Fentible/Darkest-Tiger-Gemma-9B-v1 --hf-file Darkest-Tiger-Gemma-9B-v1-IQ4_XS.gguf -c 2048
98
  ```