morriszms commited on
Commit
37ba07e
·
verified ·
1 Parent(s): 3564b9d

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,15 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ OpenHermes-2.5-neural-chat-v3-3-Slerp-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ OpenHermes-2.5-neural-chat-v3-3-Slerp-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ OpenHermes-2.5-neural-chat-v3-3-Slerp-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ OpenHermes-2.5-neural-chat-v3-3-Slerp-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ OpenHermes-2.5-neural-chat-v3-3-Slerp-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ OpenHermes-2.5-neural-chat-v3-3-Slerp-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ OpenHermes-2.5-neural-chat-v3-3-Slerp-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ OpenHermes-2.5-neural-chat-v3-3-Slerp-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ OpenHermes-2.5-neural-chat-v3-3-Slerp-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ OpenHermes-2.5-neural-chat-v3-3-Slerp-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ OpenHermes-2.5-neural-chat-v3-3-Slerp-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
47
+ OpenHermes-2.5-neural-chat-v3-3-Slerp-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
OpenHermes-2.5-neural-chat-v3-3-Slerp-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a9789622a0610a2aaa56a2bb063d6a11a4a0e61a92fc5be8d48fbddf29cdce9
3
+ size 2719242816
OpenHermes-2.5-neural-chat-v3-3-Slerp-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20598b21c12d1f19c8ff362daeb8f07717997a826ad14a7dc7278941a700e102
3
+ size 3822025280
OpenHermes-2.5-neural-chat-v3-3-Slerp-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8fe782027ab138f7cc2d99c46de59787407b0b5677ea5f5370ce01f5e57e10a7
3
+ size 3518986816
OpenHermes-2.5-neural-chat-v3-3-Slerp-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0157bdebbea13b8bf1b1f2517acee9bb447fd8011d5ab054592ef0a7b684301e
3
+ size 3164568128
OpenHermes-2.5-neural-chat-v3-3-Slerp-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:613030573159619982ab366226daab5f8fd50c62dea53a49db1e6353c5b300d0
3
+ size 4108917312
OpenHermes-2.5-neural-chat-v3-3-Slerp-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29b154135c599feb5fa66823cdf7bda679fef8566e6b54798d01aa6b42d21193
3
+ size 4368439872
OpenHermes-2.5-neural-chat-v3-3-Slerp-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a44d02d405b7288483483bb1c7e70384cf582279e27f85901151422b2878ab5
3
+ size 4140374592
OpenHermes-2.5-neural-chat-v3-3-Slerp-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6afe0c950ed7ed36f7ebfbd9b6102bea99a3e6c2dd53292142405d951ea3275
3
+ size 4997716544
OpenHermes-2.5-neural-chat-v3-3-Slerp-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19d473625eb0dade335b2ea727bf63ae1810d47d592c0f167f7d8eb682d053c0
3
+ size 5131409984
OpenHermes-2.5-neural-chat-v3-3-Slerp-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a3e48a78c16f392f24692626274bdd330bbe2523cb0c3dd26351d97f9ae3aeb
3
+ size 4997716544
OpenHermes-2.5-neural-chat-v3-3-Slerp-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3058b411c3be1cb61d36f1d973e957be32b582f1af30837bead8aa82ca45016
3
+ size 5942065728
OpenHermes-2.5-neural-chat-v3-3-Slerp-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:637f198ee62d87535fb962897121f38c155bd5163eb4a34fb9b95004ac82574d
3
+ size 7695858240
README.md ADDED
@@ -0,0 +1,162 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - merge
5
+ - TensorBlock
6
+ - GGUF
7
+ base_model: Weyaxi/OpenHermes-2.5-neural-chat-v3-3-Slerp
8
+ model-index:
9
+ - name: OpenHermes-2.5-neural-chat-v3-3-Slerp
10
+ results:
11
+ - task:
12
+ type: text-generation
13
+ name: Text Generation
14
+ dataset:
15
+ name: AI2 Reasoning Challenge (25-Shot)
16
+ type: ai2_arc
17
+ config: ARC-Challenge
18
+ split: test
19
+ args:
20
+ num_few_shot: 25
21
+ metrics:
22
+ - type: acc_norm
23
+ value: 68.09
24
+ name: normalized accuracy
25
+ - task:
26
+ type: text-generation
27
+ name: Text Generation
28
+ dataset:
29
+ name: HellaSwag (10-Shot)
30
+ type: hellaswag
31
+ split: validation
32
+ args:
33
+ num_few_shot: 10
34
+ metrics:
35
+ - type: acc_norm
36
+ value: 86.2
37
+ name: normalized accuracy
38
+ - task:
39
+ type: text-generation
40
+ name: Text Generation
41
+ dataset:
42
+ name: MMLU (5-Shot)
43
+ type: cais/mmlu
44
+ config: all
45
+ split: test
46
+ args:
47
+ num_few_shot: 5
48
+ metrics:
49
+ - type: acc
50
+ value: 64.26
51
+ name: accuracy
52
+ - task:
53
+ type: text-generation
54
+ name: Text Generation
55
+ dataset:
56
+ name: TruthfulQA (0-shot)
57
+ type: truthful_qa
58
+ config: multiple_choice
59
+ split: validation
60
+ args:
61
+ num_few_shot: 0
62
+ metrics:
63
+ - type: mc2
64
+ value: 62.78
65
+ - task:
66
+ type: text-generation
67
+ name: Text Generation
68
+ dataset:
69
+ name: Winogrande (5-shot)
70
+ type: winogrande
71
+ config: winogrande_xl
72
+ split: validation
73
+ args:
74
+ num_few_shot: 5
75
+ metrics:
76
+ - type: acc
77
+ value: 79.16
78
+ name: accuracy
79
+ - task:
80
+ type: text-generation
81
+ name: Text Generation
82
+ dataset:
83
+ name: GSM8k (5-shot)
84
+ type: gsm8k
85
+ config: main
86
+ split: test
87
+ args:
88
+ num_few_shot: 5
89
+ metrics:
90
+ - type: acc
91
+ value: 67.78
92
+ name: accuracy
93
+ ---
94
+
95
+ <div style="width: auto; margin-left: auto; margin-right: auto">
96
+ <img src="https://i.imgur.com/jC7kdl8.jpeg" alt="TensorBlock" style="width: 100%; min-width: 400px; display: block; margin: auto;">
97
+ </div>
98
+ <div style="display: flex; justify-content: space-between; width: 100%;">
99
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
100
+ <p style="margin-top: 0.5em; margin-bottom: 0em;">
101
+ Feedback and support: TensorBlock's <a href="https://x.com/tensorblock_aoi">Twitter/X</a>, <a href="https://t.me/TensorBlock">Telegram Group</a> and <a href="https://x.com/tensorblock_aoi">Discord server</a>
102
+ </p>
103
+ </div>
104
+ </div>
105
+
106
+ ## Weyaxi/OpenHermes-2.5-neural-chat-v3-3-Slerp - GGUF
107
+
108
+ This repo contains GGUF format model files for [Weyaxi/OpenHermes-2.5-neural-chat-v3-3-Slerp](https://huggingface.co/Weyaxi/OpenHermes-2.5-neural-chat-v3-3-Slerp).
109
+
110
+ The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4242](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
111
+
112
+ <div style="text-align: left; margin: 20px 0;">
113
+ <a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
114
+ Run them on the TensorBlock client using your local machine ↗
115
+ </a>
116
+ </div>
117
+
118
+ ## Prompt template
119
+
120
+ ```
121
+
122
+ ```
123
+
124
+ ## Model file specification
125
+
126
+ | Filename | Quant type | File Size | Description |
127
+ | -------- | ---------- | --------- | ----------- |
128
+ | [OpenHermes-2.5-neural-chat-v3-3-Slerp-Q2_K.gguf](https://huggingface.co/tensorblock/OpenHermes-2.5-neural-chat-v3-3-Slerp-GGUF/blob/main/OpenHermes-2.5-neural-chat-v3-3-Slerp-Q2_K.gguf) | Q2_K | 2.719 GB | smallest, significant quality loss - not recommended for most purposes |
129
+ | [OpenHermes-2.5-neural-chat-v3-3-Slerp-Q3_K_S.gguf](https://huggingface.co/tensorblock/OpenHermes-2.5-neural-chat-v3-3-Slerp-GGUF/blob/main/OpenHermes-2.5-neural-chat-v3-3-Slerp-Q3_K_S.gguf) | Q3_K_S | 3.165 GB | very small, high quality loss |
130
+ | [OpenHermes-2.5-neural-chat-v3-3-Slerp-Q3_K_M.gguf](https://huggingface.co/tensorblock/OpenHermes-2.5-neural-chat-v3-3-Slerp-GGUF/blob/main/OpenHermes-2.5-neural-chat-v3-3-Slerp-Q3_K_M.gguf) | Q3_K_M | 3.519 GB | very small, high quality loss |
131
+ | [OpenHermes-2.5-neural-chat-v3-3-Slerp-Q3_K_L.gguf](https://huggingface.co/tensorblock/OpenHermes-2.5-neural-chat-v3-3-Slerp-GGUF/blob/main/OpenHermes-2.5-neural-chat-v3-3-Slerp-Q3_K_L.gguf) | Q3_K_L | 3.822 GB | small, substantial quality loss |
132
+ | [OpenHermes-2.5-neural-chat-v3-3-Slerp-Q4_0.gguf](https://huggingface.co/tensorblock/OpenHermes-2.5-neural-chat-v3-3-Slerp-GGUF/blob/main/OpenHermes-2.5-neural-chat-v3-3-Slerp-Q4_0.gguf) | Q4_0 | 4.109 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
133
+ | [OpenHermes-2.5-neural-chat-v3-3-Slerp-Q4_K_S.gguf](https://huggingface.co/tensorblock/OpenHermes-2.5-neural-chat-v3-3-Slerp-GGUF/blob/main/OpenHermes-2.5-neural-chat-v3-3-Slerp-Q4_K_S.gguf) | Q4_K_S | 4.140 GB | small, greater quality loss |
134
+ | [OpenHermes-2.5-neural-chat-v3-3-Slerp-Q4_K_M.gguf](https://huggingface.co/tensorblock/OpenHermes-2.5-neural-chat-v3-3-Slerp-GGUF/blob/main/OpenHermes-2.5-neural-chat-v3-3-Slerp-Q4_K_M.gguf) | Q4_K_M | 4.368 GB | medium, balanced quality - recommended |
135
+ | [OpenHermes-2.5-neural-chat-v3-3-Slerp-Q5_0.gguf](https://huggingface.co/tensorblock/OpenHermes-2.5-neural-chat-v3-3-Slerp-GGUF/blob/main/OpenHermes-2.5-neural-chat-v3-3-Slerp-Q5_0.gguf) | Q5_0 | 4.998 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
136
+ | [OpenHermes-2.5-neural-chat-v3-3-Slerp-Q5_K_S.gguf](https://huggingface.co/tensorblock/OpenHermes-2.5-neural-chat-v3-3-Slerp-GGUF/blob/main/OpenHermes-2.5-neural-chat-v3-3-Slerp-Q5_K_S.gguf) | Q5_K_S | 4.998 GB | large, low quality loss - recommended |
137
+ | [OpenHermes-2.5-neural-chat-v3-3-Slerp-Q5_K_M.gguf](https://huggingface.co/tensorblock/OpenHermes-2.5-neural-chat-v3-3-Slerp-GGUF/blob/main/OpenHermes-2.5-neural-chat-v3-3-Slerp-Q5_K_M.gguf) | Q5_K_M | 5.131 GB | large, very low quality loss - recommended |
138
+ | [OpenHermes-2.5-neural-chat-v3-3-Slerp-Q6_K.gguf](https://huggingface.co/tensorblock/OpenHermes-2.5-neural-chat-v3-3-Slerp-GGUF/blob/main/OpenHermes-2.5-neural-chat-v3-3-Slerp-Q6_K.gguf) | Q6_K | 5.942 GB | very large, extremely low quality loss |
139
+ | [OpenHermes-2.5-neural-chat-v3-3-Slerp-Q8_0.gguf](https://huggingface.co/tensorblock/OpenHermes-2.5-neural-chat-v3-3-Slerp-GGUF/blob/main/OpenHermes-2.5-neural-chat-v3-3-Slerp-Q8_0.gguf) | Q8_0 | 7.696 GB | very large, extremely low quality loss - not recommended |
140
+
141
+
142
+ ## Downloading instruction
143
+
144
+ ### Command line
145
+
146
+ Firstly, install Huggingface Client
147
+
148
+ ```shell
149
+ pip install -U "huggingface_hub[cli]"
150
+ ```
151
+
152
+ Then, downoad the individual model file the a local directory
153
+
154
+ ```shell
155
+ huggingface-cli download tensorblock/OpenHermes-2.5-neural-chat-v3-3-Slerp-GGUF --include "OpenHermes-2.5-neural-chat-v3-3-Slerp-Q2_K.gguf" --local-dir MY_LOCAL_DIR
156
+ ```
157
+
158
+ If you wanna download multiple model files with a pattern (e.g., `*Q4_K*gguf`), you can try:
159
+
160
+ ```shell
161
+ huggingface-cli download tensorblock/OpenHermes-2.5-neural-chat-v3-3-Slerp-GGUF --local-dir MY_LOCAL_DIR --local-dir-use-symlinks False --include='*Q4_K*gguf'
162
+ ```