morriszms commited on
Commit
eebd1ba
·
verified ·
1 Parent(s): b95d379

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,15 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ neural-chat-7b-v3-2-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ neural-chat-7b-v3-2-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ neural-chat-7b-v3-2-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ neural-chat-7b-v3-2-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ neural-chat-7b-v3-2-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ neural-chat-7b-v3-2-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ neural-chat-7b-v3-2-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ neural-chat-7b-v3-2-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ neural-chat-7b-v3-2-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ neural-chat-7b-v3-2-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ neural-chat-7b-v3-2-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
47
+ neural-chat-7b-v3-2-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,118 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - LLMs
5
+ - mistral
6
+ - math
7
+ - Intel
8
+ - TensorBlock
9
+ - GGUF
10
+ datasets:
11
+ - meta-math/MetaMathQA
12
+ language:
13
+ - en
14
+ base_model: Intel/neural-chat-7b-v3-2
15
+ model-index:
16
+ - name: neural-chat-7b-v3-2
17
+ results:
18
+ - task:
19
+ type: Large Language Model
20
+ name: Large Language Model
21
+ dataset:
22
+ name: meta-math/MetaMathQA
23
+ type: meta-math/MetaMathQA
24
+ metrics:
25
+ - type: ARC (25-shot)
26
+ value: 67.49
27
+ name: ARC (25-shot)
28
+ verified: true
29
+ - type: HellaSwag (10-shot)
30
+ value: 83.92
31
+ name: HellaSwag (10-shot)
32
+ verified: true
33
+ - type: MMLU (5-shot)
34
+ value: 63.55
35
+ name: MMLU (5-shot)
36
+ verified: true
37
+ - type: TruthfulQA (0-shot)
38
+ value: 59.68
39
+ name: TruthfulQA (0-shot)
40
+ verified: true
41
+ - type: Winogrande (5-shot)
42
+ value: 79.95
43
+ name: Winogrande (5-shot)
44
+ verified: true
45
+ - type: GSM8K (5-shot)
46
+ value: 55.12
47
+ name: GSM8K (5-shot)
48
+ verified: true
49
+ ---
50
+
51
+ <div style="width: auto; margin-left: auto; margin-right: auto">
52
+ <img src="https://i.imgur.com/jC7kdl8.jpeg" alt="TensorBlock" style="width: 100%; min-width: 400px; display: block; margin: auto;">
53
+ </div>
54
+ <div style="display: flex; justify-content: space-between; width: 100%;">
55
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
56
+ <p style="margin-top: 0.5em; margin-bottom: 0em;">
57
+ Feedback and support: TensorBlock's <a href="https://x.com/tensorblock_aoi">Twitter/X</a>, <a href="https://t.me/TensorBlock">Telegram Group</a> and <a href="https://x.com/tensorblock_aoi">Discord server</a>
58
+ </p>
59
+ </div>
60
+ </div>
61
+
62
+ ## Intel/neural-chat-7b-v3-2 - GGUF
63
+
64
+ This repo contains GGUF format model files for [Intel/neural-chat-7b-v3-2](https://huggingface.co/Intel/neural-chat-7b-v3-2).
65
+
66
+ The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4242](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
67
+
68
+ <div style="text-align: left; margin: 20px 0;">
69
+ <a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
70
+ Run them on the TensorBlock client using your local machine ↗
71
+ </a>
72
+ </div>
73
+
74
+ ## Prompt template
75
+
76
+ ```
77
+
78
+ ```
79
+
80
+ ## Model file specification
81
+
82
+ | Filename | Quant type | File Size | Description |
83
+ | -------- | ---------- | --------- | ----------- |
84
+ | [neural-chat-7b-v3-2-Q2_K.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-2-GGUF/blob/main/neural-chat-7b-v3-2-Q2_K.gguf) | Q2_K | 2.719 GB | smallest, significant quality loss - not recommended for most purposes |
85
+ | [neural-chat-7b-v3-2-Q3_K_S.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-2-GGUF/blob/main/neural-chat-7b-v3-2-Q3_K_S.gguf) | Q3_K_S | 3.165 GB | very small, high quality loss |
86
+ | [neural-chat-7b-v3-2-Q3_K_M.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-2-GGUF/blob/main/neural-chat-7b-v3-2-Q3_K_M.gguf) | Q3_K_M | 3.519 GB | very small, high quality loss |
87
+ | [neural-chat-7b-v3-2-Q3_K_L.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-2-GGUF/blob/main/neural-chat-7b-v3-2-Q3_K_L.gguf) | Q3_K_L | 3.822 GB | small, substantial quality loss |
88
+ | [neural-chat-7b-v3-2-Q4_0.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-2-GGUF/blob/main/neural-chat-7b-v3-2-Q4_0.gguf) | Q4_0 | 4.109 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
89
+ | [neural-chat-7b-v3-2-Q4_K_S.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-2-GGUF/blob/main/neural-chat-7b-v3-2-Q4_K_S.gguf) | Q4_K_S | 4.140 GB | small, greater quality loss |
90
+ | [neural-chat-7b-v3-2-Q4_K_M.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-2-GGUF/blob/main/neural-chat-7b-v3-2-Q4_K_M.gguf) | Q4_K_M | 4.368 GB | medium, balanced quality - recommended |
91
+ | [neural-chat-7b-v3-2-Q5_0.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-2-GGUF/blob/main/neural-chat-7b-v3-2-Q5_0.gguf) | Q5_0 | 4.998 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
92
+ | [neural-chat-7b-v3-2-Q5_K_S.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-2-GGUF/blob/main/neural-chat-7b-v3-2-Q5_K_S.gguf) | Q5_K_S | 4.998 GB | large, low quality loss - recommended |
93
+ | [neural-chat-7b-v3-2-Q5_K_M.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-2-GGUF/blob/main/neural-chat-7b-v3-2-Q5_K_M.gguf) | Q5_K_M | 5.131 GB | large, very low quality loss - recommended |
94
+ | [neural-chat-7b-v3-2-Q6_K.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-2-GGUF/blob/main/neural-chat-7b-v3-2-Q6_K.gguf) | Q6_K | 5.942 GB | very large, extremely low quality loss |
95
+ | [neural-chat-7b-v3-2-Q8_0.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-2-GGUF/blob/main/neural-chat-7b-v3-2-Q8_0.gguf) | Q8_0 | 7.696 GB | very large, extremely low quality loss - not recommended |
96
+
97
+
98
+ ## Downloading instruction
99
+
100
+ ### Command line
101
+
102
+ Firstly, install Huggingface Client
103
+
104
+ ```shell
105
+ pip install -U "huggingface_hub[cli]"
106
+ ```
107
+
108
+ Then, downoad the individual model file the a local directory
109
+
110
+ ```shell
111
+ huggingface-cli download tensorblock/neural-chat-7b-v3-2-GGUF --include "neural-chat-7b-v3-2-Q2_K.gguf" --local-dir MY_LOCAL_DIR
112
+ ```
113
+
114
+ If you wanna download multiple model files with a pattern (e.g., `*Q4_K*gguf`), you can try:
115
+
116
+ ```shell
117
+ huggingface-cli download tensorblock/neural-chat-7b-v3-2-GGUF --local-dir MY_LOCAL_DIR --local-dir-use-symlinks False --include='*Q4_K*gguf'
118
+ ```
neural-chat-7b-v3-2-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:271edb6405ac7d0c8f144e8b38b751697397b6f0699f94cfe9cd374d81add0d8
3
+ size 2719242560
neural-chat-7b-v3-2-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:976bed3f407e10e29db9cbf95c29fcc0858cf5f4c5a8383a3f4155f08cd91865
3
+ size 3822025024
neural-chat-7b-v3-2-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e78355fafd7e21b0e11f442953d3981b582af50bde7d810f5d0dd7417882e875
3
+ size 3518986560
neural-chat-7b-v3-2-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:942614571234fc846171dc40acff72086f8b8006df9ac202d6af028e81e5b617
3
+ size 3164567872
neural-chat-7b-v3-2-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:733179c6d9551f81b86a9201362f22ac393a19022f3791e0e3c1dbe6de7afef7
3
+ size 4108917056
neural-chat-7b-v3-2-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c90956b182e37eee57d1e0027ea0f4049535f266dae064cd013a9534f15431cb
3
+ size 4368439616
neural-chat-7b-v3-2-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:859789faf0852445a9183cf4ef5db1587f03c98ba143d87f5fefc98b5596cb3b
3
+ size 4140374336
neural-chat-7b-v3-2-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0329fc1289ed836d4388163b9c61f09e3488f9204b2dc6a3e82dff33bee05e36
3
+ size 4997716288
neural-chat-7b-v3-2-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71bf71aaafda21b1ed51ba684e592d7a73a6c75960a77709596bd8844e9c2278
3
+ size 5131409728
neural-chat-7b-v3-2-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac71185176db65768bf8593fffe4dee9f13e17a34c2490bd3032ea877d45b5e6
3
+ size 4997716288
neural-chat-7b-v3-2-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f40f0a277b20ed299d05534f1dabb01a32ddc29abf1cbe8f7eb56d4bb28aa08d
3
+ size 5942065472
neural-chat-7b-v3-2-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1bc0ea57f64e43242c64f83c35e8b772674e018f0dec331310c55d79bb351b44
3
+ size 7695857984