DeepSeek-R1-Distill-Llama-8B-GGUF / scores /DeepSeek-R1-Distill-Llama-8B-Q3_K_S.md
eaddario's picture
Add GGUF internal file structure
66279ab verified

DeepSeek-R1-Distill-Llama-8B-Q3_K_S.gguf - GGUF Internal File Dump

  • Endian: LITTLE endian

Key Value Metadata Store

There are 36 key-value pairs in this file

POS TYPE Count Key Value
1 UINT32 1 GGUF.version 3
2 UINT64 1 GGUF.tensor_count 292
3 UINT64 1 GGUF.kv_count 33
4 STRING 1 general.architecture llama
5 STRING 1 general.type model
6 STRING 1 general.name DeepSeek R1 Distill Llama 8B
7 STRING 1 general.basename DeepSeek-R1-Distill-Llama
8 STRING 1 general.size_label 8B
9 STRING 1 general.license mit
10 UINT32 1 llama.block_count 32
11 UINT32 1 llama.context_length 131072
12 UINT32 1 llama.embedding_length 4096
13 UINT32 1 llama.feed_forward_length 14336
14 UINT32 1 llama.attention.head_count 32
15 UINT32 1 llama.attention.head_count_kv 8
16 FLOAT32 1 llama.rope.freq_base 500000.0
17 FLOAT32 1 llama.attention.layer_norm_rms_epsilon 1e-05
18 UINT32 1 llama.vocab_size 128256
19 UINT32 1 llama.rope.dimension_count 128
20 STRING 1 tokenizer.ggml.model gpt2
21 STRING 1 tokenizer.ggml.pre llama-bpe
22 [STRING] 128256 tokenizer.ggml.tokens [ !, ", #, $, %, ... ]
23 [INT32] 128256 tokenizer.ggml.token_type [ 1, 1, 1, 1, 1, 1, 1, ... ]
24 [STRING] 280147 tokenizer.ggml.merges [ Ġ Ġ, Ġ ĠĠĠ, ĠĠ ĠĠ, ĠĠĠ Ġ, i n, ... ]
25 UINT32 1 tokenizer.ggml.bos_token_id 128000
26 UINT32 1 tokenizer.ggml.eos_token_id 128001
27 UINT32 1 tokenizer.ggml.padding_token_id 128001
28 BOOL 1 tokenizer.ggml.add_bos_token True
29 BOOL 1 tokenizer.ggml.add_eos_token False
30 STRING 1 tokenizer.chat_template {% if not add_generation_promp...{{'<|Assistant|>'}}{% endif %}
31 UINT32 1 general.quantization_version 2
32 UINT32 1 general.file_type 11
33 STRING 1 quantize.imatrix.file ./imatrix/imatrix-DeepSeek-R1-Distill-Llama-8B-small.dat
34 STRING 1 quantize.imatrix.dataset ../../datasets/imatrix/calibration_all_small.txt
35 INT32 1 quantize.imatrix.entries_count 225
36 INT32 1 quantize.imatrix.chunks_count 1130

Tensors Overview ~8B Elements

Total number of elements in all tensors: 8030261312 Elements

Tensor Data Offset

This table contains the offset and data segment relative to start of file

T_ID Tensor Layer Name Data Offset (B) Data Size (B)
0 output.weight 0x779a80 0xd746000
1 output_norm.weight 0xdebfa80 0x4000
2 rope_freqs.weight 0xdec3a80 0x100
3 token_embd.weight 0xdec3b80 0xa464000
4 blk.0.attn_k.weight 0x18327b80 0x1b8000
5 blk.0.attn_norm.weight 0x184dfb80 0x4000
6 blk.0.attn_output.weight 0x184e3b80 0x6e0000
7 blk.0.attn_q.weight 0x18bc3b80 0x6e0000
8 blk.0.attn_v.weight 0x192a3b80 0x1b8000
9 blk.0.ffn_down.weight 0x1945bb80 0x1f80000
10 blk.0.ffn_gate.weight 0x1b3dbb80 0x1810000
11 blk.0.ffn_norm.weight 0x1cbebb80 0x4000
12 blk.0.ffn_up.weight 0x1cbefb80 0x1810000
13 blk.1.attn_k.weight 0x1e3ffb80 0x1b8000
14 blk.1.attn_norm.weight 0x1e5b7b80 0x4000
15 blk.1.attn_output.weight 0x1e5bbb80 0x6e0000
16 blk.1.attn_q.weight 0x1ec9bb80 0x6e0000
17 blk.1.attn_v.weight 0x1f37bb80 0x1b8000
18 blk.1.ffn_down.weight 0x1f533b80 0x1f80000
19 blk.1.ffn_gate.weight 0x214b3b80 0x1810000
20 blk.1.ffn_norm.weight 0x22cc3b80 0x4000
21 blk.1.ffn_up.weight 0x22cc7b80 0x1810000
22 blk.2.attn_k.weight 0x244d7b80 0x1b8000
23 blk.2.attn_norm.weight 0x2468fb80 0x4000
24 blk.2.attn_output.weight 0x24693b80 0x6e0000
25 blk.2.attn_q.weight 0x24d73b80 0x6e0000
26 blk.2.attn_v.weight 0x25453b80 0x1b8000
27 blk.2.ffn_down.weight 0x2560bb80 0x1f80000
28 blk.2.ffn_gate.weight 0x2758bb80 0x1810000
29 blk.2.ffn_norm.weight 0x28d9bb80 0x4000
30 blk.2.ffn_up.weight 0x28d9fb80 0x1810000
31 blk.3.attn_k.weight 0x2a5afb80 0x1b8000
32 blk.3.attn_norm.weight 0x2a767b80 0x4000
33 blk.3.attn_output.weight 0x2a76bb80 0x6e0000
34 blk.3.attn_q.weight 0x2ae4bb80 0x6e0000
35 blk.3.attn_v.weight 0x2b52bb80 0x1b8000
36 blk.3.ffn_down.weight 0x2b6e3b80 0x1f80000
37 blk.3.ffn_gate.weight 0x2d663b80 0x1810000
38 blk.3.ffn_norm.weight 0x2ee73b80 0x4000
39 blk.3.ffn_up.weight 0x2ee77b80 0x1810000
40 blk.4.attn_k.weight 0x30687b80 0x1b8000
41 blk.4.attn_norm.weight 0x3083fb80 0x4000
42 blk.4.attn_output.weight 0x30843b80 0x6e0000
43 blk.4.attn_q.weight 0x30f23b80 0x6e0000
44 blk.4.attn_v.weight 0x31603b80 0x1b8000
45 blk.4.ffn_down.weight 0x317bbb80 0x1f80000
46 blk.4.ffn_gate.weight 0x3373bb80 0x1810000
47 blk.4.ffn_norm.weight 0x34f4bb80 0x4000
48 blk.4.ffn_up.weight 0x34f4fb80 0x1810000
49 blk.5.attn_k.weight 0x3675fb80 0x1b8000
50 blk.5.attn_norm.weight 0x36917b80 0x4000
51 blk.5.attn_output.weight 0x3691bb80 0x6e0000
52 blk.5.attn_q.weight 0x36ffbb80 0x6e0000
53 blk.5.attn_v.weight 0x376dbb80 0x1b8000
54 blk.5.ffn_down.weight 0x37893b80 0x1f80000
55 blk.5.ffn_gate.weight 0x39813b80 0x1810000
56 blk.5.ffn_norm.weight 0x3b023b80 0x4000
57 blk.5.ffn_up.weight 0x3b027b80 0x1810000
58 blk.6.attn_k.weight 0x3c837b80 0x1b8000
59 blk.6.attn_norm.weight 0x3c9efb80 0x4000
60 blk.6.attn_output.weight 0x3c9f3b80 0x6e0000
61 blk.6.attn_q.weight 0x3d0d3b80 0x6e0000
62 blk.6.attn_v.weight 0x3d7b3b80 0x1b8000
63 blk.6.ffn_down.weight 0x3d96bb80 0x1f80000
64 blk.6.ffn_gate.weight 0x3f8ebb80 0x1810000
65 blk.6.ffn_norm.weight 0x410fbb80 0x4000
66 blk.6.ffn_up.weight 0x410ffb80 0x1810000
67 blk.7.attn_k.weight 0x4290fb80 0x1b8000
68 blk.7.attn_norm.weight 0x42ac7b80 0x4000
69 blk.7.attn_output.weight 0x42acbb80 0x6e0000
70 blk.7.attn_q.weight 0x431abb80 0x6e0000
71 blk.7.attn_v.weight 0x4388bb80 0x1b8000
72 blk.7.ffn_down.weight 0x43a43b80 0x1f80000
73 blk.7.ffn_gate.weight 0x459c3b80 0x1810000
74 blk.7.ffn_norm.weight 0x471d3b80 0x4000
75 blk.7.ffn_up.weight 0x471d7b80 0x1810000
76 blk.8.attn_k.weight 0x489e7b80 0x1b8000
77 blk.8.attn_norm.weight 0x48b9fb80 0x4000
78 blk.8.attn_output.weight 0x48ba3b80 0x6e0000
79 blk.8.attn_q.weight 0x49283b80 0x6e0000
80 blk.8.attn_v.weight 0x49963b80 0x1b8000
81 blk.8.ffn_down.weight 0x49b1bb80 0x1f80000
82 blk.8.ffn_gate.weight 0x4ba9bb80 0x1810000
83 blk.8.ffn_norm.weight 0x4d2abb80 0x4000
84 blk.8.ffn_up.weight 0x4d2afb80 0x1810000
85 blk.9.attn_k.weight 0x4eabfb80 0x1b8000
86 blk.9.attn_norm.weight 0x4ec77b80 0x4000
87 blk.9.attn_output.weight 0x4ec7bb80 0x6e0000
88 blk.9.attn_q.weight 0x4f35bb80 0x6e0000
89 blk.9.attn_v.weight 0x4fa3bb80 0x1b8000
90 blk.9.ffn_down.weight 0x4fbf3b80 0x1f80000
91 blk.9.ffn_gate.weight 0x51b73b80 0x1810000
92 blk.9.ffn_norm.weight 0x53383b80 0x4000
93 blk.9.ffn_up.weight 0x53387b80 0x1810000
94 blk.10.attn_k.weight 0x54b97b80 0x1b8000
95 blk.10.attn_norm.weight 0x54d4fb80 0x4000
96 blk.10.attn_output.weight 0x54d53b80 0x6e0000
97 blk.10.attn_q.weight 0x55433b80 0x6e0000
98 blk.10.attn_v.weight 0x55b13b80 0x1b8000
99 blk.10.ffn_down.weight 0x55ccbb80 0x1f80000
100 blk.10.ffn_gate.weight 0x57c4bb80 0x1810000
101 blk.10.ffn_norm.weight 0x5945bb80 0x4000
102 blk.10.ffn_up.weight 0x5945fb80 0x1810000
103 blk.11.attn_k.weight 0x5ac6fb80 0x1b8000
104 blk.11.attn_norm.weight 0x5ae27b80 0x4000
105 blk.11.attn_output.weight 0x5ae2bb80 0x6e0000
106 blk.11.attn_q.weight 0x5b50bb80 0x6e0000
107 blk.11.attn_v.weight 0x5bbebb80 0x1b8000
108 blk.11.ffn_down.weight 0x5bda3b80 0x1f80000
109 blk.11.ffn_gate.weight 0x5dd23b80 0x1810000
110 blk.11.ffn_norm.weight 0x5f533b80 0x4000
111 blk.11.ffn_up.weight 0x5f537b80 0x1810000
112 blk.12.attn_k.weight 0x60d47b80 0x1b8000
113 blk.12.attn_norm.weight 0x60effb80 0x4000
114 blk.12.attn_output.weight 0x60f03b80 0x6e0000
115 blk.12.attn_q.weight 0x615e3b80 0x6e0000
116 blk.12.attn_v.weight 0x61cc3b80 0x1b8000
117 blk.12.ffn_down.weight 0x61e7bb80 0x1f80000
118 blk.12.ffn_gate.weight 0x63dfbb80 0x1810000
119 blk.12.ffn_norm.weight 0x6560bb80 0x4000
120 blk.12.ffn_up.weight 0x6560fb80 0x1810000
121 blk.13.attn_k.weight 0x66e1fb80 0x1b8000
122 blk.13.attn_norm.weight 0x66fd7b80 0x4000
123 blk.13.attn_output.weight 0x66fdbb80 0x6e0000
124 blk.13.attn_q.weight 0x676bbb80 0x6e0000
125 blk.13.attn_v.weight 0x67d9bb80 0x1b8000
126 blk.13.ffn_down.weight 0x67f53b80 0x1f80000
127 blk.13.ffn_gate.weight 0x69ed3b80 0x1810000
128 blk.13.ffn_norm.weight 0x6b6e3b80 0x4000
129 blk.13.ffn_up.weight 0x6b6e7b80 0x1810000
130 blk.14.attn_k.weight 0x6cef7b80 0x1b8000
131 blk.14.attn_norm.weight 0x6d0afb80 0x4000
132 blk.14.attn_output.weight 0x6d0b3b80 0x6e0000
133 blk.14.attn_q.weight 0x6d793b80 0x6e0000
134 blk.14.attn_v.weight 0x6de73b80 0x1b8000
135 blk.14.ffn_down.weight 0x6e02bb80 0x1f80000
136 blk.14.ffn_gate.weight 0x6ffabb80 0x1810000
137 blk.14.ffn_norm.weight 0x717bbb80 0x4000
138 blk.14.ffn_up.weight 0x717bfb80 0x1810000
139 blk.15.attn_k.weight 0x72fcfb80 0x1b8000
140 blk.15.attn_norm.weight 0x73187b80 0x4000
141 blk.15.attn_output.weight 0x7318bb80 0x6e0000
142 blk.15.attn_q.weight 0x7386bb80 0x6e0000
143 blk.15.attn_v.weight 0x73f4bb80 0x1b8000
144 blk.15.ffn_down.weight 0x74103b80 0x1f80000
145 blk.15.ffn_gate.weight 0x76083b80 0x1810000
146 blk.15.ffn_norm.weight 0x77893b80 0x4000
147 blk.15.ffn_up.weight 0x77897b80 0x1810000
148 blk.16.attn_k.weight 0x790a7b80 0x150000
149 blk.16.attn_norm.weight 0x791f7b80 0x4000
150 blk.16.attn_output.weight 0x791fbb80 0x6e0000
151 blk.16.attn_q.weight 0x798dbb80 0x540000
152 blk.16.attn_v.weight 0x79e1bb80 0x1b8000
153 blk.16.ffn_down.weight 0x79fd3b80 0x1810000
154 blk.16.ffn_gate.weight 0x7b7e3b80 0x1260000
155 blk.16.ffn_norm.weight 0x7ca43b80 0x4000
156 blk.16.ffn_up.weight 0x7ca47b80 0x1260000
157 blk.17.attn_k.weight 0x7dca7b80 0x150000
158 blk.17.attn_norm.weight 0x7ddf7b80 0x4000
159 blk.17.attn_output.weight 0x7ddfbb80 0x6e0000
160 blk.17.attn_q.weight 0x7e4dbb80 0x540000
161 blk.17.attn_v.weight 0x7ea1bb80 0x1b8000
162 blk.17.ffn_down.weight 0x7ebd3b80 0x1810000
163 blk.17.ffn_gate.weight 0x803e3b80 0x1260000
164 blk.17.ffn_norm.weight 0x81643b80 0x4000
165 blk.17.ffn_up.weight 0x81647b80 0x1260000
166 blk.18.attn_k.weight 0x828a7b80 0x150000
167 blk.18.attn_norm.weight 0x829f7b80 0x4000
168 blk.18.attn_output.weight 0x829fbb80 0x6e0000
169 blk.18.attn_q.weight 0x830dbb80 0x540000
170 blk.18.attn_v.weight 0x8361bb80 0x1b8000
171 blk.18.ffn_down.weight 0x837d3b80 0x1810000
172 blk.18.ffn_gate.weight 0x84fe3b80 0x1260000
173 blk.18.ffn_norm.weight 0x86243b80 0x4000
174 blk.18.ffn_up.weight 0x86247b80 0x1260000
175 blk.19.attn_k.weight 0x874a7b80 0x150000
176 blk.19.attn_norm.weight 0x875f7b80 0x4000
177 blk.19.attn_output.weight 0x875fbb80 0x6e0000
178 blk.19.attn_q.weight 0x87cdbb80 0x540000
179 blk.19.attn_v.weight 0x8821bb80 0x1b8000
180 blk.19.ffn_down.weight 0x883d3b80 0x1810000
181 blk.19.ffn_gate.weight 0x89be3b80 0x1260000
182 blk.19.ffn_norm.weight 0x8ae43b80 0x4000
183 blk.19.ffn_up.weight 0x8ae47b80 0x1260000
184 blk.20.attn_k.weight 0x8c0a7b80 0x150000
185 blk.20.attn_norm.weight 0x8c1f7b80 0x4000
186 blk.20.attn_output.weight 0x8c1fbb80 0x6e0000
187 blk.20.attn_q.weight 0x8c8dbb80 0x540000
188 blk.20.attn_v.weight 0x8ce1bb80 0x1b8000
189 blk.20.ffn_down.weight 0x8cfd3b80 0x1810000
190 blk.20.ffn_gate.weight 0x8e7e3b80 0x1260000
191 blk.20.ffn_norm.weight 0x8fa43b80 0x4000
192 blk.20.ffn_up.weight 0x8fa47b80 0x1260000
193 blk.21.attn_k.weight 0x90ca7b80 0x150000
194 blk.21.attn_norm.weight 0x90df7b80 0x4000
195 blk.21.attn_output.weight 0x90dfbb80 0x6e0000
196 blk.21.attn_q.weight 0x914dbb80 0x540000
197 blk.21.attn_v.weight 0x91a1bb80 0x1b8000
198 blk.21.ffn_down.weight 0x91bd3b80 0x1810000
199 blk.21.ffn_gate.weight 0x933e3b80 0x1260000
200 blk.21.ffn_norm.weight 0x94643b80 0x4000
201 blk.21.ffn_up.weight 0x94647b80 0x1260000
202 blk.22.attn_k.weight 0x958a7b80 0x150000
203 blk.22.attn_norm.weight 0x959f7b80 0x4000
204 blk.22.attn_output.weight 0x959fbb80 0x6e0000
205 blk.22.attn_q.weight 0x960dbb80 0x540000
206 blk.22.attn_v.weight 0x9661bb80 0x1b8000
207 blk.22.ffn_down.weight 0x967d3b80 0x1810000
208 blk.22.ffn_gate.weight 0x97fe3b80 0x1260000
209 blk.22.ffn_norm.weight 0x99243b80 0x4000
210 blk.22.ffn_up.weight 0x99247b80 0x1260000
211 blk.23.attn_k.weight 0x9a4a7b80 0x150000
212 blk.23.attn_norm.weight 0x9a5f7b80 0x4000
213 blk.23.attn_output.weight 0x9a5fbb80 0x6e0000
214 blk.23.attn_q.weight 0x9acdbb80 0x540000
215 blk.23.attn_v.weight 0x9b21bb80 0x1b8000
216 blk.23.ffn_down.weight 0x9b3d3b80 0x1810000
217 blk.23.ffn_gate.weight 0x9cbe3b80 0x1260000
218 blk.23.ffn_norm.weight 0x9de43b80 0x4000
219 blk.23.ffn_up.weight 0x9de47b80 0x1260000
220 blk.24.attn_k.weight 0x9f0a7b80 0x150000
221 blk.24.attn_norm.weight 0x9f1f7b80 0x4000
222 blk.24.attn_output.weight 0x9f1fbb80 0x6e0000
223 blk.24.attn_q.weight 0x9f8dbb80 0x540000
224 blk.24.attn_v.weight 0x9fe1bb80 0x1b8000
225 blk.24.ffn_down.weight 0x9ffd3b80 0x1810000
226 blk.24.ffn_gate.weight 0xa17e3b80 0x1260000
227 blk.24.ffn_norm.weight 0xa2a43b80 0x4000
228 blk.24.ffn_up.weight 0xa2a47b80 0x1260000
229 blk.25.attn_k.weight 0xa3ca7b80 0x150000
230 blk.25.attn_norm.weight 0xa3df7b80 0x4000
231 blk.25.attn_output.weight 0xa3dfbb80 0x6e0000
232 blk.25.attn_q.weight 0xa44dbb80 0x540000
233 blk.25.attn_v.weight 0xa4a1bb80 0x1b8000
234 blk.25.ffn_down.weight 0xa4bd3b80 0x1810000
235 blk.25.ffn_gate.weight 0xa63e3b80 0x1260000
236 blk.25.ffn_norm.weight 0xa7643b80 0x4000
237 blk.25.ffn_up.weight 0xa7647b80 0x1260000
238 blk.26.attn_k.weight 0xa88a7b80 0x150000
239 blk.26.attn_norm.weight 0xa89f7b80 0x4000
240 blk.26.attn_output.weight 0xa89fbb80 0x6e0000
241 blk.26.attn_q.weight 0xa90dbb80 0x540000
242 blk.26.attn_v.weight 0xa961bb80 0x1b8000
243 blk.26.ffn_down.weight 0xa97d3b80 0x1810000
244 blk.26.ffn_gate.weight 0xaafe3b80 0x1260000
245 blk.26.ffn_norm.weight 0xac243b80 0x4000
246 blk.26.ffn_up.weight 0xac247b80 0x1260000
247 blk.27.attn_k.weight 0xad4a7b80 0x150000
248 blk.27.attn_norm.weight 0xad5f7b80 0x4000
249 blk.27.attn_output.weight 0xad5fbb80 0x6e0000
250 blk.27.attn_q.weight 0xadcdbb80 0x540000
251 blk.27.attn_v.weight 0xae21bb80 0x1b8000
252 blk.27.ffn_down.weight 0xae3d3b80 0x1810000
253 blk.27.ffn_gate.weight 0xafbe3b80 0x1260000
254 blk.27.ffn_norm.weight 0xb0e43b80 0x4000
255 blk.27.ffn_up.weight 0xb0e47b80 0x1260000
256 blk.28.attn_k.weight 0xb20a7b80 0x150000
257 blk.28.attn_norm.weight 0xb21f7b80 0x4000
258 blk.28.attn_output.weight 0xb21fbb80 0x6e0000
259 blk.28.attn_q.weight 0xb28dbb80 0x540000
260 blk.28.attn_v.weight 0xb2e1bb80 0x1b8000
261 blk.28.ffn_down.weight 0xb2fd3b80 0x1810000
262 blk.28.ffn_gate.weight 0xb47e3b80 0x1260000
263 blk.28.ffn_norm.weight 0xb5a43b80 0x4000
264 blk.28.ffn_up.weight 0xb5a47b80 0x1260000
265 blk.29.attn_k.weight 0xb6ca7b80 0x150000
266 blk.29.attn_norm.weight 0xb6df7b80 0x4000
267 blk.29.attn_output.weight 0xb6dfbb80 0x6e0000
268 blk.29.attn_q.weight 0xb74dbb80 0x540000
269 blk.29.attn_v.weight 0xb7a1bb80 0x1b8000
270 blk.29.ffn_down.weight 0xb7bd3b80 0x1810000
271 blk.29.ffn_gate.weight 0xb93e3b80 0x1260000
272 blk.29.ffn_norm.weight 0xba643b80 0x4000
273 blk.29.ffn_up.weight 0xba647b80 0x1260000
274 blk.30.attn_k.weight 0xbb8a7b80 0x150000
275 blk.30.attn_norm.weight 0xbb9f7b80 0x4000
276 blk.30.attn_output.weight 0xbb9fbb80 0x6e0000
277 blk.30.attn_q.weight 0xbc0dbb80 0x540000
278 blk.30.attn_v.weight 0xbc61bb80 0x1b8000
279 blk.30.ffn_down.weight 0xbc7d3b80 0x1810000
280 blk.30.ffn_gate.weight 0xbdfe3b80 0x1260000
281 blk.30.ffn_norm.weight 0xbf243b80 0x4000
282 blk.30.ffn_up.weight 0xbf247b80 0x1260000
283 blk.31.attn_k.weight 0xc04a7b80 0x150000
284 blk.31.attn_norm.weight 0xc05f7b80 0x4000
285 blk.31.attn_output.weight 0xc05fbb80 0x6e0000
286 blk.31.attn_q.weight 0xc0cdbb80 0x540000
287 blk.31.attn_v.weight 0xc121bb80 0x1b8000
288 blk.31.ffn_down.weight 0xc13d3b80 0x1810000
289 blk.31.ffn_gate.weight 0xc2be3b80 0x1260000
290 blk.31.ffn_norm.weight 0xc3e43b80 0x4000
291 blk.31.ffn_up.weight 0xc3e47b80 0x1260000

Base Tensor Group : ~1B Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
0 output.weight Output (W) (~525M) 525336576 4096 x 128256 x 1 x 1 Q3_K
1 output_norm.weight Output Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
2 rope_freqs.weight Rope_Freqs (W) ( 64) 64 64 x 1 x 1 x 1 F32
3 token_embd.weight Token Embedding (W) (~525M) 525336576 4096 x 128256 x 1 x 1 Q2_K
  • Total elements in base: ( ~1B) 1050677312
  • Percentage of total elements: 13.08%

Block 0 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
4 blk.0.attn_k.weight Block 0 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
5 blk.0.attn_norm.weight Block 0 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
6 blk.0.attn_output.weight Block 0 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
7 blk.0.attn_q.weight Block 0 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
8 blk.0.attn_v.weight Block 0 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
9 blk.0.ffn_down.weight Block 0 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q4_K
10 blk.0.ffn_gate.weight Block 0 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
11 blk.0.ffn_norm.weight Block 0 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
12 blk.0.ffn_up.weight Block 0 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
  • Total elements in blk.0: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 1 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
13 blk.1.attn_k.weight Block 1 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
14 blk.1.attn_norm.weight Block 1 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
15 blk.1.attn_output.weight Block 1 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
16 blk.1.attn_q.weight Block 1 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
17 blk.1.attn_v.weight Block 1 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
18 blk.1.ffn_down.weight Block 1 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q4_K
19 blk.1.ffn_gate.weight Block 1 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
20 blk.1.ffn_norm.weight Block 1 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
21 blk.1.ffn_up.weight Block 1 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
  • Total elements in blk.1: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 2 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
22 blk.2.attn_k.weight Block 2 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
23 blk.2.attn_norm.weight Block 2 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
24 blk.2.attn_output.weight Block 2 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
25 blk.2.attn_q.weight Block 2 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
26 blk.2.attn_v.weight Block 2 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
27 blk.2.ffn_down.weight Block 2 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q4_K
28 blk.2.ffn_gate.weight Block 2 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
29 blk.2.ffn_norm.weight Block 2 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
30 blk.2.ffn_up.weight Block 2 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
  • Total elements in blk.2: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 3 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
31 blk.3.attn_k.weight Block 3 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
32 blk.3.attn_norm.weight Block 3 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
33 blk.3.attn_output.weight Block 3 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
34 blk.3.attn_q.weight Block 3 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
35 blk.3.attn_v.weight Block 3 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
36 blk.3.ffn_down.weight Block 3 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q4_K
37 blk.3.ffn_gate.weight Block 3 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
38 blk.3.ffn_norm.weight Block 3 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
39 blk.3.ffn_up.weight Block 3 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
  • Total elements in blk.3: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 4 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
40 blk.4.attn_k.weight Block 4 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
41 blk.4.attn_norm.weight Block 4 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
42 blk.4.attn_output.weight Block 4 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
43 blk.4.attn_q.weight Block 4 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
44 blk.4.attn_v.weight Block 4 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
45 blk.4.ffn_down.weight Block 4 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q4_K
46 blk.4.ffn_gate.weight Block 4 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
47 blk.4.ffn_norm.weight Block 4 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
48 blk.4.ffn_up.weight Block 4 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
  • Total elements in blk.4: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 5 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
49 blk.5.attn_k.weight Block 5 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
50 blk.5.attn_norm.weight Block 5 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
51 blk.5.attn_output.weight Block 5 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
52 blk.5.attn_q.weight Block 5 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
53 blk.5.attn_v.weight Block 5 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
54 blk.5.ffn_down.weight Block 5 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q4_K
55 blk.5.ffn_gate.weight Block 5 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
56 blk.5.ffn_norm.weight Block 5 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
57 blk.5.ffn_up.weight Block 5 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
  • Total elements in blk.5: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 6 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
58 blk.6.attn_k.weight Block 6 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
59 blk.6.attn_norm.weight Block 6 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
60 blk.6.attn_output.weight Block 6 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
61 blk.6.attn_q.weight Block 6 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
62 blk.6.attn_v.weight Block 6 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
63 blk.6.ffn_down.weight Block 6 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q4_K
64 blk.6.ffn_gate.weight Block 6 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
65 blk.6.ffn_norm.weight Block 6 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
66 blk.6.ffn_up.weight Block 6 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
  • Total elements in blk.6: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 7 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
67 blk.7.attn_k.weight Block 7 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
68 blk.7.attn_norm.weight Block 7 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
69 blk.7.attn_output.weight Block 7 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
70 blk.7.attn_q.weight Block 7 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
71 blk.7.attn_v.weight Block 7 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
72 blk.7.ffn_down.weight Block 7 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q4_K
73 blk.7.ffn_gate.weight Block 7 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
74 blk.7.ffn_norm.weight Block 7 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
75 blk.7.ffn_up.weight Block 7 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
  • Total elements in blk.7: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 8 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
76 blk.8.attn_k.weight Block 8 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
77 blk.8.attn_norm.weight Block 8 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
78 blk.8.attn_output.weight Block 8 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
79 blk.8.attn_q.weight Block 8 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
80 blk.8.attn_v.weight Block 8 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
81 blk.8.ffn_down.weight Block 8 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q4_K
82 blk.8.ffn_gate.weight Block 8 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
83 blk.8.ffn_norm.weight Block 8 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
84 blk.8.ffn_up.weight Block 8 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
  • Total elements in blk.8: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 9 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
85 blk.9.attn_k.weight Block 9 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
86 blk.9.attn_norm.weight Block 9 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
87 blk.9.attn_output.weight Block 9 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
88 blk.9.attn_q.weight Block 9 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
89 blk.9.attn_v.weight Block 9 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
90 blk.9.ffn_down.weight Block 9 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q4_K
91 blk.9.ffn_gate.weight Block 9 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
92 blk.9.ffn_norm.weight Block 9 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
93 blk.9.ffn_up.weight Block 9 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
  • Total elements in blk.9: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 10 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
94 blk.10.attn_k.weight Block 10 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
95 blk.10.attn_norm.weight Block 10 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
96 blk.10.attn_output.weight Block 10 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
97 blk.10.attn_q.weight Block 10 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
98 blk.10.attn_v.weight Block 10 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
99 blk.10.ffn_down.weight Block 10 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q4_K
100 blk.10.ffn_gate.weight Block 10 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
101 blk.10.ffn_norm.weight Block 10 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
102 blk.10.ffn_up.weight Block 10 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
  • Total elements in blk.10: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 11 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
103 blk.11.attn_k.weight Block 11 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
104 blk.11.attn_norm.weight Block 11 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
105 blk.11.attn_output.weight Block 11 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
106 blk.11.attn_q.weight Block 11 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
107 blk.11.attn_v.weight Block 11 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
108 blk.11.ffn_down.weight Block 11 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q4_K
109 blk.11.ffn_gate.weight Block 11 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
110 blk.11.ffn_norm.weight Block 11 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
111 blk.11.ffn_up.weight Block 11 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
  • Total elements in blk.11: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 12 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
112 blk.12.attn_k.weight Block 12 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
113 blk.12.attn_norm.weight Block 12 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
114 blk.12.attn_output.weight Block 12 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
115 blk.12.attn_q.weight Block 12 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
116 blk.12.attn_v.weight Block 12 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
117 blk.12.ffn_down.weight Block 12 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q4_K
118 blk.12.ffn_gate.weight Block 12 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
119 blk.12.ffn_norm.weight Block 12 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
120 blk.12.ffn_up.weight Block 12 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
  • Total elements in blk.12: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 13 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
121 blk.13.attn_k.weight Block 13 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
122 blk.13.attn_norm.weight Block 13 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
123 blk.13.attn_output.weight Block 13 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
124 blk.13.attn_q.weight Block 13 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
125 blk.13.attn_v.weight Block 13 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
126 blk.13.ffn_down.weight Block 13 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q4_K
127 blk.13.ffn_gate.weight Block 13 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
128 blk.13.ffn_norm.weight Block 13 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
129 blk.13.ffn_up.weight Block 13 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
  • Total elements in blk.13: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 14 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
130 blk.14.attn_k.weight Block 14 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
131 blk.14.attn_norm.weight Block 14 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
132 blk.14.attn_output.weight Block 14 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
133 blk.14.attn_q.weight Block 14 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
134 blk.14.attn_v.weight Block 14 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
135 blk.14.ffn_down.weight Block 14 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q4_K
136 blk.14.ffn_gate.weight Block 14 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
137 blk.14.ffn_norm.weight Block 14 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
138 blk.14.ffn_up.weight Block 14 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
  • Total elements in blk.14: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 15 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
139 blk.15.attn_k.weight Block 15 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
140 blk.15.attn_norm.weight Block 15 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
141 blk.15.attn_output.weight Block 15 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
142 blk.15.attn_q.weight Block 15 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
143 blk.15.attn_v.weight Block 15 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
144 blk.15.ffn_down.weight Block 15 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q4_K
145 blk.15.ffn_gate.weight Block 15 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
146 blk.15.ffn_norm.weight Block 15 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
147 blk.15.ffn_up.weight Block 15 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q3_K
  • Total elements in blk.15: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 16 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
148 blk.16.attn_k.weight Block 16 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q2_K
149 blk.16.attn_norm.weight Block 16 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
150 blk.16.attn_output.weight Block 16 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
151 blk.16.attn_q.weight Block 16 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q2_K
152 blk.16.attn_v.weight Block 16 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
153 blk.16.ffn_down.weight Block 16 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q3_K
154 blk.16.ffn_gate.weight Block 16 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
155 blk.16.ffn_norm.weight Block 16 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
156 blk.16.ffn_up.weight Block 16 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
  • Total elements in blk.16: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 17 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
157 blk.17.attn_k.weight Block 17 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q2_K
158 blk.17.attn_norm.weight Block 17 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
159 blk.17.attn_output.weight Block 17 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
160 blk.17.attn_q.weight Block 17 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q2_K
161 blk.17.attn_v.weight Block 17 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
162 blk.17.ffn_down.weight Block 17 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q3_K
163 blk.17.ffn_gate.weight Block 17 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
164 blk.17.ffn_norm.weight Block 17 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
165 blk.17.ffn_up.weight Block 17 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
  • Total elements in blk.17: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 18 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
166 blk.18.attn_k.weight Block 18 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q2_K
167 blk.18.attn_norm.weight Block 18 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
168 blk.18.attn_output.weight Block 18 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
169 blk.18.attn_q.weight Block 18 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q2_K
170 blk.18.attn_v.weight Block 18 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
171 blk.18.ffn_down.weight Block 18 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q3_K
172 blk.18.ffn_gate.weight Block 18 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
173 blk.18.ffn_norm.weight Block 18 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
174 blk.18.ffn_up.weight Block 18 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
  • Total elements in blk.18: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 19 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
175 blk.19.attn_k.weight Block 19 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q2_K
176 blk.19.attn_norm.weight Block 19 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
177 blk.19.attn_output.weight Block 19 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
178 blk.19.attn_q.weight Block 19 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q2_K
179 blk.19.attn_v.weight Block 19 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
180 blk.19.ffn_down.weight Block 19 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q3_K
181 blk.19.ffn_gate.weight Block 19 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
182 blk.19.ffn_norm.weight Block 19 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
183 blk.19.ffn_up.weight Block 19 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
  • Total elements in blk.19: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 20 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
184 blk.20.attn_k.weight Block 20 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q2_K
185 blk.20.attn_norm.weight Block 20 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
186 blk.20.attn_output.weight Block 20 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
187 blk.20.attn_q.weight Block 20 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q2_K
188 blk.20.attn_v.weight Block 20 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
189 blk.20.ffn_down.weight Block 20 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q3_K
190 blk.20.ffn_gate.weight Block 20 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
191 blk.20.ffn_norm.weight Block 20 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
192 blk.20.ffn_up.weight Block 20 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
  • Total elements in blk.20: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 21 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
193 blk.21.attn_k.weight Block 21 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q2_K
194 blk.21.attn_norm.weight Block 21 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
195 blk.21.attn_output.weight Block 21 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
196 blk.21.attn_q.weight Block 21 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q2_K
197 blk.21.attn_v.weight Block 21 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
198 blk.21.ffn_down.weight Block 21 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q3_K
199 blk.21.ffn_gate.weight Block 21 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
200 blk.21.ffn_norm.weight Block 21 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
201 blk.21.ffn_up.weight Block 21 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
  • Total elements in blk.21: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 22 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
202 blk.22.attn_k.weight Block 22 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q2_K
203 blk.22.attn_norm.weight Block 22 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
204 blk.22.attn_output.weight Block 22 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
205 blk.22.attn_q.weight Block 22 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q2_K
206 blk.22.attn_v.weight Block 22 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
207 blk.22.ffn_down.weight Block 22 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q3_K
208 blk.22.ffn_gate.weight Block 22 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
209 blk.22.ffn_norm.weight Block 22 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
210 blk.22.ffn_up.weight Block 22 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
  • Total elements in blk.22: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 23 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
211 blk.23.attn_k.weight Block 23 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q2_K
212 blk.23.attn_norm.weight Block 23 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
213 blk.23.attn_output.weight Block 23 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
214 blk.23.attn_q.weight Block 23 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q2_K
215 blk.23.attn_v.weight Block 23 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
216 blk.23.ffn_down.weight Block 23 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q3_K
217 blk.23.ffn_gate.weight Block 23 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
218 blk.23.ffn_norm.weight Block 23 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
219 blk.23.ffn_up.weight Block 23 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
  • Total elements in blk.23: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 24 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
220 blk.24.attn_k.weight Block 24 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q2_K
221 blk.24.attn_norm.weight Block 24 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
222 blk.24.attn_output.weight Block 24 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
223 blk.24.attn_q.weight Block 24 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q2_K
224 blk.24.attn_v.weight Block 24 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
225 blk.24.ffn_down.weight Block 24 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q3_K
226 blk.24.ffn_gate.weight Block 24 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
227 blk.24.ffn_norm.weight Block 24 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
228 blk.24.ffn_up.weight Block 24 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
  • Total elements in blk.24: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 25 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
229 blk.25.attn_k.weight Block 25 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q2_K
230 blk.25.attn_norm.weight Block 25 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
231 blk.25.attn_output.weight Block 25 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
232 blk.25.attn_q.weight Block 25 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q2_K
233 blk.25.attn_v.weight Block 25 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
234 blk.25.ffn_down.weight Block 25 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q3_K
235 blk.25.ffn_gate.weight Block 25 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
236 blk.25.ffn_norm.weight Block 25 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
237 blk.25.ffn_up.weight Block 25 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
  • Total elements in blk.25: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 26 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
238 blk.26.attn_k.weight Block 26 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q2_K
239 blk.26.attn_norm.weight Block 26 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
240 blk.26.attn_output.weight Block 26 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
241 blk.26.attn_q.weight Block 26 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q2_K
242 blk.26.attn_v.weight Block 26 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
243 blk.26.ffn_down.weight Block 26 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q3_K
244 blk.26.ffn_gate.weight Block 26 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
245 blk.26.ffn_norm.weight Block 26 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
246 blk.26.ffn_up.weight Block 26 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
  • Total elements in blk.26: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 27 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
247 blk.27.attn_k.weight Block 27 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q2_K
248 blk.27.attn_norm.weight Block 27 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
249 blk.27.attn_output.weight Block 27 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
250 blk.27.attn_q.weight Block 27 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q2_K
251 blk.27.attn_v.weight Block 27 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
252 blk.27.ffn_down.weight Block 27 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q3_K
253 blk.27.ffn_gate.weight Block 27 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
254 blk.27.ffn_norm.weight Block 27 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
255 blk.27.ffn_up.weight Block 27 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
  • Total elements in blk.27: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 28 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
256 blk.28.attn_k.weight Block 28 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q2_K
257 blk.28.attn_norm.weight Block 28 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
258 blk.28.attn_output.weight Block 28 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
259 blk.28.attn_q.weight Block 28 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q2_K
260 blk.28.attn_v.weight Block 28 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
261 blk.28.ffn_down.weight Block 28 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q3_K
262 blk.28.ffn_gate.weight Block 28 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
263 blk.28.ffn_norm.weight Block 28 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
264 blk.28.ffn_up.weight Block 28 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
  • Total elements in blk.28: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 29 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
265 blk.29.attn_k.weight Block 29 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q2_K
266 blk.29.attn_norm.weight Block 29 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
267 blk.29.attn_output.weight Block 29 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
268 blk.29.attn_q.weight Block 29 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q2_K
269 blk.29.attn_v.weight Block 29 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
270 blk.29.ffn_down.weight Block 29 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q3_K
271 blk.29.ffn_gate.weight Block 29 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
272 blk.29.ffn_norm.weight Block 29 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
273 blk.29.ffn_up.weight Block 29 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
  • Total elements in blk.29: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 30 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
274 blk.30.attn_k.weight Block 30 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q2_K
275 blk.30.attn_norm.weight Block 30 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
276 blk.30.attn_output.weight Block 30 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
277 blk.30.attn_q.weight Block 30 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q2_K
278 blk.30.attn_v.weight Block 30 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
279 blk.30.ffn_down.weight Block 30 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q3_K
280 blk.30.ffn_gate.weight Block 30 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
281 blk.30.ffn_norm.weight Block 30 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
282 blk.30.ffn_up.weight Block 30 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
  • Total elements in blk.30: (~218M) 218112000
  • Percentage of total elements: 2.72%

Block 31 Tensor Group : ~218M Elements

T_ID Tensor Layer Name Human Friendly Tensor Layer Name Elements Shape Type
283 blk.31.attn_k.weight Block 31 Attention Key (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q2_K
284 blk.31.attn_norm.weight Block 31 Attention Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
285 blk.31.attn_output.weight Block 31 Attention Output (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q3_K
286 blk.31.attn_q.weight Block 31 Attention Query (W) (~17M) 16777216 4096 x 4096 x 1 x 1 Q2_K
287 blk.31.attn_v.weight Block 31 Attention Value (W) ( ~4M) 4194304 4096 x 1024 x 1 x 1 Q3_K
288 blk.31.ffn_down.weight Block 31 Feed-Forward Network "Down" (W) (~59M) 58720256 14336 x 4096 x 1 x 1 Q3_K
289 blk.31.ffn_gate.weight Block 31 Feed-Forward Network "Gate" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
290 blk.31.ffn_norm.weight Block 31 Feed-Forward Network Normalization (W) ( ~4K) 4096 4096 x 1 x 1 x 1 F32
291 blk.31.ffn_up.weight Block 31 Feed-Forward Network "Up" (W) (~59M) 58720256 4096 x 14336 x 1 x 1 Q2_K
  • Total elements in blk.31: (~218M) 218112000
  • Percentage of total elements: 2.72%