diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000000000000000000000000000000000000..52373fe24473b1aa44333d318f578ae6bf04b49b --- /dev/null +++ b/.gitattributes @@ -0,0 +1,36 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +tokenizer.json filter=lfs diff=lfs merge=lfs -text diff --git a/chat_template.jinja b/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..33089ace1be88f22a10fe861ad49718d5d886090 --- /dev/null +++ b/chat_template.jinja @@ -0,0 +1,109 @@ +{{- bos_token }} +{%- if custom_tools is defined %} + {%- set tools = custom_tools %} +{%- endif %} +{%- if not tools_in_user_message is defined %} + {%- set tools_in_user_message = true %} +{%- endif %} +{%- if not date_string is defined %} + {%- set date_string = "26 Jul 2024" %} +{%- endif %} +{%- if not tools is defined %} + {%- set tools = none %} +{%- endif %} + +{#- This block extracts the system message, so we can slot it into the right place. #} +{%- if messages[0]['role'] == 'system' %} + {%- set system_message = messages[0]['content']|trim %} + {%- set messages = messages[1:] %} +{%- else %} + {%- set system_message = "" %} +{%- endif %} + +{#- System message + builtin tools #} +{{- "<|start_header_id|>system<|end_header_id|>\n\n" }} +{%- if builtin_tools is defined or tools is not none %} + {{- "Environment: ipython\n" }} +{%- endif %} +{%- if builtin_tools is defined %} + {{- "Tools: " + builtin_tools | reject('equalto', 'code_interpreter') | join(", ") + "\n\n"}} +{%- endif %} +{{- "Cutting Knowledge Date: December 2023\n" }} +{{- "Today Date: " + date_string + "\n\n" }} +{%- if tools is not none and not tools_in_user_message %} + {{- "You have access to the following functions. To call a function, please respond with JSON for a function call." }} + {{- 'Respond in the format {"name": function name, "parameters": dictionary of argument name and its value}.' }} + {{- "Do not use variables.\n\n" }} + {%- for t in tools %} + {{- t | tojson(indent=4) }} + {{- "\n\n" }} + {%- endfor %} +{%- endif %} +{{- system_message }} +{{- "<|eot_id|>" }} + +{#- Custom tools are passed in a user message with some extra guidance #} +{%- if tools_in_user_message and not tools is none %} + {#- Extract the first user message so we can plug it in here #} + {%- if messages | length != 0 %} + {%- set first_user_message = messages[0]['content']|trim %} + {%- set messages = messages[1:] %} + {%- else %} + {{- raise_exception("Cannot put tools in the first user message when there's no first user message!") }} +{%- endif %} + {{- '<|start_header_id|>user<|end_header_id|>\n\n' -}} + {{- "Given the following functions, please respond with a JSON for a function call " }} + {{- "with its proper arguments that best answers the given prompt.\n\n" }} + {{- 'Respond in the format {"name": function name, "parameters": dictionary of argument name and its value}.' }} + {{- "Do not use variables.\n\n" }} + {%- for t in tools %} + {{- t | tojson(indent=4) }} + {{- "\n\n" }} + {%- endfor %} + {{- first_user_message + "<|eot_id|>"}} +{%- endif %} + +{%- for message in messages %} + {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %} + {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' }} + {%- elif 'tool_calls' in message %} + {%- if not message.tool_calls|length == 1 %} + {{- raise_exception("This model only supports single tool-calls at once!") }} + {%- endif %} + {%- set tool_call = message.tool_calls[0].function %} + {%- if builtin_tools is defined and tool_call.name in builtin_tools %} + {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' -}} + {{- "<|python_tag|>" + tool_call.name + ".call(" }} + {%- for arg_name, arg_val in tool_call.arguments | items %} + {{- arg_name + '="' + arg_val + '"' }} + {%- if not loop.last %} + {{- ", " }} + {%- endif %} + {%- endfor %} + {{- ")" }} + {%- else %} + {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' -}} + {{- '{"name": "' + tool_call.name + '", ' }} + {{- '"parameters": ' }} + {{- tool_call.arguments | tojson }} + {{- "}" }} + {%- endif %} + {%- if builtin_tools is defined %} + {#- This means we're in ipython mode #} + {{- "<|eom_id|>" }} + {%- else %} + {{- "<|eot_id|>" }} + {%- endif %} + {%- elif message.role == "tool" or message.role == "ipython" %} + {{- "<|start_header_id|>ipython<|end_header_id|>\n\n" }} + {%- if message.content is mapping or message.content is iterable %} + {{- message.content | tojson }} + {%- else %} + {{- message.content }} + {%- endif %} + {{- "<|eot_id|>" }} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' }} +{%- endif %} diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..57f61315d6ee3cd9802bfa06f4fce532134c30ec --- /dev/null +++ b/config.json @@ -0,0 +1,61 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 128000, + "eos_token_id": [ + 128001, + 128008, + 128009 + ], + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 16384, + "initializer_range": 0.02, + "intermediate_size": 53248, + "max_position_embeddings": 131072, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 128, + "num_hidden_layers": 126, + "num_key_value_heads": 8, + "pad_token_id": 128009, + "pretraining_tp": 1, + "quantization_config": { + "bits": 4, + "checkpoint_format": "gptq", + "desc_act": true, + "group_size": 128, + "lm_head": false, + "meta": { + "damp_auto_increment": 0.0025, + "damp_percent": 0.01, + "mse": 0.0, + "quantizer": [ + "gptqmodel:2.2.0" + ], + "static_groups": false, + "true_sequential": true, + "uri": "https://github.com/modelcloud/gptqmodel" + }, + "pack_dtype": "int32", + "quant_method": "gptq", + "sym": true + }, + "rms_norm_eps": 1e-05, + "rope_scaling": { + "factor": 8.0, + "high_freq_factor": 4.0, + "low_freq_factor": 1.0, + "original_max_position_embeddings": 8192, + "rope_type": "llama3" + }, + "rope_theta": 500000.0, + "tie_word_embeddings": false, + "torch_dtype": "float16", + "transformers_version": "4.52.3", + "use_cache": false, + "vocab_size": 128256 +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..329d2aad7a332f09a4a3a035e5eb1f37a8ae84a2 --- /dev/null +++ b/generation_config.json @@ -0,0 +1,12 @@ +{ + "bos_token_id": 128000, + "do_sample": true, + "eos_token_id": [ + 128001, + 128008, + 128009 + ], + "temperature": 0.6, + "top_p": 0.9, + "transformers_version": "4.52.3" +} diff --git a/model-00001-of-00057.safetensors b/model-00001-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9de4ad533c9798ae03fa76eee9ed81d9d4e97409 --- /dev/null +++ b/model-00001-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7624c458c474b56d12517e688dc99e949d4f9a82b9bcce7f354d832fbb618bce +size 4202692752 diff --git a/model-00002-of-00057.safetensors b/model-00002-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b21c6e8dfae0936a8fde0df3802593468e3c0570 --- /dev/null +++ b/model-00002-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2273109ebd9c32da07c5896a82fc84ed2d49f0b59e88020ada806f8566034b90 +size 3610157512 diff --git a/model-00003-of-00057.safetensors b/model-00003-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f957e5b29c41bfea1111dba328caa730def5cde8 --- /dev/null +++ b/model-00003-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69545b23a8d00f9d7f310a30366bfb4b7b587d68c37c4b755af8902d62244d58 +size 3766852704 diff --git a/model-00004-of-00057.safetensors b/model-00004-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5b643ebc1a6f84eaa5d5f7f91ff2b34c7bd588e2 --- /dev/null +++ b/model-00004-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e4fbcca1a82ca89d2fca583b8aa795b828b2b4240bc79b75108394ed3f4bbbf1 +size 3766852696 diff --git a/model-00005-of-00057.safetensors b/model-00005-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..804daa40ee86556d1ef8290fe696c584d0ac623b --- /dev/null +++ b/model-00005-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aaabe63fa37db61d7077ea3a6bfdab7c2f30bbe490cd878dca69efb41625b4cd +size 3924157096 diff --git a/model-00006-of-00057.safetensors b/model-00006-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..aff95790e48807c824603f9ccd6da9e73d6addb7 --- /dev/null +++ b/model-00006-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35a95f6529eb6e8292f1e70aecd3a84cf1adc5c78afb1e60e1b6c2d50b232228 +size 3906379360 diff --git a/model-00007-of-00057.safetensors b/model-00007-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5c9a5a975eda29dcfa7483921bd866aa5aa9f26e --- /dev/null +++ b/model-00007-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ffe19d6348bb3a292f88ddcd2ceb3d4137d47ab35a212fe968c857fa046b9fff +size 3766852760 diff --git a/model-00008-of-00057.safetensors b/model-00008-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..036bf66cecc8d20b2ceb169b60909d68753feb45 --- /dev/null +++ b/model-00008-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f711a4c2cf3309c26244db0aa00037dccafc46e17749d671105b54e3596b7997 +size 3924157168 diff --git a/model-00009-of-00057.safetensors b/model-00009-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..52fbdbe4bae60ae356cbb5d92d332229521c61ca --- /dev/null +++ b/model-00009-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1dafa31e5fab861e6c827758a6399432368549852c5f2408767ca1f0851bcbe9 +size 3906379376 diff --git a/model-00010-of-00057.safetensors b/model-00010-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7d0e007c301424ad0fad75985b273a0e39fe372f --- /dev/null +++ b/model-00010-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5f661329ba429c12498a9222f4eae8cbe965a19d660370f090e65e70df3badb +size 3766852760 diff --git a/model-00011-of-00057.safetensors b/model-00011-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..996893367585899adb8e427b5162802aab9b6ebc --- /dev/null +++ b/model-00011-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ffef6133eca40965ca7b42213ea9c3f778a42b1ebc9ac6b7dbcf3578514d9764 +size 3924157168 diff --git a/model-00012-of-00057.safetensors b/model-00012-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..370ee6b1140fe279f8530f1ee61e95015e9647e2 --- /dev/null +++ b/model-00012-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:61e072defe939da4598732701a1e3e68892b71a132ef554846172d341d53fbc7 +size 3906379376 diff --git a/model-00013-of-00057.safetensors b/model-00013-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4265d3616a0fd36c39b88b4ced3efbf53a04c1d0 --- /dev/null +++ b/model-00013-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9b0fca69d12f2181d015de7f6b2590b938f30c9c12d5e734e41ec969d5af962a +size 3766852760 diff --git a/model-00014-of-00057.safetensors b/model-00014-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bdb4440d52ad200ddde83fbd40b14eb0ef20973c --- /dev/null +++ b/model-00014-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d94ff5866762e9ddc5f2f1c10636a07e21cc263ece1d32f9a53c8fa8f2fffcfd +size 3924157168 diff --git a/model-00015-of-00057.safetensors b/model-00015-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..30ae6223d2fee486be1e92c058d6dce7071d7823 --- /dev/null +++ b/model-00015-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71876de1934b2cefa9daffaabd32693ce89b0c379ab23d3b84465dad0f6f7851 +size 3906379376 diff --git a/model-00016-of-00057.safetensors b/model-00016-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1c42e5a7b6818930fb174339a858e0bec42c7c5c --- /dev/null +++ b/model-00016-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a67cd828f99eb67dc22cc6935a4340cce8134cbbf70e2ecc581387d90258ef04 +size 3766852760 diff --git a/model-00017-of-00057.safetensors b/model-00017-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0c518c109d6eefb228d0c6ecff5df1d5ee616286 --- /dev/null +++ b/model-00017-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9b4a54c97ef55f0fb236ab73e096a8a19525d4c646f304b82a6799898aa97043 +size 3924157168 diff --git a/model-00018-of-00057.safetensors b/model-00018-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2bef0eb1bf2a62f1a488811a0e46b296f791131c --- /dev/null +++ b/model-00018-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1a551605f4c687713d9f34ee5b1d4fec3720f926e3a49fede27f7ff16d4a9cd7 +size 3906379376 diff --git a/model-00019-of-00057.safetensors b/model-00019-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..162ec7ab1a49002f35d579f811aeb650b560f57d --- /dev/null +++ b/model-00019-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3f4b4de3b09ecbf6881e7a2beb993478cec3a8ae66a7b63c061c3b4ba16a5d0f +size 3766852760 diff --git a/model-00020-of-00057.safetensors b/model-00020-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..456add052fbc92576b1187b46b36264fb9cf78b3 --- /dev/null +++ b/model-00020-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f176a35288a98a68b2c1bc7d9ac72d286191e54d36ef7e7bd90151977e3b503c +size 3924157168 diff --git a/model-00021-of-00057.safetensors b/model-00021-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..eeeb1fc3abab34581015f1395e3bca0a592ab3cc --- /dev/null +++ b/model-00021-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:999f85c7bfb3249856f7cc120634a5479a52abcb44c8fe679a4aaeba07329e95 +size 3906379376 diff --git a/model-00022-of-00057.safetensors b/model-00022-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cc4fddcc2cff75b5f04030c9a7bacaa8b2f15ef1 --- /dev/null +++ b/model-00022-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:716b2898a85266f32cd950f6841f7860a02dab5f42b5989c65a7bbf6e9649530 +size 3766852760 diff --git a/model-00023-of-00057.safetensors b/model-00023-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bd7534a374a5f63660cc238a273175cf520ef0f7 --- /dev/null +++ b/model-00023-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a1464c3f203cc31460c75eed343201520aa9c4c1a5baadbd9e41d12e0f34b253 +size 3924157168 diff --git a/model-00024-of-00057.safetensors b/model-00024-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..25aa8f1091f5ac6a2fb1308736d9590645502a30 --- /dev/null +++ b/model-00024-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a7549bde4e58ea3182553ae621d6e5599ef4c8f5332d39c3a712e0c5330df1e +size 3906379376 diff --git a/model-00025-of-00057.safetensors b/model-00025-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0577d27ee55a97ee36bcc8e5942ad58f7cf0ff81 --- /dev/null +++ b/model-00025-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c57f5b8f5b12a43113fcf16fce5421c62d7e166fcb98eecc67ebe8e20bfced4e +size 3766852760 diff --git a/model-00026-of-00057.safetensors b/model-00026-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7cc9453e36e80a8dbb235ee064b0d90a4a094040 --- /dev/null +++ b/model-00026-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e5ac7fc81c65d701421a8d22f127aad3f1698100f4c080851eac31ae1e80d58 +size 3924157168 diff --git a/model-00027-of-00057.safetensors b/model-00027-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7962697a21f3eec902f936cd87b36e3cb01c9112 --- /dev/null +++ b/model-00027-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c754ac23b4f6bededdac2ff14c96e2d57f0ad5cf94fb49662a3ea6fd62c4d91 +size 3906379376 diff --git a/model-00028-of-00057.safetensors b/model-00028-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..20526f3f32adfa7bbb1da7f2df8d6c4a1a1006ab --- /dev/null +++ b/model-00028-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:394f940704410a8e66294ec25e5e6ee8015f4562e0000d91f6b4b3aef832ba5f +size 3766852760 diff --git a/model-00029-of-00057.safetensors b/model-00029-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dd8b1db604d1893983c7896d3dbadf6b004453fe --- /dev/null +++ b/model-00029-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:986276a4d8da96eac622bff6da02a59fc58e86012f773aab22cff383bc4475bf +size 3924157168 diff --git a/model-00030-of-00057.safetensors b/model-00030-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d8fe8553056c9348ce1766d1eec52f9969640349 --- /dev/null +++ b/model-00030-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1bddbf3292de7f20bc250d1c1aba5c69136fb03cdfe6c6620aeee97978e70332 +size 3906379376 diff --git a/model-00031-of-00057.safetensors b/model-00031-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7194413d1ef66063bb6a2650b4a4ee6ad83bf5b9 --- /dev/null +++ b/model-00031-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d37b2d013cb7335f524f34a58bd29998cc4fc127d38b1d1bd3b4d7830c965646 +size 3766852760 diff --git a/model-00032-of-00057.safetensors b/model-00032-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f7d5745bc6f49f8404460e7143dae01fadb70a37 --- /dev/null +++ b/model-00032-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eb9754f361a12da03cf1fb730f44be80f2f0584961b4b47303b4bcccafb38323 +size 3924157168 diff --git a/model-00033-of-00057.safetensors b/model-00033-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dcd82addec3e62646058cd9dc392aa10a12db280 --- /dev/null +++ b/model-00033-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bb6ca5429c799ff298a703c866abbc66919de3ff597d59599bf5b29edb35227b +size 3906379376 diff --git a/model-00034-of-00057.safetensors b/model-00034-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..929486400b85da94c732c630e429a1d6248f91aa --- /dev/null +++ b/model-00034-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f41fe986bf5090094bbcee1d7d5f9bda2cf51f569bdf6703658c0384c89e63fd +size 3766852760 diff --git a/model-00035-of-00057.safetensors b/model-00035-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cd2535da93a4a0c310fe7eecbf4168c853f7220f --- /dev/null +++ b/model-00035-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69b7310c8d8da7f63751ee548ec1ade8d9e1438929087aa884e82b7796293168 +size 3924157168 diff --git a/model-00036-of-00057.safetensors b/model-00036-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f5718fcfac6f2ff42db7dda9ffb388f723e83857 --- /dev/null +++ b/model-00036-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ce156be7652fda3125c3af4471a0d6f894dbec82a7b2eb367315a02b80cdabc +size 3906379376 diff --git a/model-00037-of-00057.safetensors b/model-00037-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b6c17ba099b73517a544834cbfd45e61e28b27eb --- /dev/null +++ b/model-00037-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7874dbad8f3137d75dfd60127f23a2f7c140e362ac6007dd8bbe8b5396f24677 +size 3766852760 diff --git a/model-00038-of-00057.safetensors b/model-00038-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d2d6c0f7e2dcec855e2cc57861e7aa45a4a21eab --- /dev/null +++ b/model-00038-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:34fff1444be87828a8631a44da7b4ff42e91e6d01faf7ae59c41e03d641d3138 +size 3924157168 diff --git a/model-00039-of-00057.safetensors b/model-00039-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..48568e84d38af559609539424056d76079f569aa --- /dev/null +++ b/model-00039-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:09b9603e2cc096803d7667c28f9ee1f96a63bc29d0e22faab314de3cb38892b5 +size 3906379376 diff --git a/model-00040-of-00057.safetensors b/model-00040-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..302ca6f42f4a9e5b611ab11317f906c225c3253c --- /dev/null +++ b/model-00040-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:97f73fd88f37eb3e9db30b9627185ba8760a14f824797af038f81ffe75359e88 +size 3766852760 diff --git a/model-00041-of-00057.safetensors b/model-00041-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..77d70a301642d10e42387f204a5c39ffc4a489fa --- /dev/null +++ b/model-00041-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:489e5f6f31f540213fe21e8307e77aa4418b8cecb4eb1497e63f71ce631add91 +size 3924157168 diff --git a/model-00042-of-00057.safetensors b/model-00042-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..da7d1c2d70c89d29280e79725bb11acaf9b2089b --- /dev/null +++ b/model-00042-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f0f65690c49179c1376e98030cf1816c5ef612a9413e2697207229cc28b63204 +size 3906379376 diff --git a/model-00043-of-00057.safetensors b/model-00043-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b00ef36db07d039d2031f4267d2c138ae4b9e0a6 --- /dev/null +++ b/model-00043-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9131b19fd4bcd5b711d4261d326d7ffad2583f47614ab60ec4308938a5fa8d28 +size 3766852760 diff --git a/model-00044-of-00057.safetensors b/model-00044-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a049f98610d448bd82fb24868e699945f539c0bf --- /dev/null +++ b/model-00044-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:310e7f7a2742f63294d8cfa3d7b28392354b7e2c6103ef1e0c97744a102bd8e3 +size 3924157160 diff --git a/model-00045-of-00057.safetensors b/model-00045-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..33cf2db7d81915919bcc494947e6a1439caedc2e --- /dev/null +++ b/model-00045-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a46f9c4cac9404020d6ff75f21a08ee333d96f59ac21662c9af6aaf6d10f26aa +size 3906379448 diff --git a/model-00046-of-00057.safetensors b/model-00046-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..eb248df10602a0d2c33cb55687cd29187d20542f --- /dev/null +++ b/model-00046-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:113193c3f0939f49a2ec944072a6ad1ac688d70b279ffc02ec09b552993f7170 +size 3766852824 diff --git a/model-00047-of-00057.safetensors b/model-00047-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..54a531120f790ab92ec4af5e832df140ec2d445f --- /dev/null +++ b/model-00047-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce45ac145534108756de73af80ca7b6022bed18415ce553e2d2a2bb37b3f01a9 +size 3924157248 diff --git a/model-00048-of-00057.safetensors b/model-00048-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..09330090b666a84cd06a0bece45370093411f240 --- /dev/null +++ b/model-00048-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f6926bd789c284af0d8995410b7d7e14ac6b702efb00d371b5d67413b36f157a +size 3906379448 diff --git a/model-00049-of-00057.safetensors b/model-00049-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0eac240d245d136143fd215cbe84a62a4f2e4af5 --- /dev/null +++ b/model-00049-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b937ec74f62da33830d6486ad17447802521dc3c7eb797f7f77ad8369bcb3cae +size 3766852824 diff --git a/model-00050-of-00057.safetensors b/model-00050-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..59390c8b52c7021c21b7f1a458fc1ec3e83a67d1 --- /dev/null +++ b/model-00050-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db263caad50d1ed45c29fea3b5fc33b2d93af9b5ba12cd894afab6e4f72dc7f1 +size 3924157248 diff --git a/model-00051-of-00057.safetensors b/model-00051-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..89f387367fd4a5650eed2acb4c3d107435ccee33 --- /dev/null +++ b/model-00051-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eeba0611c5416823a12c09551ada31ea43fe794458aa1d0f02c07a301983ee15 +size 3906379448 diff --git a/model-00052-of-00057.safetensors b/model-00052-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b38838e4cb83ab6d1145e61ce753596d4a0ba348 --- /dev/null +++ b/model-00052-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a4e71d47890c66ee38d112f3caf407aa504f496c4d4d33a67991c7e8c140ce8a +size 3766852824 diff --git a/model-00053-of-00057.safetensors b/model-00053-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3b92a71bd2d007a465321e944474aa8268c581b8 --- /dev/null +++ b/model-00053-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b5a28f5bacdd20c4bb16c2476f02de2afd9fe1e61708e1d29b6cbb90d50a142 +size 3924157248 diff --git a/model-00054-of-00057.safetensors b/model-00054-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5885d0f10191a8b6ed7e2515041384a360c61f1c --- /dev/null +++ b/model-00054-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1adcb56947c94190a7fb6dba1e12fa47f75465b0fce8e8d68c6a1205b173c80d +size 3906379448 diff --git a/model-00055-of-00057.safetensors b/model-00055-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2a22e16dc151b14e8c5bf2f8a6f61d630de78fd2 --- /dev/null +++ b/model-00055-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:96157919c5b9d697cfb8bbd0faa29e5dae6cee256036d4d76bda46a7437a1701 +size 3766852824 diff --git a/model-00056-of-00057.safetensors b/model-00056-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cd7191663ae9f8ef893099763fbc6962bca0a2a7 --- /dev/null +++ b/model-00056-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1cd6e5bf6b00c9124959d1df923a46124d5006ddd9f82db91d0fdd06e998241c +size 4202692736 diff --git a/model-00057-of-00057.safetensors b/model-00057-of-00057.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..78e90e95677aa2f0a9b88958d461886d8c6b8222 --- /dev/null +++ b/model-00057-of-00057.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:49f094ea421914f2b0279ddd7fe1a71a3d6a12bb756e5902e3d284b925d083f9 +size 453559072 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..cb1d02fd1b485e88e53b65bb78b8afb718aa2820 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,3790 @@ +{ + "metadata": { + "total_size": 217157992448 + }, + "weight_map": { + "lm_head.weight": "model-00056-of-00057.safetensors", + "model.embed_tokens.weight": "model-00001-of-00057.safetensors", + "model.layers.0.input_layernorm.weight": "model-00002-of-00057.safetensors", + "model.layers.0.mlp.down_proj.g_idx": "model-00002-of-00057.safetensors", + "model.layers.0.mlp.down_proj.qweight": "model-00002-of-00057.safetensors", + "model.layers.0.mlp.down_proj.qzeros": "model-00002-of-00057.safetensors", + "model.layers.0.mlp.down_proj.scales": "model-00002-of-00057.safetensors", + "model.layers.0.mlp.gate_proj.g_idx": "model-00002-of-00057.safetensors", + "model.layers.0.mlp.gate_proj.qweight": "model-00002-of-00057.safetensors", + "model.layers.0.mlp.gate_proj.qzeros": "model-00002-of-00057.safetensors", + "model.layers.0.mlp.gate_proj.scales": "model-00002-of-00057.safetensors", + "model.layers.0.mlp.up_proj.g_idx": "model-00002-of-00057.safetensors", + "model.layers.0.mlp.up_proj.qweight": "model-00002-of-00057.safetensors", + "model.layers.0.mlp.up_proj.qzeros": "model-00002-of-00057.safetensors", + "model.layers.0.mlp.up_proj.scales": "model-00002-of-00057.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00002-of-00057.safetensors", + "model.layers.0.self_attn.k_proj.g_idx": "model-00002-of-00057.safetensors", + "model.layers.0.self_attn.k_proj.qweight": "model-00002-of-00057.safetensors", + "model.layers.0.self_attn.k_proj.qzeros": "model-00002-of-00057.safetensors", + "model.layers.0.self_attn.k_proj.scales": "model-00002-of-00057.safetensors", + "model.layers.0.self_attn.o_proj.g_idx": "model-00002-of-00057.safetensors", + "model.layers.0.self_attn.o_proj.qweight": "model-00002-of-00057.safetensors", + "model.layers.0.self_attn.o_proj.qzeros": "model-00002-of-00057.safetensors", + "model.layers.0.self_attn.o_proj.scales": "model-00002-of-00057.safetensors", + "model.layers.0.self_attn.q_proj.g_idx": "model-00002-of-00057.safetensors", + "model.layers.0.self_attn.q_proj.qweight": "model-00002-of-00057.safetensors", + "model.layers.0.self_attn.q_proj.qzeros": "model-00002-of-00057.safetensors", + "model.layers.0.self_attn.q_proj.scales": "model-00002-of-00057.safetensors", + "model.layers.0.self_attn.v_proj.g_idx": "model-00002-of-00057.safetensors", + "model.layers.0.self_attn.v_proj.qweight": "model-00002-of-00057.safetensors", + "model.layers.0.self_attn.v_proj.qzeros": "model-00002-of-00057.safetensors", + "model.layers.0.self_attn.v_proj.scales": "model-00002-of-00057.safetensors", + "model.layers.1.input_layernorm.weight": "model-00002-of-00057.safetensors", + "model.layers.1.mlp.down_proj.g_idx": "model-00002-of-00057.safetensors", + "model.layers.1.mlp.down_proj.qweight": "model-00002-of-00057.safetensors", + "model.layers.1.mlp.down_proj.qzeros": "model-00002-of-00057.safetensors", + "model.layers.1.mlp.down_proj.scales": "model-00002-of-00057.safetensors", + "model.layers.1.mlp.gate_proj.g_idx": "model-00002-of-00057.safetensors", + "model.layers.1.mlp.gate_proj.qweight": "model-00002-of-00057.safetensors", + "model.layers.1.mlp.gate_proj.qzeros": "model-00002-of-00057.safetensors", + "model.layers.1.mlp.gate_proj.scales": "model-00002-of-00057.safetensors", + "model.layers.1.mlp.up_proj.g_idx": "model-00002-of-00057.safetensors", + "model.layers.1.mlp.up_proj.qweight": "model-00002-of-00057.safetensors", + "model.layers.1.mlp.up_proj.qzeros": "model-00002-of-00057.safetensors", + "model.layers.1.mlp.up_proj.scales": "model-00002-of-00057.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00002-of-00057.safetensors", + "model.layers.1.self_attn.k_proj.g_idx": "model-00002-of-00057.safetensors", + "model.layers.1.self_attn.k_proj.qweight": "model-00002-of-00057.safetensors", + "model.layers.1.self_attn.k_proj.qzeros": "model-00002-of-00057.safetensors", + "model.layers.1.self_attn.k_proj.scales": "model-00002-of-00057.safetensors", + "model.layers.1.self_attn.o_proj.g_idx": "model-00002-of-00057.safetensors", + "model.layers.1.self_attn.o_proj.qweight": "model-00002-of-00057.safetensors", + "model.layers.1.self_attn.o_proj.qzeros": "model-00002-of-00057.safetensors", + "model.layers.1.self_attn.o_proj.scales": "model-00002-of-00057.safetensors", + "model.layers.1.self_attn.q_proj.g_idx": "model-00002-of-00057.safetensors", + "model.layers.1.self_attn.q_proj.qweight": "model-00002-of-00057.safetensors", + "model.layers.1.self_attn.q_proj.qzeros": "model-00002-of-00057.safetensors", + "model.layers.1.self_attn.q_proj.scales": "model-00002-of-00057.safetensors", + "model.layers.1.self_attn.v_proj.g_idx": "model-00002-of-00057.safetensors", + "model.layers.1.self_attn.v_proj.qweight": "model-00002-of-00057.safetensors", + "model.layers.1.self_attn.v_proj.qzeros": "model-00002-of-00057.safetensors", + "model.layers.1.self_attn.v_proj.scales": "model-00002-of-00057.safetensors", + "model.layers.10.input_layernorm.weight": "model-00006-of-00057.safetensors", + "model.layers.10.mlp.down_proj.g_idx": "model-00006-of-00057.safetensors", + "model.layers.10.mlp.down_proj.qweight": "model-00006-of-00057.safetensors", + "model.layers.10.mlp.down_proj.qzeros": "model-00006-of-00057.safetensors", + "model.layers.10.mlp.down_proj.scales": "model-00006-of-00057.safetensors", + "model.layers.10.mlp.gate_proj.g_idx": "model-00006-of-00057.safetensors", + "model.layers.10.mlp.gate_proj.qweight": "model-00006-of-00057.safetensors", + "model.layers.10.mlp.gate_proj.qzeros": "model-00006-of-00057.safetensors", + "model.layers.10.mlp.gate_proj.scales": "model-00006-of-00057.safetensors", + "model.layers.10.mlp.up_proj.g_idx": "model-00006-of-00057.safetensors", + "model.layers.10.mlp.up_proj.qweight": "model-00006-of-00057.safetensors", + "model.layers.10.mlp.up_proj.qzeros": "model-00006-of-00057.safetensors", + "model.layers.10.mlp.up_proj.scales": "model-00006-of-00057.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00006-of-00057.safetensors", + "model.layers.10.self_attn.k_proj.g_idx": "model-00006-of-00057.safetensors", + "model.layers.10.self_attn.k_proj.qweight": "model-00006-of-00057.safetensors", + "model.layers.10.self_attn.k_proj.qzeros": "model-00006-of-00057.safetensors", + "model.layers.10.self_attn.k_proj.scales": "model-00006-of-00057.safetensors", + "model.layers.10.self_attn.o_proj.g_idx": "model-00006-of-00057.safetensors", + "model.layers.10.self_attn.o_proj.qweight": "model-00006-of-00057.safetensors", + "model.layers.10.self_attn.o_proj.qzeros": "model-00006-of-00057.safetensors", + "model.layers.10.self_attn.o_proj.scales": "model-00006-of-00057.safetensors", + "model.layers.10.self_attn.q_proj.g_idx": "model-00006-of-00057.safetensors", + "model.layers.10.self_attn.q_proj.qweight": "model-00006-of-00057.safetensors", + "model.layers.10.self_attn.q_proj.qzeros": "model-00006-of-00057.safetensors", + "model.layers.10.self_attn.q_proj.scales": "model-00006-of-00057.safetensors", + "model.layers.10.self_attn.v_proj.g_idx": "model-00006-of-00057.safetensors", + "model.layers.10.self_attn.v_proj.qweight": "model-00006-of-00057.safetensors", + "model.layers.10.self_attn.v_proj.qzeros": "model-00006-of-00057.safetensors", + "model.layers.10.self_attn.v_proj.scales": "model-00006-of-00057.safetensors", + "model.layers.100.input_layernorm.weight": "model-00045-of-00057.safetensors", + "model.layers.100.mlp.down_proj.g_idx": "model-00045-of-00057.safetensors", + "model.layers.100.mlp.down_proj.qweight": "model-00045-of-00057.safetensors", + "model.layers.100.mlp.down_proj.qzeros": "model-00045-of-00057.safetensors", + "model.layers.100.mlp.down_proj.scales": "model-00045-of-00057.safetensors", + "model.layers.100.mlp.gate_proj.g_idx": "model-00045-of-00057.safetensors", + "model.layers.100.mlp.gate_proj.qweight": "model-00045-of-00057.safetensors", + "model.layers.100.mlp.gate_proj.qzeros": "model-00045-of-00057.safetensors", + "model.layers.100.mlp.gate_proj.scales": "model-00045-of-00057.safetensors", + "model.layers.100.mlp.up_proj.g_idx": "model-00045-of-00057.safetensors", + "model.layers.100.mlp.up_proj.qweight": "model-00045-of-00057.safetensors", + "model.layers.100.mlp.up_proj.qzeros": "model-00045-of-00057.safetensors", + "model.layers.100.mlp.up_proj.scales": "model-00045-of-00057.safetensors", + "model.layers.100.post_attention_layernorm.weight": "model-00045-of-00057.safetensors", + "model.layers.100.self_attn.k_proj.g_idx": "model-00044-of-00057.safetensors", + "model.layers.100.self_attn.k_proj.qweight": "model-00044-of-00057.safetensors", + "model.layers.100.self_attn.k_proj.qzeros": "model-00044-of-00057.safetensors", + "model.layers.100.self_attn.k_proj.scales": "model-00044-of-00057.safetensors", + "model.layers.100.self_attn.o_proj.g_idx": "model-00045-of-00057.safetensors", + "model.layers.100.self_attn.o_proj.qweight": "model-00045-of-00057.safetensors", + "model.layers.100.self_attn.o_proj.qzeros": "model-00045-of-00057.safetensors", + "model.layers.100.self_attn.o_proj.scales": "model-00045-of-00057.safetensors", + "model.layers.100.self_attn.q_proj.g_idx": "model-00044-of-00057.safetensors", + "model.layers.100.self_attn.q_proj.qweight": "model-00044-of-00057.safetensors", + "model.layers.100.self_attn.q_proj.qzeros": "model-00044-of-00057.safetensors", + "model.layers.100.self_attn.q_proj.scales": "model-00044-of-00057.safetensors", + "model.layers.100.self_attn.v_proj.g_idx": "model-00044-of-00057.safetensors", + "model.layers.100.self_attn.v_proj.qweight": "model-00044-of-00057.safetensors", + "model.layers.100.self_attn.v_proj.qzeros": "model-00044-of-00057.safetensors", + "model.layers.100.self_attn.v_proj.scales": "model-00044-of-00057.safetensors", + "model.layers.101.input_layernorm.weight": "model-00045-of-00057.safetensors", + "model.layers.101.mlp.down_proj.g_idx": "model-00045-of-00057.safetensors", + "model.layers.101.mlp.down_proj.qweight": "model-00045-of-00057.safetensors", + "model.layers.101.mlp.down_proj.qzeros": "model-00045-of-00057.safetensors", + "model.layers.101.mlp.down_proj.scales": "model-00045-of-00057.safetensors", + "model.layers.101.mlp.gate_proj.g_idx": "model-00045-of-00057.safetensors", + "model.layers.101.mlp.gate_proj.qweight": "model-00045-of-00057.safetensors", + "model.layers.101.mlp.gate_proj.qzeros": "model-00045-of-00057.safetensors", + "model.layers.101.mlp.gate_proj.scales": "model-00045-of-00057.safetensors", + "model.layers.101.mlp.up_proj.g_idx": "model-00045-of-00057.safetensors", + "model.layers.101.mlp.up_proj.qweight": "model-00045-of-00057.safetensors", + "model.layers.101.mlp.up_proj.qzeros": "model-00045-of-00057.safetensors", + "model.layers.101.mlp.up_proj.scales": "model-00045-of-00057.safetensors", + "model.layers.101.post_attention_layernorm.weight": "model-00045-of-00057.safetensors", + "model.layers.101.self_attn.k_proj.g_idx": "model-00045-of-00057.safetensors", + "model.layers.101.self_attn.k_proj.qweight": "model-00045-of-00057.safetensors", + "model.layers.101.self_attn.k_proj.qzeros": "model-00045-of-00057.safetensors", + "model.layers.101.self_attn.k_proj.scales": "model-00045-of-00057.safetensors", + "model.layers.101.self_attn.o_proj.g_idx": "model-00045-of-00057.safetensors", + "model.layers.101.self_attn.o_proj.qweight": "model-00045-of-00057.safetensors", + "model.layers.101.self_attn.o_proj.qzeros": "model-00045-of-00057.safetensors", + "model.layers.101.self_attn.o_proj.scales": "model-00045-of-00057.safetensors", + "model.layers.101.self_attn.q_proj.g_idx": "model-00045-of-00057.safetensors", + "model.layers.101.self_attn.q_proj.qweight": "model-00045-of-00057.safetensors", + "model.layers.101.self_attn.q_proj.qzeros": "model-00045-of-00057.safetensors", + "model.layers.101.self_attn.q_proj.scales": "model-00045-of-00057.safetensors", + "model.layers.101.self_attn.v_proj.g_idx": "model-00045-of-00057.safetensors", + "model.layers.101.self_attn.v_proj.qweight": "model-00045-of-00057.safetensors", + "model.layers.101.self_attn.v_proj.qzeros": "model-00045-of-00057.safetensors", + "model.layers.101.self_attn.v_proj.scales": "model-00045-of-00057.safetensors", + "model.layers.102.input_layernorm.weight": "model-00046-of-00057.safetensors", + "model.layers.102.mlp.down_proj.g_idx": "model-00046-of-00057.safetensors", + "model.layers.102.mlp.down_proj.qweight": "model-00046-of-00057.safetensors", + "model.layers.102.mlp.down_proj.qzeros": "model-00046-of-00057.safetensors", + "model.layers.102.mlp.down_proj.scales": "model-00046-of-00057.safetensors", + "model.layers.102.mlp.gate_proj.g_idx": "model-00045-of-00057.safetensors", + "model.layers.102.mlp.gate_proj.qweight": "model-00045-of-00057.safetensors", + "model.layers.102.mlp.gate_proj.qzeros": "model-00045-of-00057.safetensors", + "model.layers.102.mlp.gate_proj.scales": "model-00045-of-00057.safetensors", + "model.layers.102.mlp.up_proj.g_idx": "model-00046-of-00057.safetensors", + "model.layers.102.mlp.up_proj.qweight": "model-00046-of-00057.safetensors", + "model.layers.102.mlp.up_proj.qzeros": "model-00046-of-00057.safetensors", + "model.layers.102.mlp.up_proj.scales": "model-00046-of-00057.safetensors", + "model.layers.102.post_attention_layernorm.weight": "model-00046-of-00057.safetensors", + "model.layers.102.self_attn.k_proj.g_idx": "model-00045-of-00057.safetensors", + "model.layers.102.self_attn.k_proj.qweight": "model-00045-of-00057.safetensors", + "model.layers.102.self_attn.k_proj.qzeros": "model-00045-of-00057.safetensors", + "model.layers.102.self_attn.k_proj.scales": "model-00045-of-00057.safetensors", + "model.layers.102.self_attn.o_proj.g_idx": "model-00045-of-00057.safetensors", + "model.layers.102.self_attn.o_proj.qweight": "model-00045-of-00057.safetensors", + "model.layers.102.self_attn.o_proj.qzeros": "model-00045-of-00057.safetensors", + "model.layers.102.self_attn.o_proj.scales": "model-00045-of-00057.safetensors", + "model.layers.102.self_attn.q_proj.g_idx": "model-00045-of-00057.safetensors", + "model.layers.102.self_attn.q_proj.qweight": "model-00045-of-00057.safetensors", + "model.layers.102.self_attn.q_proj.qzeros": "model-00045-of-00057.safetensors", + "model.layers.102.self_attn.q_proj.scales": "model-00045-of-00057.safetensors", + "model.layers.102.self_attn.v_proj.g_idx": "model-00045-of-00057.safetensors", + "model.layers.102.self_attn.v_proj.qweight": "model-00045-of-00057.safetensors", + "model.layers.102.self_attn.v_proj.qzeros": "model-00045-of-00057.safetensors", + "model.layers.102.self_attn.v_proj.scales": "model-00045-of-00057.safetensors", + "model.layers.103.input_layernorm.weight": "model-00046-of-00057.safetensors", + "model.layers.103.mlp.down_proj.g_idx": "model-00046-of-00057.safetensors", + "model.layers.103.mlp.down_proj.qweight": "model-00046-of-00057.safetensors", + "model.layers.103.mlp.down_proj.qzeros": "model-00046-of-00057.safetensors", + "model.layers.103.mlp.down_proj.scales": "model-00046-of-00057.safetensors", + "model.layers.103.mlp.gate_proj.g_idx": "model-00046-of-00057.safetensors", + "model.layers.103.mlp.gate_proj.qweight": "model-00046-of-00057.safetensors", + "model.layers.103.mlp.gate_proj.qzeros": "model-00046-of-00057.safetensors", + "model.layers.103.mlp.gate_proj.scales": "model-00046-of-00057.safetensors", + "model.layers.103.mlp.up_proj.g_idx": "model-00046-of-00057.safetensors", + "model.layers.103.mlp.up_proj.qweight": "model-00046-of-00057.safetensors", + "model.layers.103.mlp.up_proj.qzeros": "model-00046-of-00057.safetensors", + "model.layers.103.mlp.up_proj.scales": "model-00046-of-00057.safetensors", + "model.layers.103.post_attention_layernorm.weight": "model-00046-of-00057.safetensors", + "model.layers.103.self_attn.k_proj.g_idx": "model-00046-of-00057.safetensors", + "model.layers.103.self_attn.k_proj.qweight": "model-00046-of-00057.safetensors", + "model.layers.103.self_attn.k_proj.qzeros": "model-00046-of-00057.safetensors", + "model.layers.103.self_attn.k_proj.scales": "model-00046-of-00057.safetensors", + "model.layers.103.self_attn.o_proj.g_idx": "model-00046-of-00057.safetensors", + "model.layers.103.self_attn.o_proj.qweight": "model-00046-of-00057.safetensors", + "model.layers.103.self_attn.o_proj.qzeros": "model-00046-of-00057.safetensors", + "model.layers.103.self_attn.o_proj.scales": "model-00046-of-00057.safetensors", + "model.layers.103.self_attn.q_proj.g_idx": "model-00046-of-00057.safetensors", + "model.layers.103.self_attn.q_proj.qweight": "model-00046-of-00057.safetensors", + "model.layers.103.self_attn.q_proj.qzeros": "model-00046-of-00057.safetensors", + "model.layers.103.self_attn.q_proj.scales": "model-00046-of-00057.safetensors", + "model.layers.103.self_attn.v_proj.g_idx": "model-00046-of-00057.safetensors", + "model.layers.103.self_attn.v_proj.qweight": "model-00046-of-00057.safetensors", + "model.layers.103.self_attn.v_proj.qzeros": "model-00046-of-00057.safetensors", + "model.layers.103.self_attn.v_proj.scales": "model-00046-of-00057.safetensors", + "model.layers.104.input_layernorm.weight": "model-00047-of-00057.safetensors", + "model.layers.104.mlp.down_proj.g_idx": "model-00047-of-00057.safetensors", + "model.layers.104.mlp.down_proj.qweight": "model-00047-of-00057.safetensors", + "model.layers.104.mlp.down_proj.qzeros": "model-00047-of-00057.safetensors", + "model.layers.104.mlp.down_proj.scales": "model-00047-of-00057.safetensors", + "model.layers.104.mlp.gate_proj.g_idx": "model-00046-of-00057.safetensors", + "model.layers.104.mlp.gate_proj.qweight": "model-00046-of-00057.safetensors", + "model.layers.104.mlp.gate_proj.qzeros": "model-00046-of-00057.safetensors", + "model.layers.104.mlp.gate_proj.scales": "model-00046-of-00057.safetensors", + "model.layers.104.mlp.up_proj.g_idx": "model-00046-of-00057.safetensors", + "model.layers.104.mlp.up_proj.qweight": "model-00046-of-00057.safetensors", + "model.layers.104.mlp.up_proj.qzeros": "model-00046-of-00057.safetensors", + "model.layers.104.mlp.up_proj.scales": "model-00046-of-00057.safetensors", + "model.layers.104.post_attention_layernorm.weight": "model-00047-of-00057.safetensors", + "model.layers.104.self_attn.k_proj.g_idx": "model-00046-of-00057.safetensors", + "model.layers.104.self_attn.k_proj.qweight": "model-00046-of-00057.safetensors", + "model.layers.104.self_attn.k_proj.qzeros": "model-00046-of-00057.safetensors", + "model.layers.104.self_attn.k_proj.scales": "model-00046-of-00057.safetensors", + "model.layers.104.self_attn.o_proj.g_idx": "model-00046-of-00057.safetensors", + "model.layers.104.self_attn.o_proj.qweight": "model-00046-of-00057.safetensors", + "model.layers.104.self_attn.o_proj.qzeros": "model-00046-of-00057.safetensors", + "model.layers.104.self_attn.o_proj.scales": "model-00046-of-00057.safetensors", + "model.layers.104.self_attn.q_proj.g_idx": "model-00046-of-00057.safetensors", + "model.layers.104.self_attn.q_proj.qweight": "model-00046-of-00057.safetensors", + "model.layers.104.self_attn.q_proj.qzeros": "model-00046-of-00057.safetensors", + "model.layers.104.self_attn.q_proj.scales": "model-00046-of-00057.safetensors", + "model.layers.104.self_attn.v_proj.g_idx": "model-00046-of-00057.safetensors", + "model.layers.104.self_attn.v_proj.qweight": "model-00046-of-00057.safetensors", + "model.layers.104.self_attn.v_proj.qzeros": "model-00046-of-00057.safetensors", + "model.layers.104.self_attn.v_proj.scales": "model-00046-of-00057.safetensors", + "model.layers.105.input_layernorm.weight": "model-00047-of-00057.safetensors", + "model.layers.105.mlp.down_proj.g_idx": "model-00047-of-00057.safetensors", + "model.layers.105.mlp.down_proj.qweight": "model-00047-of-00057.safetensors", + "model.layers.105.mlp.down_proj.qzeros": "model-00047-of-00057.safetensors", + "model.layers.105.mlp.down_proj.scales": "model-00047-of-00057.safetensors", + "model.layers.105.mlp.gate_proj.g_idx": "model-00047-of-00057.safetensors", + "model.layers.105.mlp.gate_proj.qweight": "model-00047-of-00057.safetensors", + "model.layers.105.mlp.gate_proj.qzeros": "model-00047-of-00057.safetensors", + "model.layers.105.mlp.gate_proj.scales": "model-00047-of-00057.safetensors", + "model.layers.105.mlp.up_proj.g_idx": "model-00047-of-00057.safetensors", + "model.layers.105.mlp.up_proj.qweight": "model-00047-of-00057.safetensors", + "model.layers.105.mlp.up_proj.qzeros": "model-00047-of-00057.safetensors", + "model.layers.105.mlp.up_proj.scales": "model-00047-of-00057.safetensors", + "model.layers.105.post_attention_layernorm.weight": "model-00047-of-00057.safetensors", + "model.layers.105.self_attn.k_proj.g_idx": "model-00047-of-00057.safetensors", + "model.layers.105.self_attn.k_proj.qweight": "model-00047-of-00057.safetensors", + "model.layers.105.self_attn.k_proj.qzeros": "model-00047-of-00057.safetensors", + "model.layers.105.self_attn.k_proj.scales": "model-00047-of-00057.safetensors", + "model.layers.105.self_attn.o_proj.g_idx": "model-00047-of-00057.safetensors", + "model.layers.105.self_attn.o_proj.qweight": "model-00047-of-00057.safetensors", + "model.layers.105.self_attn.o_proj.qzeros": "model-00047-of-00057.safetensors", + "model.layers.105.self_attn.o_proj.scales": "model-00047-of-00057.safetensors", + "model.layers.105.self_attn.q_proj.g_idx": "model-00047-of-00057.safetensors", + "model.layers.105.self_attn.q_proj.qweight": "model-00047-of-00057.safetensors", + "model.layers.105.self_attn.q_proj.qzeros": "model-00047-of-00057.safetensors", + "model.layers.105.self_attn.q_proj.scales": "model-00047-of-00057.safetensors", + "model.layers.105.self_attn.v_proj.g_idx": "model-00047-of-00057.safetensors", + "model.layers.105.self_attn.v_proj.qweight": "model-00047-of-00057.safetensors", + "model.layers.105.self_attn.v_proj.qzeros": "model-00047-of-00057.safetensors", + "model.layers.105.self_attn.v_proj.scales": "model-00047-of-00057.safetensors", + "model.layers.106.input_layernorm.weight": "model-00047-of-00057.safetensors", + "model.layers.106.mlp.down_proj.g_idx": "model-00047-of-00057.safetensors", + "model.layers.106.mlp.down_proj.qweight": "model-00047-of-00057.safetensors", + "model.layers.106.mlp.down_proj.qzeros": "model-00047-of-00057.safetensors", + "model.layers.106.mlp.down_proj.scales": "model-00047-of-00057.safetensors", + "model.layers.106.mlp.gate_proj.g_idx": "model-00047-of-00057.safetensors", + "model.layers.106.mlp.gate_proj.qweight": "model-00047-of-00057.safetensors", + "model.layers.106.mlp.gate_proj.qzeros": "model-00047-of-00057.safetensors", + "model.layers.106.mlp.gate_proj.scales": "model-00047-of-00057.safetensors", + "model.layers.106.mlp.up_proj.g_idx": "model-00047-of-00057.safetensors", + "model.layers.106.mlp.up_proj.qweight": "model-00047-of-00057.safetensors", + "model.layers.106.mlp.up_proj.qzeros": "model-00047-of-00057.safetensors", + "model.layers.106.mlp.up_proj.scales": "model-00047-of-00057.safetensors", + "model.layers.106.post_attention_layernorm.weight": "model-00047-of-00057.safetensors", + "model.layers.106.self_attn.k_proj.g_idx": "model-00047-of-00057.safetensors", + "model.layers.106.self_attn.k_proj.qweight": "model-00047-of-00057.safetensors", + "model.layers.106.self_attn.k_proj.qzeros": "model-00047-of-00057.safetensors", + "model.layers.106.self_attn.k_proj.scales": "model-00047-of-00057.safetensors", + "model.layers.106.self_attn.o_proj.g_idx": "model-00047-of-00057.safetensors", + "model.layers.106.self_attn.o_proj.qweight": "model-00047-of-00057.safetensors", + "model.layers.106.self_attn.o_proj.qzeros": "model-00047-of-00057.safetensors", + "model.layers.106.self_attn.o_proj.scales": "model-00047-of-00057.safetensors", + "model.layers.106.self_attn.q_proj.g_idx": "model-00047-of-00057.safetensors", + "model.layers.106.self_attn.q_proj.qweight": "model-00047-of-00057.safetensors", + "model.layers.106.self_attn.q_proj.qzeros": "model-00047-of-00057.safetensors", + "model.layers.106.self_attn.q_proj.scales": "model-00047-of-00057.safetensors", + "model.layers.106.self_attn.v_proj.g_idx": "model-00047-of-00057.safetensors", + "model.layers.106.self_attn.v_proj.qweight": "model-00047-of-00057.safetensors", + "model.layers.106.self_attn.v_proj.qzeros": "model-00047-of-00057.safetensors", + "model.layers.106.self_attn.v_proj.scales": "model-00047-of-00057.safetensors", + "model.layers.107.input_layernorm.weight": "model-00048-of-00057.safetensors", + "model.layers.107.mlp.down_proj.g_idx": "model-00048-of-00057.safetensors", + "model.layers.107.mlp.down_proj.qweight": "model-00048-of-00057.safetensors", + "model.layers.107.mlp.down_proj.qzeros": "model-00048-of-00057.safetensors", + "model.layers.107.mlp.down_proj.scales": "model-00048-of-00057.safetensors", + "model.layers.107.mlp.gate_proj.g_idx": "model-00048-of-00057.safetensors", + "model.layers.107.mlp.gate_proj.qweight": "model-00048-of-00057.safetensors", + "model.layers.107.mlp.gate_proj.qzeros": "model-00048-of-00057.safetensors", + "model.layers.107.mlp.gate_proj.scales": "model-00048-of-00057.safetensors", + "model.layers.107.mlp.up_proj.g_idx": "model-00048-of-00057.safetensors", + "model.layers.107.mlp.up_proj.qweight": "model-00048-of-00057.safetensors", + "model.layers.107.mlp.up_proj.qzeros": "model-00048-of-00057.safetensors", + "model.layers.107.mlp.up_proj.scales": "model-00048-of-00057.safetensors", + "model.layers.107.post_attention_layernorm.weight": "model-00048-of-00057.safetensors", + "model.layers.107.self_attn.k_proj.g_idx": "model-00047-of-00057.safetensors", + "model.layers.107.self_attn.k_proj.qweight": "model-00047-of-00057.safetensors", + "model.layers.107.self_attn.k_proj.qzeros": "model-00047-of-00057.safetensors", + "model.layers.107.self_attn.k_proj.scales": "model-00047-of-00057.safetensors", + "model.layers.107.self_attn.o_proj.g_idx": "model-00048-of-00057.safetensors", + "model.layers.107.self_attn.o_proj.qweight": "model-00048-of-00057.safetensors", + "model.layers.107.self_attn.o_proj.qzeros": "model-00048-of-00057.safetensors", + "model.layers.107.self_attn.o_proj.scales": "model-00048-of-00057.safetensors", + "model.layers.107.self_attn.q_proj.g_idx": "model-00047-of-00057.safetensors", + "model.layers.107.self_attn.q_proj.qweight": "model-00047-of-00057.safetensors", + "model.layers.107.self_attn.q_proj.qzeros": "model-00047-of-00057.safetensors", + "model.layers.107.self_attn.q_proj.scales": "model-00047-of-00057.safetensors", + "model.layers.107.self_attn.v_proj.g_idx": "model-00047-of-00057.safetensors", + "model.layers.107.self_attn.v_proj.qweight": "model-00047-of-00057.safetensors", + "model.layers.107.self_attn.v_proj.qzeros": "model-00047-of-00057.safetensors", + "model.layers.107.self_attn.v_proj.scales": "model-00047-of-00057.safetensors", + "model.layers.108.input_layernorm.weight": "model-00048-of-00057.safetensors", + "model.layers.108.mlp.down_proj.g_idx": "model-00048-of-00057.safetensors", + "model.layers.108.mlp.down_proj.qweight": "model-00048-of-00057.safetensors", + "model.layers.108.mlp.down_proj.qzeros": "model-00048-of-00057.safetensors", + "model.layers.108.mlp.down_proj.scales": "model-00048-of-00057.safetensors", + "model.layers.108.mlp.gate_proj.g_idx": "model-00048-of-00057.safetensors", + "model.layers.108.mlp.gate_proj.qweight": "model-00048-of-00057.safetensors", + "model.layers.108.mlp.gate_proj.qzeros": "model-00048-of-00057.safetensors", + "model.layers.108.mlp.gate_proj.scales": "model-00048-of-00057.safetensors", + "model.layers.108.mlp.up_proj.g_idx": "model-00048-of-00057.safetensors", + "model.layers.108.mlp.up_proj.qweight": "model-00048-of-00057.safetensors", + "model.layers.108.mlp.up_proj.qzeros": "model-00048-of-00057.safetensors", + "model.layers.108.mlp.up_proj.scales": "model-00048-of-00057.safetensors", + "model.layers.108.post_attention_layernorm.weight": "model-00048-of-00057.safetensors", + "model.layers.108.self_attn.k_proj.g_idx": "model-00048-of-00057.safetensors", + "model.layers.108.self_attn.k_proj.qweight": "model-00048-of-00057.safetensors", + "model.layers.108.self_attn.k_proj.qzeros": "model-00048-of-00057.safetensors", + "model.layers.108.self_attn.k_proj.scales": "model-00048-of-00057.safetensors", + "model.layers.108.self_attn.o_proj.g_idx": "model-00048-of-00057.safetensors", + "model.layers.108.self_attn.o_proj.qweight": "model-00048-of-00057.safetensors", + "model.layers.108.self_attn.o_proj.qzeros": "model-00048-of-00057.safetensors", + "model.layers.108.self_attn.o_proj.scales": "model-00048-of-00057.safetensors", + "model.layers.108.self_attn.q_proj.g_idx": "model-00048-of-00057.safetensors", + "model.layers.108.self_attn.q_proj.qweight": "model-00048-of-00057.safetensors", + "model.layers.108.self_attn.q_proj.qzeros": "model-00048-of-00057.safetensors", + "model.layers.108.self_attn.q_proj.scales": "model-00048-of-00057.safetensors", + "model.layers.108.self_attn.v_proj.g_idx": "model-00048-of-00057.safetensors", + "model.layers.108.self_attn.v_proj.qweight": "model-00048-of-00057.safetensors", + "model.layers.108.self_attn.v_proj.qzeros": "model-00048-of-00057.safetensors", + "model.layers.108.self_attn.v_proj.scales": "model-00048-of-00057.safetensors", + "model.layers.109.input_layernorm.weight": "model-00049-of-00057.safetensors", + "model.layers.109.mlp.down_proj.g_idx": "model-00049-of-00057.safetensors", + "model.layers.109.mlp.down_proj.qweight": "model-00049-of-00057.safetensors", + "model.layers.109.mlp.down_proj.qzeros": "model-00049-of-00057.safetensors", + "model.layers.109.mlp.down_proj.scales": "model-00049-of-00057.safetensors", + "model.layers.109.mlp.gate_proj.g_idx": "model-00048-of-00057.safetensors", + "model.layers.109.mlp.gate_proj.qweight": "model-00048-of-00057.safetensors", + "model.layers.109.mlp.gate_proj.qzeros": "model-00048-of-00057.safetensors", + "model.layers.109.mlp.gate_proj.scales": "model-00048-of-00057.safetensors", + "model.layers.109.mlp.up_proj.g_idx": "model-00049-of-00057.safetensors", + "model.layers.109.mlp.up_proj.qweight": "model-00049-of-00057.safetensors", + "model.layers.109.mlp.up_proj.qzeros": "model-00049-of-00057.safetensors", + "model.layers.109.mlp.up_proj.scales": "model-00049-of-00057.safetensors", + "model.layers.109.post_attention_layernorm.weight": "model-00049-of-00057.safetensors", + "model.layers.109.self_attn.k_proj.g_idx": "model-00048-of-00057.safetensors", + "model.layers.109.self_attn.k_proj.qweight": "model-00048-of-00057.safetensors", + "model.layers.109.self_attn.k_proj.qzeros": "model-00048-of-00057.safetensors", + "model.layers.109.self_attn.k_proj.scales": "model-00048-of-00057.safetensors", + "model.layers.109.self_attn.o_proj.g_idx": "model-00048-of-00057.safetensors", + "model.layers.109.self_attn.o_proj.qweight": "model-00048-of-00057.safetensors", + "model.layers.109.self_attn.o_proj.qzeros": "model-00048-of-00057.safetensors", + "model.layers.109.self_attn.o_proj.scales": "model-00048-of-00057.safetensors", + "model.layers.109.self_attn.q_proj.g_idx": "model-00048-of-00057.safetensors", + "model.layers.109.self_attn.q_proj.qweight": "model-00048-of-00057.safetensors", + "model.layers.109.self_attn.q_proj.qzeros": "model-00048-of-00057.safetensors", + "model.layers.109.self_attn.q_proj.scales": "model-00048-of-00057.safetensors", + "model.layers.109.self_attn.v_proj.g_idx": "model-00048-of-00057.safetensors", + "model.layers.109.self_attn.v_proj.qweight": "model-00048-of-00057.safetensors", + "model.layers.109.self_attn.v_proj.qzeros": "model-00048-of-00057.safetensors", + "model.layers.109.self_attn.v_proj.scales": "model-00048-of-00057.safetensors", + "model.layers.11.input_layernorm.weight": "model-00007-of-00057.safetensors", + "model.layers.11.mlp.down_proj.g_idx": "model-00007-of-00057.safetensors", + "model.layers.11.mlp.down_proj.qweight": "model-00007-of-00057.safetensors", + "model.layers.11.mlp.down_proj.qzeros": "model-00007-of-00057.safetensors", + "model.layers.11.mlp.down_proj.scales": "model-00007-of-00057.safetensors", + "model.layers.11.mlp.gate_proj.g_idx": "model-00006-of-00057.safetensors", + "model.layers.11.mlp.gate_proj.qweight": "model-00006-of-00057.safetensors", + "model.layers.11.mlp.gate_proj.qzeros": "model-00006-of-00057.safetensors", + "model.layers.11.mlp.gate_proj.scales": "model-00006-of-00057.safetensors", + "model.layers.11.mlp.up_proj.g_idx": "model-00007-of-00057.safetensors", + "model.layers.11.mlp.up_proj.qweight": "model-00007-of-00057.safetensors", + "model.layers.11.mlp.up_proj.qzeros": "model-00007-of-00057.safetensors", + "model.layers.11.mlp.up_proj.scales": "model-00007-of-00057.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00007-of-00057.safetensors", + "model.layers.11.self_attn.k_proj.g_idx": "model-00006-of-00057.safetensors", + "model.layers.11.self_attn.k_proj.qweight": "model-00006-of-00057.safetensors", + "model.layers.11.self_attn.k_proj.qzeros": "model-00006-of-00057.safetensors", + "model.layers.11.self_attn.k_proj.scales": "model-00006-of-00057.safetensors", + "model.layers.11.self_attn.o_proj.g_idx": "model-00006-of-00057.safetensors", + "model.layers.11.self_attn.o_proj.qweight": "model-00006-of-00057.safetensors", + "model.layers.11.self_attn.o_proj.qzeros": "model-00006-of-00057.safetensors", + "model.layers.11.self_attn.o_proj.scales": "model-00006-of-00057.safetensors", + "model.layers.11.self_attn.q_proj.g_idx": "model-00006-of-00057.safetensors", + "model.layers.11.self_attn.q_proj.qweight": "model-00006-of-00057.safetensors", + "model.layers.11.self_attn.q_proj.qzeros": "model-00006-of-00057.safetensors", + "model.layers.11.self_attn.q_proj.scales": "model-00006-of-00057.safetensors", + "model.layers.11.self_attn.v_proj.g_idx": "model-00006-of-00057.safetensors", + "model.layers.11.self_attn.v_proj.qweight": "model-00006-of-00057.safetensors", + "model.layers.11.self_attn.v_proj.qzeros": "model-00006-of-00057.safetensors", + "model.layers.11.self_attn.v_proj.scales": "model-00006-of-00057.safetensors", + "model.layers.110.input_layernorm.weight": "model-00049-of-00057.safetensors", + "model.layers.110.mlp.down_proj.g_idx": "model-00049-of-00057.safetensors", + "model.layers.110.mlp.down_proj.qweight": "model-00049-of-00057.safetensors", + "model.layers.110.mlp.down_proj.qzeros": "model-00049-of-00057.safetensors", + "model.layers.110.mlp.down_proj.scales": "model-00049-of-00057.safetensors", + "model.layers.110.mlp.gate_proj.g_idx": "model-00049-of-00057.safetensors", + "model.layers.110.mlp.gate_proj.qweight": "model-00049-of-00057.safetensors", + "model.layers.110.mlp.gate_proj.qzeros": "model-00049-of-00057.safetensors", + "model.layers.110.mlp.gate_proj.scales": "model-00049-of-00057.safetensors", + "model.layers.110.mlp.up_proj.g_idx": "model-00049-of-00057.safetensors", + "model.layers.110.mlp.up_proj.qweight": "model-00049-of-00057.safetensors", + "model.layers.110.mlp.up_proj.qzeros": "model-00049-of-00057.safetensors", + "model.layers.110.mlp.up_proj.scales": "model-00049-of-00057.safetensors", + "model.layers.110.post_attention_layernorm.weight": "model-00049-of-00057.safetensors", + "model.layers.110.self_attn.k_proj.g_idx": "model-00049-of-00057.safetensors", + "model.layers.110.self_attn.k_proj.qweight": "model-00049-of-00057.safetensors", + "model.layers.110.self_attn.k_proj.qzeros": "model-00049-of-00057.safetensors", + "model.layers.110.self_attn.k_proj.scales": "model-00049-of-00057.safetensors", + "model.layers.110.self_attn.o_proj.g_idx": "model-00049-of-00057.safetensors", + "model.layers.110.self_attn.o_proj.qweight": "model-00049-of-00057.safetensors", + "model.layers.110.self_attn.o_proj.qzeros": "model-00049-of-00057.safetensors", + "model.layers.110.self_attn.o_proj.scales": "model-00049-of-00057.safetensors", + "model.layers.110.self_attn.q_proj.g_idx": "model-00049-of-00057.safetensors", + "model.layers.110.self_attn.q_proj.qweight": "model-00049-of-00057.safetensors", + "model.layers.110.self_attn.q_proj.qzeros": "model-00049-of-00057.safetensors", + "model.layers.110.self_attn.q_proj.scales": "model-00049-of-00057.safetensors", + "model.layers.110.self_attn.v_proj.g_idx": "model-00049-of-00057.safetensors", + "model.layers.110.self_attn.v_proj.qweight": "model-00049-of-00057.safetensors", + "model.layers.110.self_attn.v_proj.qzeros": "model-00049-of-00057.safetensors", + "model.layers.110.self_attn.v_proj.scales": "model-00049-of-00057.safetensors", + "model.layers.111.input_layernorm.weight": "model-00050-of-00057.safetensors", + "model.layers.111.mlp.down_proj.g_idx": "model-00050-of-00057.safetensors", + "model.layers.111.mlp.down_proj.qweight": "model-00050-of-00057.safetensors", + "model.layers.111.mlp.down_proj.qzeros": "model-00050-of-00057.safetensors", + "model.layers.111.mlp.down_proj.scales": "model-00050-of-00057.safetensors", + "model.layers.111.mlp.gate_proj.g_idx": "model-00049-of-00057.safetensors", + "model.layers.111.mlp.gate_proj.qweight": "model-00049-of-00057.safetensors", + "model.layers.111.mlp.gate_proj.qzeros": "model-00049-of-00057.safetensors", + "model.layers.111.mlp.gate_proj.scales": "model-00049-of-00057.safetensors", + "model.layers.111.mlp.up_proj.g_idx": "model-00049-of-00057.safetensors", + "model.layers.111.mlp.up_proj.qweight": "model-00049-of-00057.safetensors", + "model.layers.111.mlp.up_proj.qzeros": "model-00049-of-00057.safetensors", + "model.layers.111.mlp.up_proj.scales": "model-00049-of-00057.safetensors", + "model.layers.111.post_attention_layernorm.weight": "model-00050-of-00057.safetensors", + "model.layers.111.self_attn.k_proj.g_idx": "model-00049-of-00057.safetensors", + "model.layers.111.self_attn.k_proj.qweight": "model-00049-of-00057.safetensors", + "model.layers.111.self_attn.k_proj.qzeros": "model-00049-of-00057.safetensors", + "model.layers.111.self_attn.k_proj.scales": "model-00049-of-00057.safetensors", + "model.layers.111.self_attn.o_proj.g_idx": "model-00049-of-00057.safetensors", + "model.layers.111.self_attn.o_proj.qweight": "model-00049-of-00057.safetensors", + "model.layers.111.self_attn.o_proj.qzeros": "model-00049-of-00057.safetensors", + "model.layers.111.self_attn.o_proj.scales": "model-00049-of-00057.safetensors", + "model.layers.111.self_attn.q_proj.g_idx": "model-00049-of-00057.safetensors", + "model.layers.111.self_attn.q_proj.qweight": "model-00049-of-00057.safetensors", + "model.layers.111.self_attn.q_proj.qzeros": "model-00049-of-00057.safetensors", + "model.layers.111.self_attn.q_proj.scales": "model-00049-of-00057.safetensors", + "model.layers.111.self_attn.v_proj.g_idx": "model-00049-of-00057.safetensors", + "model.layers.111.self_attn.v_proj.qweight": "model-00049-of-00057.safetensors", + "model.layers.111.self_attn.v_proj.qzeros": "model-00049-of-00057.safetensors", + "model.layers.111.self_attn.v_proj.scales": "model-00049-of-00057.safetensors", + "model.layers.112.input_layernorm.weight": "model-00050-of-00057.safetensors", + "model.layers.112.mlp.down_proj.g_idx": "model-00050-of-00057.safetensors", + "model.layers.112.mlp.down_proj.qweight": "model-00050-of-00057.safetensors", + "model.layers.112.mlp.down_proj.qzeros": "model-00050-of-00057.safetensors", + "model.layers.112.mlp.down_proj.scales": "model-00050-of-00057.safetensors", + "model.layers.112.mlp.gate_proj.g_idx": "model-00050-of-00057.safetensors", + "model.layers.112.mlp.gate_proj.qweight": "model-00050-of-00057.safetensors", + "model.layers.112.mlp.gate_proj.qzeros": "model-00050-of-00057.safetensors", + "model.layers.112.mlp.gate_proj.scales": "model-00050-of-00057.safetensors", + "model.layers.112.mlp.up_proj.g_idx": "model-00050-of-00057.safetensors", + "model.layers.112.mlp.up_proj.qweight": "model-00050-of-00057.safetensors", + "model.layers.112.mlp.up_proj.qzeros": "model-00050-of-00057.safetensors", + "model.layers.112.mlp.up_proj.scales": "model-00050-of-00057.safetensors", + "model.layers.112.post_attention_layernorm.weight": "model-00050-of-00057.safetensors", + "model.layers.112.self_attn.k_proj.g_idx": "model-00050-of-00057.safetensors", + "model.layers.112.self_attn.k_proj.qweight": "model-00050-of-00057.safetensors", + "model.layers.112.self_attn.k_proj.qzeros": "model-00050-of-00057.safetensors", + "model.layers.112.self_attn.k_proj.scales": "model-00050-of-00057.safetensors", + "model.layers.112.self_attn.o_proj.g_idx": "model-00050-of-00057.safetensors", + "model.layers.112.self_attn.o_proj.qweight": "model-00050-of-00057.safetensors", + "model.layers.112.self_attn.o_proj.qzeros": "model-00050-of-00057.safetensors", + "model.layers.112.self_attn.o_proj.scales": "model-00050-of-00057.safetensors", + "model.layers.112.self_attn.q_proj.g_idx": "model-00050-of-00057.safetensors", + "model.layers.112.self_attn.q_proj.qweight": "model-00050-of-00057.safetensors", + "model.layers.112.self_attn.q_proj.qzeros": "model-00050-of-00057.safetensors", + "model.layers.112.self_attn.q_proj.scales": "model-00050-of-00057.safetensors", + "model.layers.112.self_attn.v_proj.g_idx": "model-00050-of-00057.safetensors", + "model.layers.112.self_attn.v_proj.qweight": "model-00050-of-00057.safetensors", + "model.layers.112.self_attn.v_proj.qzeros": "model-00050-of-00057.safetensors", + "model.layers.112.self_attn.v_proj.scales": "model-00050-of-00057.safetensors", + "model.layers.113.input_layernorm.weight": "model-00050-of-00057.safetensors", + "model.layers.113.mlp.down_proj.g_idx": "model-00050-of-00057.safetensors", + "model.layers.113.mlp.down_proj.qweight": "model-00050-of-00057.safetensors", + "model.layers.113.mlp.down_proj.qzeros": "model-00050-of-00057.safetensors", + "model.layers.113.mlp.down_proj.scales": "model-00050-of-00057.safetensors", + "model.layers.113.mlp.gate_proj.g_idx": "model-00050-of-00057.safetensors", + "model.layers.113.mlp.gate_proj.qweight": "model-00050-of-00057.safetensors", + "model.layers.113.mlp.gate_proj.qzeros": "model-00050-of-00057.safetensors", + "model.layers.113.mlp.gate_proj.scales": "model-00050-of-00057.safetensors", + "model.layers.113.mlp.up_proj.g_idx": "model-00050-of-00057.safetensors", + "model.layers.113.mlp.up_proj.qweight": "model-00050-of-00057.safetensors", + "model.layers.113.mlp.up_proj.qzeros": "model-00050-of-00057.safetensors", + "model.layers.113.mlp.up_proj.scales": "model-00050-of-00057.safetensors", + "model.layers.113.post_attention_layernorm.weight": "model-00050-of-00057.safetensors", + "model.layers.113.self_attn.k_proj.g_idx": "model-00050-of-00057.safetensors", + "model.layers.113.self_attn.k_proj.qweight": "model-00050-of-00057.safetensors", + "model.layers.113.self_attn.k_proj.qzeros": "model-00050-of-00057.safetensors", + "model.layers.113.self_attn.k_proj.scales": "model-00050-of-00057.safetensors", + "model.layers.113.self_attn.o_proj.g_idx": "model-00050-of-00057.safetensors", + "model.layers.113.self_attn.o_proj.qweight": "model-00050-of-00057.safetensors", + "model.layers.113.self_attn.o_proj.qzeros": "model-00050-of-00057.safetensors", + "model.layers.113.self_attn.o_proj.scales": "model-00050-of-00057.safetensors", + "model.layers.113.self_attn.q_proj.g_idx": "model-00050-of-00057.safetensors", + "model.layers.113.self_attn.q_proj.qweight": "model-00050-of-00057.safetensors", + "model.layers.113.self_attn.q_proj.qzeros": "model-00050-of-00057.safetensors", + "model.layers.113.self_attn.q_proj.scales": "model-00050-of-00057.safetensors", + "model.layers.113.self_attn.v_proj.g_idx": "model-00050-of-00057.safetensors", + "model.layers.113.self_attn.v_proj.qweight": "model-00050-of-00057.safetensors", + "model.layers.113.self_attn.v_proj.qzeros": "model-00050-of-00057.safetensors", + "model.layers.113.self_attn.v_proj.scales": "model-00050-of-00057.safetensors", + "model.layers.114.input_layernorm.weight": "model-00051-of-00057.safetensors", + "model.layers.114.mlp.down_proj.g_idx": "model-00051-of-00057.safetensors", + "model.layers.114.mlp.down_proj.qweight": "model-00051-of-00057.safetensors", + "model.layers.114.mlp.down_proj.qzeros": "model-00051-of-00057.safetensors", + "model.layers.114.mlp.down_proj.scales": "model-00051-of-00057.safetensors", + "model.layers.114.mlp.gate_proj.g_idx": "model-00051-of-00057.safetensors", + "model.layers.114.mlp.gate_proj.qweight": "model-00051-of-00057.safetensors", + "model.layers.114.mlp.gate_proj.qzeros": "model-00051-of-00057.safetensors", + "model.layers.114.mlp.gate_proj.scales": "model-00051-of-00057.safetensors", + "model.layers.114.mlp.up_proj.g_idx": "model-00051-of-00057.safetensors", + "model.layers.114.mlp.up_proj.qweight": "model-00051-of-00057.safetensors", + "model.layers.114.mlp.up_proj.qzeros": "model-00051-of-00057.safetensors", + "model.layers.114.mlp.up_proj.scales": "model-00051-of-00057.safetensors", + "model.layers.114.post_attention_layernorm.weight": "model-00051-of-00057.safetensors", + "model.layers.114.self_attn.k_proj.g_idx": "model-00050-of-00057.safetensors", + "model.layers.114.self_attn.k_proj.qweight": "model-00050-of-00057.safetensors", + "model.layers.114.self_attn.k_proj.qzeros": "model-00050-of-00057.safetensors", + "model.layers.114.self_attn.k_proj.scales": "model-00050-of-00057.safetensors", + "model.layers.114.self_attn.o_proj.g_idx": "model-00051-of-00057.safetensors", + "model.layers.114.self_attn.o_proj.qweight": "model-00051-of-00057.safetensors", + "model.layers.114.self_attn.o_proj.qzeros": "model-00051-of-00057.safetensors", + "model.layers.114.self_attn.o_proj.scales": "model-00051-of-00057.safetensors", + "model.layers.114.self_attn.q_proj.g_idx": "model-00050-of-00057.safetensors", + "model.layers.114.self_attn.q_proj.qweight": "model-00050-of-00057.safetensors", + "model.layers.114.self_attn.q_proj.qzeros": "model-00050-of-00057.safetensors", + "model.layers.114.self_attn.q_proj.scales": "model-00050-of-00057.safetensors", + "model.layers.114.self_attn.v_proj.g_idx": "model-00050-of-00057.safetensors", + "model.layers.114.self_attn.v_proj.qweight": "model-00050-of-00057.safetensors", + "model.layers.114.self_attn.v_proj.qzeros": "model-00050-of-00057.safetensors", + "model.layers.114.self_attn.v_proj.scales": "model-00050-of-00057.safetensors", + "model.layers.115.input_layernorm.weight": "model-00051-of-00057.safetensors", + "model.layers.115.mlp.down_proj.g_idx": "model-00051-of-00057.safetensors", + "model.layers.115.mlp.down_proj.qweight": "model-00051-of-00057.safetensors", + "model.layers.115.mlp.down_proj.qzeros": "model-00051-of-00057.safetensors", + "model.layers.115.mlp.down_proj.scales": "model-00051-of-00057.safetensors", + "model.layers.115.mlp.gate_proj.g_idx": "model-00051-of-00057.safetensors", + "model.layers.115.mlp.gate_proj.qweight": "model-00051-of-00057.safetensors", + "model.layers.115.mlp.gate_proj.qzeros": "model-00051-of-00057.safetensors", + "model.layers.115.mlp.gate_proj.scales": "model-00051-of-00057.safetensors", + "model.layers.115.mlp.up_proj.g_idx": "model-00051-of-00057.safetensors", + "model.layers.115.mlp.up_proj.qweight": "model-00051-of-00057.safetensors", + "model.layers.115.mlp.up_proj.qzeros": "model-00051-of-00057.safetensors", + "model.layers.115.mlp.up_proj.scales": "model-00051-of-00057.safetensors", + "model.layers.115.post_attention_layernorm.weight": "model-00051-of-00057.safetensors", + "model.layers.115.self_attn.k_proj.g_idx": "model-00051-of-00057.safetensors", + "model.layers.115.self_attn.k_proj.qweight": "model-00051-of-00057.safetensors", + "model.layers.115.self_attn.k_proj.qzeros": "model-00051-of-00057.safetensors", + "model.layers.115.self_attn.k_proj.scales": "model-00051-of-00057.safetensors", + "model.layers.115.self_attn.o_proj.g_idx": "model-00051-of-00057.safetensors", + "model.layers.115.self_attn.o_proj.qweight": "model-00051-of-00057.safetensors", + "model.layers.115.self_attn.o_proj.qzeros": "model-00051-of-00057.safetensors", + "model.layers.115.self_attn.o_proj.scales": "model-00051-of-00057.safetensors", + "model.layers.115.self_attn.q_proj.g_idx": "model-00051-of-00057.safetensors", + "model.layers.115.self_attn.q_proj.qweight": "model-00051-of-00057.safetensors", + "model.layers.115.self_attn.q_proj.qzeros": "model-00051-of-00057.safetensors", + "model.layers.115.self_attn.q_proj.scales": "model-00051-of-00057.safetensors", + "model.layers.115.self_attn.v_proj.g_idx": "model-00051-of-00057.safetensors", + "model.layers.115.self_attn.v_proj.qweight": "model-00051-of-00057.safetensors", + "model.layers.115.self_attn.v_proj.qzeros": "model-00051-of-00057.safetensors", + "model.layers.115.self_attn.v_proj.scales": "model-00051-of-00057.safetensors", + "model.layers.116.input_layernorm.weight": "model-00052-of-00057.safetensors", + "model.layers.116.mlp.down_proj.g_idx": "model-00052-of-00057.safetensors", + "model.layers.116.mlp.down_proj.qweight": "model-00052-of-00057.safetensors", + "model.layers.116.mlp.down_proj.qzeros": "model-00052-of-00057.safetensors", + "model.layers.116.mlp.down_proj.scales": "model-00052-of-00057.safetensors", + "model.layers.116.mlp.gate_proj.g_idx": "model-00051-of-00057.safetensors", + "model.layers.116.mlp.gate_proj.qweight": "model-00051-of-00057.safetensors", + "model.layers.116.mlp.gate_proj.qzeros": "model-00051-of-00057.safetensors", + "model.layers.116.mlp.gate_proj.scales": "model-00051-of-00057.safetensors", + "model.layers.116.mlp.up_proj.g_idx": "model-00052-of-00057.safetensors", + "model.layers.116.mlp.up_proj.qweight": "model-00052-of-00057.safetensors", + "model.layers.116.mlp.up_proj.qzeros": "model-00052-of-00057.safetensors", + "model.layers.116.mlp.up_proj.scales": "model-00052-of-00057.safetensors", + "model.layers.116.post_attention_layernorm.weight": "model-00052-of-00057.safetensors", + "model.layers.116.self_attn.k_proj.g_idx": "model-00051-of-00057.safetensors", + "model.layers.116.self_attn.k_proj.qweight": "model-00051-of-00057.safetensors", + "model.layers.116.self_attn.k_proj.qzeros": "model-00051-of-00057.safetensors", + "model.layers.116.self_attn.k_proj.scales": "model-00051-of-00057.safetensors", + "model.layers.116.self_attn.o_proj.g_idx": "model-00051-of-00057.safetensors", + "model.layers.116.self_attn.o_proj.qweight": "model-00051-of-00057.safetensors", + "model.layers.116.self_attn.o_proj.qzeros": "model-00051-of-00057.safetensors", + "model.layers.116.self_attn.o_proj.scales": "model-00051-of-00057.safetensors", + "model.layers.116.self_attn.q_proj.g_idx": "model-00051-of-00057.safetensors", + "model.layers.116.self_attn.q_proj.qweight": "model-00051-of-00057.safetensors", + "model.layers.116.self_attn.q_proj.qzeros": "model-00051-of-00057.safetensors", + "model.layers.116.self_attn.q_proj.scales": "model-00051-of-00057.safetensors", + "model.layers.116.self_attn.v_proj.g_idx": "model-00051-of-00057.safetensors", + "model.layers.116.self_attn.v_proj.qweight": "model-00051-of-00057.safetensors", + "model.layers.116.self_attn.v_proj.qzeros": "model-00051-of-00057.safetensors", + "model.layers.116.self_attn.v_proj.scales": "model-00051-of-00057.safetensors", + "model.layers.117.input_layernorm.weight": "model-00052-of-00057.safetensors", + "model.layers.117.mlp.down_proj.g_idx": "model-00052-of-00057.safetensors", + "model.layers.117.mlp.down_proj.qweight": "model-00052-of-00057.safetensors", + "model.layers.117.mlp.down_proj.qzeros": "model-00052-of-00057.safetensors", + "model.layers.117.mlp.down_proj.scales": "model-00052-of-00057.safetensors", + "model.layers.117.mlp.gate_proj.g_idx": "model-00052-of-00057.safetensors", + "model.layers.117.mlp.gate_proj.qweight": "model-00052-of-00057.safetensors", + "model.layers.117.mlp.gate_proj.qzeros": "model-00052-of-00057.safetensors", + "model.layers.117.mlp.gate_proj.scales": "model-00052-of-00057.safetensors", + "model.layers.117.mlp.up_proj.g_idx": "model-00052-of-00057.safetensors", + "model.layers.117.mlp.up_proj.qweight": "model-00052-of-00057.safetensors", + "model.layers.117.mlp.up_proj.qzeros": "model-00052-of-00057.safetensors", + "model.layers.117.mlp.up_proj.scales": "model-00052-of-00057.safetensors", + "model.layers.117.post_attention_layernorm.weight": "model-00052-of-00057.safetensors", + "model.layers.117.self_attn.k_proj.g_idx": "model-00052-of-00057.safetensors", + "model.layers.117.self_attn.k_proj.qweight": "model-00052-of-00057.safetensors", + "model.layers.117.self_attn.k_proj.qzeros": "model-00052-of-00057.safetensors", + "model.layers.117.self_attn.k_proj.scales": "model-00052-of-00057.safetensors", + "model.layers.117.self_attn.o_proj.g_idx": "model-00052-of-00057.safetensors", + "model.layers.117.self_attn.o_proj.qweight": "model-00052-of-00057.safetensors", + "model.layers.117.self_attn.o_proj.qzeros": "model-00052-of-00057.safetensors", + "model.layers.117.self_attn.o_proj.scales": "model-00052-of-00057.safetensors", + "model.layers.117.self_attn.q_proj.g_idx": "model-00052-of-00057.safetensors", + "model.layers.117.self_attn.q_proj.qweight": "model-00052-of-00057.safetensors", + "model.layers.117.self_attn.q_proj.qzeros": "model-00052-of-00057.safetensors", + "model.layers.117.self_attn.q_proj.scales": "model-00052-of-00057.safetensors", + "model.layers.117.self_attn.v_proj.g_idx": "model-00052-of-00057.safetensors", + "model.layers.117.self_attn.v_proj.qweight": "model-00052-of-00057.safetensors", + "model.layers.117.self_attn.v_proj.qzeros": "model-00052-of-00057.safetensors", + "model.layers.117.self_attn.v_proj.scales": "model-00052-of-00057.safetensors", + "model.layers.118.input_layernorm.weight": "model-00053-of-00057.safetensors", + "model.layers.118.mlp.down_proj.g_idx": "model-00053-of-00057.safetensors", + "model.layers.118.mlp.down_proj.qweight": "model-00053-of-00057.safetensors", + "model.layers.118.mlp.down_proj.qzeros": "model-00053-of-00057.safetensors", + "model.layers.118.mlp.down_proj.scales": "model-00053-of-00057.safetensors", + "model.layers.118.mlp.gate_proj.g_idx": "model-00052-of-00057.safetensors", + "model.layers.118.mlp.gate_proj.qweight": "model-00052-of-00057.safetensors", + "model.layers.118.mlp.gate_proj.qzeros": "model-00052-of-00057.safetensors", + "model.layers.118.mlp.gate_proj.scales": "model-00052-of-00057.safetensors", + "model.layers.118.mlp.up_proj.g_idx": "model-00052-of-00057.safetensors", + "model.layers.118.mlp.up_proj.qweight": "model-00052-of-00057.safetensors", + "model.layers.118.mlp.up_proj.qzeros": "model-00052-of-00057.safetensors", + "model.layers.118.mlp.up_proj.scales": "model-00052-of-00057.safetensors", + "model.layers.118.post_attention_layernorm.weight": "model-00053-of-00057.safetensors", + "model.layers.118.self_attn.k_proj.g_idx": "model-00052-of-00057.safetensors", + "model.layers.118.self_attn.k_proj.qweight": "model-00052-of-00057.safetensors", + "model.layers.118.self_attn.k_proj.qzeros": "model-00052-of-00057.safetensors", + "model.layers.118.self_attn.k_proj.scales": "model-00052-of-00057.safetensors", + "model.layers.118.self_attn.o_proj.g_idx": "model-00052-of-00057.safetensors", + "model.layers.118.self_attn.o_proj.qweight": "model-00052-of-00057.safetensors", + "model.layers.118.self_attn.o_proj.qzeros": "model-00052-of-00057.safetensors", + "model.layers.118.self_attn.o_proj.scales": "model-00052-of-00057.safetensors", + "model.layers.118.self_attn.q_proj.g_idx": "model-00052-of-00057.safetensors", + "model.layers.118.self_attn.q_proj.qweight": "model-00052-of-00057.safetensors", + "model.layers.118.self_attn.q_proj.qzeros": "model-00052-of-00057.safetensors", + "model.layers.118.self_attn.q_proj.scales": "model-00052-of-00057.safetensors", + "model.layers.118.self_attn.v_proj.g_idx": "model-00052-of-00057.safetensors", + "model.layers.118.self_attn.v_proj.qweight": "model-00052-of-00057.safetensors", + "model.layers.118.self_attn.v_proj.qzeros": "model-00052-of-00057.safetensors", + "model.layers.118.self_attn.v_proj.scales": "model-00052-of-00057.safetensors", + "model.layers.119.input_layernorm.weight": "model-00053-of-00057.safetensors", + "model.layers.119.mlp.down_proj.g_idx": "model-00053-of-00057.safetensors", + "model.layers.119.mlp.down_proj.qweight": "model-00053-of-00057.safetensors", + "model.layers.119.mlp.down_proj.qzeros": "model-00053-of-00057.safetensors", + "model.layers.119.mlp.down_proj.scales": "model-00053-of-00057.safetensors", + "model.layers.119.mlp.gate_proj.g_idx": "model-00053-of-00057.safetensors", + "model.layers.119.mlp.gate_proj.qweight": "model-00053-of-00057.safetensors", + "model.layers.119.mlp.gate_proj.qzeros": "model-00053-of-00057.safetensors", + "model.layers.119.mlp.gate_proj.scales": "model-00053-of-00057.safetensors", + "model.layers.119.mlp.up_proj.g_idx": "model-00053-of-00057.safetensors", + "model.layers.119.mlp.up_proj.qweight": "model-00053-of-00057.safetensors", + "model.layers.119.mlp.up_proj.qzeros": "model-00053-of-00057.safetensors", + "model.layers.119.mlp.up_proj.scales": "model-00053-of-00057.safetensors", + "model.layers.119.post_attention_layernorm.weight": "model-00053-of-00057.safetensors", + "model.layers.119.self_attn.k_proj.g_idx": "model-00053-of-00057.safetensors", + "model.layers.119.self_attn.k_proj.qweight": "model-00053-of-00057.safetensors", + "model.layers.119.self_attn.k_proj.qzeros": "model-00053-of-00057.safetensors", + "model.layers.119.self_attn.k_proj.scales": "model-00053-of-00057.safetensors", + "model.layers.119.self_attn.o_proj.g_idx": "model-00053-of-00057.safetensors", + "model.layers.119.self_attn.o_proj.qweight": "model-00053-of-00057.safetensors", + "model.layers.119.self_attn.o_proj.qzeros": "model-00053-of-00057.safetensors", + "model.layers.119.self_attn.o_proj.scales": "model-00053-of-00057.safetensors", + "model.layers.119.self_attn.q_proj.g_idx": "model-00053-of-00057.safetensors", + "model.layers.119.self_attn.q_proj.qweight": "model-00053-of-00057.safetensors", + "model.layers.119.self_attn.q_proj.qzeros": "model-00053-of-00057.safetensors", + "model.layers.119.self_attn.q_proj.scales": "model-00053-of-00057.safetensors", + "model.layers.119.self_attn.v_proj.g_idx": "model-00053-of-00057.safetensors", + "model.layers.119.self_attn.v_proj.qweight": "model-00053-of-00057.safetensors", + "model.layers.119.self_attn.v_proj.qzeros": "model-00053-of-00057.safetensors", + "model.layers.119.self_attn.v_proj.scales": "model-00053-of-00057.safetensors", + "model.layers.12.input_layernorm.weight": "model-00007-of-00057.safetensors", + "model.layers.12.mlp.down_proj.g_idx": "model-00007-of-00057.safetensors", + "model.layers.12.mlp.down_proj.qweight": "model-00007-of-00057.safetensors", + "model.layers.12.mlp.down_proj.qzeros": "model-00007-of-00057.safetensors", + "model.layers.12.mlp.down_proj.scales": "model-00007-of-00057.safetensors", + "model.layers.12.mlp.gate_proj.g_idx": "model-00007-of-00057.safetensors", + "model.layers.12.mlp.gate_proj.qweight": "model-00007-of-00057.safetensors", + "model.layers.12.mlp.gate_proj.qzeros": "model-00007-of-00057.safetensors", + "model.layers.12.mlp.gate_proj.scales": "model-00007-of-00057.safetensors", + "model.layers.12.mlp.up_proj.g_idx": "model-00007-of-00057.safetensors", + "model.layers.12.mlp.up_proj.qweight": "model-00007-of-00057.safetensors", + "model.layers.12.mlp.up_proj.qzeros": "model-00007-of-00057.safetensors", + "model.layers.12.mlp.up_proj.scales": "model-00007-of-00057.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00007-of-00057.safetensors", + "model.layers.12.self_attn.k_proj.g_idx": "model-00007-of-00057.safetensors", + "model.layers.12.self_attn.k_proj.qweight": "model-00007-of-00057.safetensors", + "model.layers.12.self_attn.k_proj.qzeros": "model-00007-of-00057.safetensors", + "model.layers.12.self_attn.k_proj.scales": "model-00007-of-00057.safetensors", + "model.layers.12.self_attn.o_proj.g_idx": "model-00007-of-00057.safetensors", + "model.layers.12.self_attn.o_proj.qweight": "model-00007-of-00057.safetensors", + "model.layers.12.self_attn.o_proj.qzeros": "model-00007-of-00057.safetensors", + "model.layers.12.self_attn.o_proj.scales": "model-00007-of-00057.safetensors", + "model.layers.12.self_attn.q_proj.g_idx": "model-00007-of-00057.safetensors", + "model.layers.12.self_attn.q_proj.qweight": "model-00007-of-00057.safetensors", + "model.layers.12.self_attn.q_proj.qzeros": "model-00007-of-00057.safetensors", + "model.layers.12.self_attn.q_proj.scales": "model-00007-of-00057.safetensors", + "model.layers.12.self_attn.v_proj.g_idx": "model-00007-of-00057.safetensors", + "model.layers.12.self_attn.v_proj.qweight": "model-00007-of-00057.safetensors", + "model.layers.12.self_attn.v_proj.qzeros": "model-00007-of-00057.safetensors", + "model.layers.12.self_attn.v_proj.scales": "model-00007-of-00057.safetensors", + "model.layers.120.input_layernorm.weight": "model-00053-of-00057.safetensors", + "model.layers.120.mlp.down_proj.g_idx": "model-00053-of-00057.safetensors", + "model.layers.120.mlp.down_proj.qweight": "model-00053-of-00057.safetensors", + "model.layers.120.mlp.down_proj.qzeros": "model-00053-of-00057.safetensors", + "model.layers.120.mlp.down_proj.scales": "model-00053-of-00057.safetensors", + "model.layers.120.mlp.gate_proj.g_idx": "model-00053-of-00057.safetensors", + "model.layers.120.mlp.gate_proj.qweight": "model-00053-of-00057.safetensors", + "model.layers.120.mlp.gate_proj.qzeros": "model-00053-of-00057.safetensors", + "model.layers.120.mlp.gate_proj.scales": "model-00053-of-00057.safetensors", + "model.layers.120.mlp.up_proj.g_idx": "model-00053-of-00057.safetensors", + "model.layers.120.mlp.up_proj.qweight": "model-00053-of-00057.safetensors", + "model.layers.120.mlp.up_proj.qzeros": "model-00053-of-00057.safetensors", + "model.layers.120.mlp.up_proj.scales": "model-00053-of-00057.safetensors", + "model.layers.120.post_attention_layernorm.weight": "model-00053-of-00057.safetensors", + "model.layers.120.self_attn.k_proj.g_idx": "model-00053-of-00057.safetensors", + "model.layers.120.self_attn.k_proj.qweight": "model-00053-of-00057.safetensors", + "model.layers.120.self_attn.k_proj.qzeros": "model-00053-of-00057.safetensors", + "model.layers.120.self_attn.k_proj.scales": "model-00053-of-00057.safetensors", + "model.layers.120.self_attn.o_proj.g_idx": "model-00053-of-00057.safetensors", + "model.layers.120.self_attn.o_proj.qweight": "model-00053-of-00057.safetensors", + "model.layers.120.self_attn.o_proj.qzeros": "model-00053-of-00057.safetensors", + "model.layers.120.self_attn.o_proj.scales": "model-00053-of-00057.safetensors", + "model.layers.120.self_attn.q_proj.g_idx": "model-00053-of-00057.safetensors", + "model.layers.120.self_attn.q_proj.qweight": "model-00053-of-00057.safetensors", + "model.layers.120.self_attn.q_proj.qzeros": "model-00053-of-00057.safetensors", + "model.layers.120.self_attn.q_proj.scales": "model-00053-of-00057.safetensors", + "model.layers.120.self_attn.v_proj.g_idx": "model-00053-of-00057.safetensors", + "model.layers.120.self_attn.v_proj.qweight": "model-00053-of-00057.safetensors", + "model.layers.120.self_attn.v_proj.qzeros": "model-00053-of-00057.safetensors", + "model.layers.120.self_attn.v_proj.scales": "model-00053-of-00057.safetensors", + "model.layers.121.input_layernorm.weight": "model-00054-of-00057.safetensors", + "model.layers.121.mlp.down_proj.g_idx": "model-00054-of-00057.safetensors", + "model.layers.121.mlp.down_proj.qweight": "model-00054-of-00057.safetensors", + "model.layers.121.mlp.down_proj.qzeros": "model-00054-of-00057.safetensors", + "model.layers.121.mlp.down_proj.scales": "model-00054-of-00057.safetensors", + "model.layers.121.mlp.gate_proj.g_idx": "model-00054-of-00057.safetensors", + "model.layers.121.mlp.gate_proj.qweight": "model-00054-of-00057.safetensors", + "model.layers.121.mlp.gate_proj.qzeros": "model-00054-of-00057.safetensors", + "model.layers.121.mlp.gate_proj.scales": "model-00054-of-00057.safetensors", + "model.layers.121.mlp.up_proj.g_idx": "model-00054-of-00057.safetensors", + "model.layers.121.mlp.up_proj.qweight": "model-00054-of-00057.safetensors", + "model.layers.121.mlp.up_proj.qzeros": "model-00054-of-00057.safetensors", + "model.layers.121.mlp.up_proj.scales": "model-00054-of-00057.safetensors", + "model.layers.121.post_attention_layernorm.weight": "model-00054-of-00057.safetensors", + "model.layers.121.self_attn.k_proj.g_idx": "model-00053-of-00057.safetensors", + "model.layers.121.self_attn.k_proj.qweight": "model-00053-of-00057.safetensors", + "model.layers.121.self_attn.k_proj.qzeros": "model-00053-of-00057.safetensors", + "model.layers.121.self_attn.k_proj.scales": "model-00053-of-00057.safetensors", + "model.layers.121.self_attn.o_proj.g_idx": "model-00054-of-00057.safetensors", + "model.layers.121.self_attn.o_proj.qweight": "model-00054-of-00057.safetensors", + "model.layers.121.self_attn.o_proj.qzeros": "model-00054-of-00057.safetensors", + "model.layers.121.self_attn.o_proj.scales": "model-00054-of-00057.safetensors", + "model.layers.121.self_attn.q_proj.g_idx": "model-00053-of-00057.safetensors", + "model.layers.121.self_attn.q_proj.qweight": "model-00053-of-00057.safetensors", + "model.layers.121.self_attn.q_proj.qzeros": "model-00053-of-00057.safetensors", + "model.layers.121.self_attn.q_proj.scales": "model-00053-of-00057.safetensors", + "model.layers.121.self_attn.v_proj.g_idx": "model-00053-of-00057.safetensors", + "model.layers.121.self_attn.v_proj.qweight": "model-00053-of-00057.safetensors", + "model.layers.121.self_attn.v_proj.qzeros": "model-00053-of-00057.safetensors", + "model.layers.121.self_attn.v_proj.scales": "model-00053-of-00057.safetensors", + "model.layers.122.input_layernorm.weight": "model-00054-of-00057.safetensors", + "model.layers.122.mlp.down_proj.g_idx": "model-00054-of-00057.safetensors", + "model.layers.122.mlp.down_proj.qweight": "model-00054-of-00057.safetensors", + "model.layers.122.mlp.down_proj.qzeros": "model-00054-of-00057.safetensors", + "model.layers.122.mlp.down_proj.scales": "model-00054-of-00057.safetensors", + "model.layers.122.mlp.gate_proj.g_idx": "model-00054-of-00057.safetensors", + "model.layers.122.mlp.gate_proj.qweight": "model-00054-of-00057.safetensors", + "model.layers.122.mlp.gate_proj.qzeros": "model-00054-of-00057.safetensors", + "model.layers.122.mlp.gate_proj.scales": "model-00054-of-00057.safetensors", + "model.layers.122.mlp.up_proj.g_idx": "model-00054-of-00057.safetensors", + "model.layers.122.mlp.up_proj.qweight": "model-00054-of-00057.safetensors", + "model.layers.122.mlp.up_proj.qzeros": "model-00054-of-00057.safetensors", + "model.layers.122.mlp.up_proj.scales": "model-00054-of-00057.safetensors", + "model.layers.122.post_attention_layernorm.weight": "model-00054-of-00057.safetensors", + "model.layers.122.self_attn.k_proj.g_idx": "model-00054-of-00057.safetensors", + "model.layers.122.self_attn.k_proj.qweight": "model-00054-of-00057.safetensors", + "model.layers.122.self_attn.k_proj.qzeros": "model-00054-of-00057.safetensors", + "model.layers.122.self_attn.k_proj.scales": "model-00054-of-00057.safetensors", + "model.layers.122.self_attn.o_proj.g_idx": "model-00054-of-00057.safetensors", + "model.layers.122.self_attn.o_proj.qweight": "model-00054-of-00057.safetensors", + "model.layers.122.self_attn.o_proj.qzeros": "model-00054-of-00057.safetensors", + "model.layers.122.self_attn.o_proj.scales": "model-00054-of-00057.safetensors", + "model.layers.122.self_attn.q_proj.g_idx": "model-00054-of-00057.safetensors", + "model.layers.122.self_attn.q_proj.qweight": "model-00054-of-00057.safetensors", + "model.layers.122.self_attn.q_proj.qzeros": "model-00054-of-00057.safetensors", + "model.layers.122.self_attn.q_proj.scales": "model-00054-of-00057.safetensors", + "model.layers.122.self_attn.v_proj.g_idx": "model-00054-of-00057.safetensors", + "model.layers.122.self_attn.v_proj.qweight": "model-00054-of-00057.safetensors", + "model.layers.122.self_attn.v_proj.qzeros": "model-00054-of-00057.safetensors", + "model.layers.122.self_attn.v_proj.scales": "model-00054-of-00057.safetensors", + "model.layers.123.input_layernorm.weight": "model-00055-of-00057.safetensors", + "model.layers.123.mlp.down_proj.g_idx": "model-00055-of-00057.safetensors", + "model.layers.123.mlp.down_proj.qweight": "model-00055-of-00057.safetensors", + "model.layers.123.mlp.down_proj.qzeros": "model-00055-of-00057.safetensors", + "model.layers.123.mlp.down_proj.scales": "model-00055-of-00057.safetensors", + "model.layers.123.mlp.gate_proj.g_idx": "model-00054-of-00057.safetensors", + "model.layers.123.mlp.gate_proj.qweight": "model-00054-of-00057.safetensors", + "model.layers.123.mlp.gate_proj.qzeros": "model-00054-of-00057.safetensors", + "model.layers.123.mlp.gate_proj.scales": "model-00054-of-00057.safetensors", + "model.layers.123.mlp.up_proj.g_idx": "model-00055-of-00057.safetensors", + "model.layers.123.mlp.up_proj.qweight": "model-00055-of-00057.safetensors", + "model.layers.123.mlp.up_proj.qzeros": "model-00055-of-00057.safetensors", + "model.layers.123.mlp.up_proj.scales": "model-00055-of-00057.safetensors", + "model.layers.123.post_attention_layernorm.weight": "model-00055-of-00057.safetensors", + "model.layers.123.self_attn.k_proj.g_idx": "model-00054-of-00057.safetensors", + "model.layers.123.self_attn.k_proj.qweight": "model-00054-of-00057.safetensors", + "model.layers.123.self_attn.k_proj.qzeros": "model-00054-of-00057.safetensors", + "model.layers.123.self_attn.k_proj.scales": "model-00054-of-00057.safetensors", + "model.layers.123.self_attn.o_proj.g_idx": "model-00054-of-00057.safetensors", + "model.layers.123.self_attn.o_proj.qweight": "model-00054-of-00057.safetensors", + "model.layers.123.self_attn.o_proj.qzeros": "model-00054-of-00057.safetensors", + "model.layers.123.self_attn.o_proj.scales": "model-00054-of-00057.safetensors", + "model.layers.123.self_attn.q_proj.g_idx": "model-00054-of-00057.safetensors", + "model.layers.123.self_attn.q_proj.qweight": "model-00054-of-00057.safetensors", + "model.layers.123.self_attn.q_proj.qzeros": "model-00054-of-00057.safetensors", + "model.layers.123.self_attn.q_proj.scales": "model-00054-of-00057.safetensors", + "model.layers.123.self_attn.v_proj.g_idx": "model-00054-of-00057.safetensors", + "model.layers.123.self_attn.v_proj.qweight": "model-00054-of-00057.safetensors", + "model.layers.123.self_attn.v_proj.qzeros": "model-00054-of-00057.safetensors", + "model.layers.123.self_attn.v_proj.scales": "model-00054-of-00057.safetensors", + "model.layers.124.input_layernorm.weight": "model-00055-of-00057.safetensors", + "model.layers.124.mlp.down_proj.g_idx": "model-00055-of-00057.safetensors", + "model.layers.124.mlp.down_proj.qweight": "model-00055-of-00057.safetensors", + "model.layers.124.mlp.down_proj.qzeros": "model-00055-of-00057.safetensors", + "model.layers.124.mlp.down_proj.scales": "model-00055-of-00057.safetensors", + "model.layers.124.mlp.gate_proj.g_idx": "model-00055-of-00057.safetensors", + "model.layers.124.mlp.gate_proj.qweight": "model-00055-of-00057.safetensors", + "model.layers.124.mlp.gate_proj.qzeros": "model-00055-of-00057.safetensors", + "model.layers.124.mlp.gate_proj.scales": "model-00055-of-00057.safetensors", + "model.layers.124.mlp.up_proj.g_idx": "model-00055-of-00057.safetensors", + "model.layers.124.mlp.up_proj.qweight": "model-00055-of-00057.safetensors", + "model.layers.124.mlp.up_proj.qzeros": "model-00055-of-00057.safetensors", + "model.layers.124.mlp.up_proj.scales": "model-00055-of-00057.safetensors", + "model.layers.124.post_attention_layernorm.weight": "model-00055-of-00057.safetensors", + "model.layers.124.self_attn.k_proj.g_idx": "model-00055-of-00057.safetensors", + "model.layers.124.self_attn.k_proj.qweight": "model-00055-of-00057.safetensors", + "model.layers.124.self_attn.k_proj.qzeros": "model-00055-of-00057.safetensors", + "model.layers.124.self_attn.k_proj.scales": "model-00055-of-00057.safetensors", + "model.layers.124.self_attn.o_proj.g_idx": "model-00055-of-00057.safetensors", + "model.layers.124.self_attn.o_proj.qweight": "model-00055-of-00057.safetensors", + "model.layers.124.self_attn.o_proj.qzeros": "model-00055-of-00057.safetensors", + "model.layers.124.self_attn.o_proj.scales": "model-00055-of-00057.safetensors", + "model.layers.124.self_attn.q_proj.g_idx": "model-00055-of-00057.safetensors", + "model.layers.124.self_attn.q_proj.qweight": "model-00055-of-00057.safetensors", + "model.layers.124.self_attn.q_proj.qzeros": "model-00055-of-00057.safetensors", + "model.layers.124.self_attn.q_proj.scales": "model-00055-of-00057.safetensors", + "model.layers.124.self_attn.v_proj.g_idx": "model-00055-of-00057.safetensors", + "model.layers.124.self_attn.v_proj.qweight": "model-00055-of-00057.safetensors", + "model.layers.124.self_attn.v_proj.qzeros": "model-00055-of-00057.safetensors", + "model.layers.124.self_attn.v_proj.scales": "model-00055-of-00057.safetensors", + "model.layers.125.input_layernorm.weight": "model-00057-of-00057.safetensors", + "model.layers.125.mlp.down_proj.g_idx": "model-00057-of-00057.safetensors", + "model.layers.125.mlp.down_proj.qweight": "model-00057-of-00057.safetensors", + "model.layers.125.mlp.down_proj.qzeros": "model-00057-of-00057.safetensors", + "model.layers.125.mlp.down_proj.scales": "model-00057-of-00057.safetensors", + "model.layers.125.mlp.gate_proj.g_idx": "model-00055-of-00057.safetensors", + "model.layers.125.mlp.gate_proj.qweight": "model-00055-of-00057.safetensors", + "model.layers.125.mlp.gate_proj.qzeros": "model-00055-of-00057.safetensors", + "model.layers.125.mlp.gate_proj.scales": "model-00055-of-00057.safetensors", + "model.layers.125.mlp.up_proj.g_idx": "model-00055-of-00057.safetensors", + "model.layers.125.mlp.up_proj.qweight": "model-00055-of-00057.safetensors", + "model.layers.125.mlp.up_proj.qzeros": "model-00055-of-00057.safetensors", + "model.layers.125.mlp.up_proj.scales": "model-00055-of-00057.safetensors", + "model.layers.125.post_attention_layernorm.weight": "model-00057-of-00057.safetensors", + "model.layers.125.self_attn.k_proj.g_idx": "model-00055-of-00057.safetensors", + "model.layers.125.self_attn.k_proj.qweight": "model-00055-of-00057.safetensors", + "model.layers.125.self_attn.k_proj.qzeros": "model-00055-of-00057.safetensors", + "model.layers.125.self_attn.k_proj.scales": "model-00055-of-00057.safetensors", + "model.layers.125.self_attn.o_proj.g_idx": "model-00055-of-00057.safetensors", + "model.layers.125.self_attn.o_proj.qweight": "model-00055-of-00057.safetensors", + "model.layers.125.self_attn.o_proj.qzeros": "model-00055-of-00057.safetensors", + "model.layers.125.self_attn.o_proj.scales": "model-00055-of-00057.safetensors", + "model.layers.125.self_attn.q_proj.g_idx": "model-00055-of-00057.safetensors", + "model.layers.125.self_attn.q_proj.qweight": "model-00055-of-00057.safetensors", + "model.layers.125.self_attn.q_proj.qzeros": "model-00055-of-00057.safetensors", + "model.layers.125.self_attn.q_proj.scales": "model-00055-of-00057.safetensors", + "model.layers.125.self_attn.v_proj.g_idx": "model-00055-of-00057.safetensors", + "model.layers.125.self_attn.v_proj.qweight": "model-00055-of-00057.safetensors", + "model.layers.125.self_attn.v_proj.qzeros": "model-00055-of-00057.safetensors", + "model.layers.125.self_attn.v_proj.scales": "model-00055-of-00057.safetensors", + "model.layers.13.input_layernorm.weight": "model-00008-of-00057.safetensors", + "model.layers.13.mlp.down_proj.g_idx": "model-00008-of-00057.safetensors", + "model.layers.13.mlp.down_proj.qweight": "model-00008-of-00057.safetensors", + "model.layers.13.mlp.down_proj.qzeros": "model-00008-of-00057.safetensors", + "model.layers.13.mlp.down_proj.scales": "model-00008-of-00057.safetensors", + "model.layers.13.mlp.gate_proj.g_idx": "model-00007-of-00057.safetensors", + "model.layers.13.mlp.gate_proj.qweight": "model-00007-of-00057.safetensors", + "model.layers.13.mlp.gate_proj.qzeros": "model-00007-of-00057.safetensors", + "model.layers.13.mlp.gate_proj.scales": "model-00007-of-00057.safetensors", + "model.layers.13.mlp.up_proj.g_idx": "model-00007-of-00057.safetensors", + "model.layers.13.mlp.up_proj.qweight": "model-00007-of-00057.safetensors", + "model.layers.13.mlp.up_proj.qzeros": "model-00007-of-00057.safetensors", + "model.layers.13.mlp.up_proj.scales": "model-00007-of-00057.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00008-of-00057.safetensors", + "model.layers.13.self_attn.k_proj.g_idx": "model-00007-of-00057.safetensors", + "model.layers.13.self_attn.k_proj.qweight": "model-00007-of-00057.safetensors", + "model.layers.13.self_attn.k_proj.qzeros": "model-00007-of-00057.safetensors", + "model.layers.13.self_attn.k_proj.scales": "model-00007-of-00057.safetensors", + "model.layers.13.self_attn.o_proj.g_idx": "model-00007-of-00057.safetensors", + "model.layers.13.self_attn.o_proj.qweight": "model-00007-of-00057.safetensors", + "model.layers.13.self_attn.o_proj.qzeros": "model-00007-of-00057.safetensors", + "model.layers.13.self_attn.o_proj.scales": "model-00007-of-00057.safetensors", + "model.layers.13.self_attn.q_proj.g_idx": "model-00007-of-00057.safetensors", + "model.layers.13.self_attn.q_proj.qweight": "model-00007-of-00057.safetensors", + "model.layers.13.self_attn.q_proj.qzeros": "model-00007-of-00057.safetensors", + "model.layers.13.self_attn.q_proj.scales": "model-00007-of-00057.safetensors", + "model.layers.13.self_attn.v_proj.g_idx": "model-00007-of-00057.safetensors", + "model.layers.13.self_attn.v_proj.qweight": "model-00007-of-00057.safetensors", + "model.layers.13.self_attn.v_proj.qzeros": "model-00007-of-00057.safetensors", + "model.layers.13.self_attn.v_proj.scales": "model-00007-of-00057.safetensors", + "model.layers.14.input_layernorm.weight": "model-00008-of-00057.safetensors", + "model.layers.14.mlp.down_proj.g_idx": "model-00008-of-00057.safetensors", + "model.layers.14.mlp.down_proj.qweight": "model-00008-of-00057.safetensors", + "model.layers.14.mlp.down_proj.qzeros": "model-00008-of-00057.safetensors", + "model.layers.14.mlp.down_proj.scales": "model-00008-of-00057.safetensors", + "model.layers.14.mlp.gate_proj.g_idx": "model-00008-of-00057.safetensors", + "model.layers.14.mlp.gate_proj.qweight": "model-00008-of-00057.safetensors", + "model.layers.14.mlp.gate_proj.qzeros": "model-00008-of-00057.safetensors", + "model.layers.14.mlp.gate_proj.scales": "model-00008-of-00057.safetensors", + "model.layers.14.mlp.up_proj.g_idx": "model-00008-of-00057.safetensors", + "model.layers.14.mlp.up_proj.qweight": "model-00008-of-00057.safetensors", + "model.layers.14.mlp.up_proj.qzeros": "model-00008-of-00057.safetensors", + "model.layers.14.mlp.up_proj.scales": "model-00008-of-00057.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00008-of-00057.safetensors", + "model.layers.14.self_attn.k_proj.g_idx": "model-00008-of-00057.safetensors", + "model.layers.14.self_attn.k_proj.qweight": "model-00008-of-00057.safetensors", + "model.layers.14.self_attn.k_proj.qzeros": "model-00008-of-00057.safetensors", + "model.layers.14.self_attn.k_proj.scales": "model-00008-of-00057.safetensors", + "model.layers.14.self_attn.o_proj.g_idx": "model-00008-of-00057.safetensors", + "model.layers.14.self_attn.o_proj.qweight": "model-00008-of-00057.safetensors", + "model.layers.14.self_attn.o_proj.qzeros": "model-00008-of-00057.safetensors", + "model.layers.14.self_attn.o_proj.scales": "model-00008-of-00057.safetensors", + "model.layers.14.self_attn.q_proj.g_idx": "model-00008-of-00057.safetensors", + "model.layers.14.self_attn.q_proj.qweight": "model-00008-of-00057.safetensors", + "model.layers.14.self_attn.q_proj.qzeros": "model-00008-of-00057.safetensors", + "model.layers.14.self_attn.q_proj.scales": "model-00008-of-00057.safetensors", + "model.layers.14.self_attn.v_proj.g_idx": "model-00008-of-00057.safetensors", + "model.layers.14.self_attn.v_proj.qweight": "model-00008-of-00057.safetensors", + "model.layers.14.self_attn.v_proj.qzeros": "model-00008-of-00057.safetensors", + "model.layers.14.self_attn.v_proj.scales": "model-00008-of-00057.safetensors", + "model.layers.15.input_layernorm.weight": "model-00008-of-00057.safetensors", + "model.layers.15.mlp.down_proj.g_idx": "model-00008-of-00057.safetensors", + "model.layers.15.mlp.down_proj.qweight": "model-00008-of-00057.safetensors", + "model.layers.15.mlp.down_proj.qzeros": "model-00008-of-00057.safetensors", + "model.layers.15.mlp.down_proj.scales": "model-00008-of-00057.safetensors", + "model.layers.15.mlp.gate_proj.g_idx": "model-00008-of-00057.safetensors", + "model.layers.15.mlp.gate_proj.qweight": "model-00008-of-00057.safetensors", + "model.layers.15.mlp.gate_proj.qzeros": "model-00008-of-00057.safetensors", + "model.layers.15.mlp.gate_proj.scales": "model-00008-of-00057.safetensors", + "model.layers.15.mlp.up_proj.g_idx": "model-00008-of-00057.safetensors", + "model.layers.15.mlp.up_proj.qweight": "model-00008-of-00057.safetensors", + "model.layers.15.mlp.up_proj.qzeros": "model-00008-of-00057.safetensors", + "model.layers.15.mlp.up_proj.scales": "model-00008-of-00057.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00008-of-00057.safetensors", + "model.layers.15.self_attn.k_proj.g_idx": "model-00008-of-00057.safetensors", + "model.layers.15.self_attn.k_proj.qweight": "model-00008-of-00057.safetensors", + "model.layers.15.self_attn.k_proj.qzeros": "model-00008-of-00057.safetensors", + "model.layers.15.self_attn.k_proj.scales": "model-00008-of-00057.safetensors", + "model.layers.15.self_attn.o_proj.g_idx": "model-00008-of-00057.safetensors", + "model.layers.15.self_attn.o_proj.qweight": "model-00008-of-00057.safetensors", + "model.layers.15.self_attn.o_proj.qzeros": "model-00008-of-00057.safetensors", + "model.layers.15.self_attn.o_proj.scales": "model-00008-of-00057.safetensors", + "model.layers.15.self_attn.q_proj.g_idx": "model-00008-of-00057.safetensors", + "model.layers.15.self_attn.q_proj.qweight": "model-00008-of-00057.safetensors", + "model.layers.15.self_attn.q_proj.qzeros": "model-00008-of-00057.safetensors", + "model.layers.15.self_attn.q_proj.scales": "model-00008-of-00057.safetensors", + "model.layers.15.self_attn.v_proj.g_idx": "model-00008-of-00057.safetensors", + "model.layers.15.self_attn.v_proj.qweight": "model-00008-of-00057.safetensors", + "model.layers.15.self_attn.v_proj.qzeros": "model-00008-of-00057.safetensors", + "model.layers.15.self_attn.v_proj.scales": "model-00008-of-00057.safetensors", + "model.layers.16.input_layernorm.weight": "model-00009-of-00057.safetensors", + "model.layers.16.mlp.down_proj.g_idx": "model-00009-of-00057.safetensors", + "model.layers.16.mlp.down_proj.qweight": "model-00009-of-00057.safetensors", + "model.layers.16.mlp.down_proj.qzeros": "model-00009-of-00057.safetensors", + "model.layers.16.mlp.down_proj.scales": "model-00009-of-00057.safetensors", + "model.layers.16.mlp.gate_proj.g_idx": "model-00009-of-00057.safetensors", + "model.layers.16.mlp.gate_proj.qweight": "model-00009-of-00057.safetensors", + "model.layers.16.mlp.gate_proj.qzeros": "model-00009-of-00057.safetensors", + "model.layers.16.mlp.gate_proj.scales": "model-00009-of-00057.safetensors", + "model.layers.16.mlp.up_proj.g_idx": "model-00009-of-00057.safetensors", + "model.layers.16.mlp.up_proj.qweight": "model-00009-of-00057.safetensors", + "model.layers.16.mlp.up_proj.qzeros": "model-00009-of-00057.safetensors", + "model.layers.16.mlp.up_proj.scales": "model-00009-of-00057.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00009-of-00057.safetensors", + "model.layers.16.self_attn.k_proj.g_idx": "model-00008-of-00057.safetensors", + "model.layers.16.self_attn.k_proj.qweight": "model-00008-of-00057.safetensors", + "model.layers.16.self_attn.k_proj.qzeros": "model-00008-of-00057.safetensors", + "model.layers.16.self_attn.k_proj.scales": "model-00008-of-00057.safetensors", + "model.layers.16.self_attn.o_proj.g_idx": "model-00009-of-00057.safetensors", + "model.layers.16.self_attn.o_proj.qweight": "model-00009-of-00057.safetensors", + "model.layers.16.self_attn.o_proj.qzeros": "model-00009-of-00057.safetensors", + "model.layers.16.self_attn.o_proj.scales": "model-00009-of-00057.safetensors", + "model.layers.16.self_attn.q_proj.g_idx": "model-00008-of-00057.safetensors", + "model.layers.16.self_attn.q_proj.qweight": "model-00008-of-00057.safetensors", + "model.layers.16.self_attn.q_proj.qzeros": "model-00008-of-00057.safetensors", + "model.layers.16.self_attn.q_proj.scales": "model-00008-of-00057.safetensors", + "model.layers.16.self_attn.v_proj.g_idx": "model-00008-of-00057.safetensors", + "model.layers.16.self_attn.v_proj.qweight": "model-00008-of-00057.safetensors", + "model.layers.16.self_attn.v_proj.qzeros": "model-00008-of-00057.safetensors", + "model.layers.16.self_attn.v_proj.scales": "model-00008-of-00057.safetensors", + "model.layers.17.input_layernorm.weight": "model-00009-of-00057.safetensors", + "model.layers.17.mlp.down_proj.g_idx": "model-00009-of-00057.safetensors", + "model.layers.17.mlp.down_proj.qweight": "model-00009-of-00057.safetensors", + "model.layers.17.mlp.down_proj.qzeros": "model-00009-of-00057.safetensors", + "model.layers.17.mlp.down_proj.scales": "model-00009-of-00057.safetensors", + "model.layers.17.mlp.gate_proj.g_idx": "model-00009-of-00057.safetensors", + "model.layers.17.mlp.gate_proj.qweight": "model-00009-of-00057.safetensors", + "model.layers.17.mlp.gate_proj.qzeros": "model-00009-of-00057.safetensors", + "model.layers.17.mlp.gate_proj.scales": "model-00009-of-00057.safetensors", + "model.layers.17.mlp.up_proj.g_idx": "model-00009-of-00057.safetensors", + "model.layers.17.mlp.up_proj.qweight": "model-00009-of-00057.safetensors", + "model.layers.17.mlp.up_proj.qzeros": "model-00009-of-00057.safetensors", + "model.layers.17.mlp.up_proj.scales": "model-00009-of-00057.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00009-of-00057.safetensors", + "model.layers.17.self_attn.k_proj.g_idx": "model-00009-of-00057.safetensors", + "model.layers.17.self_attn.k_proj.qweight": "model-00009-of-00057.safetensors", + "model.layers.17.self_attn.k_proj.qzeros": "model-00009-of-00057.safetensors", + "model.layers.17.self_attn.k_proj.scales": "model-00009-of-00057.safetensors", + "model.layers.17.self_attn.o_proj.g_idx": "model-00009-of-00057.safetensors", + "model.layers.17.self_attn.o_proj.qweight": "model-00009-of-00057.safetensors", + "model.layers.17.self_attn.o_proj.qzeros": "model-00009-of-00057.safetensors", + "model.layers.17.self_attn.o_proj.scales": "model-00009-of-00057.safetensors", + "model.layers.17.self_attn.q_proj.g_idx": "model-00009-of-00057.safetensors", + "model.layers.17.self_attn.q_proj.qweight": "model-00009-of-00057.safetensors", + "model.layers.17.self_attn.q_proj.qzeros": "model-00009-of-00057.safetensors", + "model.layers.17.self_attn.q_proj.scales": "model-00009-of-00057.safetensors", + "model.layers.17.self_attn.v_proj.g_idx": "model-00009-of-00057.safetensors", + "model.layers.17.self_attn.v_proj.qweight": "model-00009-of-00057.safetensors", + "model.layers.17.self_attn.v_proj.qzeros": "model-00009-of-00057.safetensors", + "model.layers.17.self_attn.v_proj.scales": "model-00009-of-00057.safetensors", + "model.layers.18.input_layernorm.weight": "model-00010-of-00057.safetensors", + "model.layers.18.mlp.down_proj.g_idx": "model-00010-of-00057.safetensors", + "model.layers.18.mlp.down_proj.qweight": "model-00010-of-00057.safetensors", + "model.layers.18.mlp.down_proj.qzeros": "model-00010-of-00057.safetensors", + "model.layers.18.mlp.down_proj.scales": "model-00010-of-00057.safetensors", + "model.layers.18.mlp.gate_proj.g_idx": "model-00009-of-00057.safetensors", + "model.layers.18.mlp.gate_proj.qweight": "model-00009-of-00057.safetensors", + "model.layers.18.mlp.gate_proj.qzeros": "model-00009-of-00057.safetensors", + "model.layers.18.mlp.gate_proj.scales": "model-00009-of-00057.safetensors", + "model.layers.18.mlp.up_proj.g_idx": "model-00010-of-00057.safetensors", + "model.layers.18.mlp.up_proj.qweight": "model-00010-of-00057.safetensors", + "model.layers.18.mlp.up_proj.qzeros": "model-00010-of-00057.safetensors", + "model.layers.18.mlp.up_proj.scales": "model-00010-of-00057.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00010-of-00057.safetensors", + "model.layers.18.self_attn.k_proj.g_idx": "model-00009-of-00057.safetensors", + "model.layers.18.self_attn.k_proj.qweight": "model-00009-of-00057.safetensors", + "model.layers.18.self_attn.k_proj.qzeros": "model-00009-of-00057.safetensors", + "model.layers.18.self_attn.k_proj.scales": "model-00009-of-00057.safetensors", + "model.layers.18.self_attn.o_proj.g_idx": "model-00009-of-00057.safetensors", + "model.layers.18.self_attn.o_proj.qweight": "model-00009-of-00057.safetensors", + "model.layers.18.self_attn.o_proj.qzeros": "model-00009-of-00057.safetensors", + "model.layers.18.self_attn.o_proj.scales": "model-00009-of-00057.safetensors", + "model.layers.18.self_attn.q_proj.g_idx": "model-00009-of-00057.safetensors", + "model.layers.18.self_attn.q_proj.qweight": "model-00009-of-00057.safetensors", + "model.layers.18.self_attn.q_proj.qzeros": "model-00009-of-00057.safetensors", + "model.layers.18.self_attn.q_proj.scales": "model-00009-of-00057.safetensors", + "model.layers.18.self_attn.v_proj.g_idx": "model-00009-of-00057.safetensors", + "model.layers.18.self_attn.v_proj.qweight": "model-00009-of-00057.safetensors", + "model.layers.18.self_attn.v_proj.qzeros": "model-00009-of-00057.safetensors", + "model.layers.18.self_attn.v_proj.scales": "model-00009-of-00057.safetensors", + "model.layers.19.input_layernorm.weight": "model-00010-of-00057.safetensors", + "model.layers.19.mlp.down_proj.g_idx": "model-00010-of-00057.safetensors", + "model.layers.19.mlp.down_proj.qweight": "model-00010-of-00057.safetensors", + "model.layers.19.mlp.down_proj.qzeros": "model-00010-of-00057.safetensors", + "model.layers.19.mlp.down_proj.scales": "model-00010-of-00057.safetensors", + "model.layers.19.mlp.gate_proj.g_idx": "model-00010-of-00057.safetensors", + "model.layers.19.mlp.gate_proj.qweight": "model-00010-of-00057.safetensors", + "model.layers.19.mlp.gate_proj.qzeros": "model-00010-of-00057.safetensors", + "model.layers.19.mlp.gate_proj.scales": "model-00010-of-00057.safetensors", + "model.layers.19.mlp.up_proj.g_idx": "model-00010-of-00057.safetensors", + "model.layers.19.mlp.up_proj.qweight": "model-00010-of-00057.safetensors", + "model.layers.19.mlp.up_proj.qzeros": "model-00010-of-00057.safetensors", + "model.layers.19.mlp.up_proj.scales": "model-00010-of-00057.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00010-of-00057.safetensors", + "model.layers.19.self_attn.k_proj.g_idx": "model-00010-of-00057.safetensors", + "model.layers.19.self_attn.k_proj.qweight": "model-00010-of-00057.safetensors", + "model.layers.19.self_attn.k_proj.qzeros": "model-00010-of-00057.safetensors", + "model.layers.19.self_attn.k_proj.scales": "model-00010-of-00057.safetensors", + "model.layers.19.self_attn.o_proj.g_idx": "model-00010-of-00057.safetensors", + "model.layers.19.self_attn.o_proj.qweight": "model-00010-of-00057.safetensors", + "model.layers.19.self_attn.o_proj.qzeros": "model-00010-of-00057.safetensors", + "model.layers.19.self_attn.o_proj.scales": "model-00010-of-00057.safetensors", + "model.layers.19.self_attn.q_proj.g_idx": "model-00010-of-00057.safetensors", + "model.layers.19.self_attn.q_proj.qweight": "model-00010-of-00057.safetensors", + "model.layers.19.self_attn.q_proj.qzeros": "model-00010-of-00057.safetensors", + "model.layers.19.self_attn.q_proj.scales": "model-00010-of-00057.safetensors", + "model.layers.19.self_attn.v_proj.g_idx": "model-00010-of-00057.safetensors", + "model.layers.19.self_attn.v_proj.qweight": "model-00010-of-00057.safetensors", + "model.layers.19.self_attn.v_proj.qzeros": "model-00010-of-00057.safetensors", + "model.layers.19.self_attn.v_proj.scales": "model-00010-of-00057.safetensors", + "model.layers.2.input_layernorm.weight": "model-00003-of-00057.safetensors", + "model.layers.2.mlp.down_proj.g_idx": "model-00003-of-00057.safetensors", + "model.layers.2.mlp.down_proj.qweight": "model-00003-of-00057.safetensors", + "model.layers.2.mlp.down_proj.qzeros": "model-00003-of-00057.safetensors", + "model.layers.2.mlp.down_proj.scales": "model-00003-of-00057.safetensors", + "model.layers.2.mlp.gate_proj.g_idx": "model-00003-of-00057.safetensors", + "model.layers.2.mlp.gate_proj.qweight": "model-00003-of-00057.safetensors", + "model.layers.2.mlp.gate_proj.qzeros": "model-00003-of-00057.safetensors", + "model.layers.2.mlp.gate_proj.scales": "model-00003-of-00057.safetensors", + "model.layers.2.mlp.up_proj.g_idx": "model-00003-of-00057.safetensors", + "model.layers.2.mlp.up_proj.qweight": "model-00003-of-00057.safetensors", + "model.layers.2.mlp.up_proj.qzeros": "model-00003-of-00057.safetensors", + "model.layers.2.mlp.up_proj.scales": "model-00003-of-00057.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00003-of-00057.safetensors", + "model.layers.2.self_attn.k_proj.g_idx": "model-00002-of-00057.safetensors", + "model.layers.2.self_attn.k_proj.qweight": "model-00002-of-00057.safetensors", + "model.layers.2.self_attn.k_proj.qzeros": "model-00002-of-00057.safetensors", + "model.layers.2.self_attn.k_proj.scales": "model-00002-of-00057.safetensors", + "model.layers.2.self_attn.o_proj.g_idx": "model-00002-of-00057.safetensors", + "model.layers.2.self_attn.o_proj.qweight": "model-00002-of-00057.safetensors", + "model.layers.2.self_attn.o_proj.qzeros": "model-00002-of-00057.safetensors", + "model.layers.2.self_attn.o_proj.scales": "model-00002-of-00057.safetensors", + "model.layers.2.self_attn.q_proj.g_idx": "model-00002-of-00057.safetensors", + "model.layers.2.self_attn.q_proj.qweight": "model-00002-of-00057.safetensors", + "model.layers.2.self_attn.q_proj.qzeros": "model-00002-of-00057.safetensors", + "model.layers.2.self_attn.q_proj.scales": "model-00002-of-00057.safetensors", + "model.layers.2.self_attn.v_proj.g_idx": "model-00002-of-00057.safetensors", + "model.layers.2.self_attn.v_proj.qweight": "model-00002-of-00057.safetensors", + "model.layers.2.self_attn.v_proj.qzeros": "model-00002-of-00057.safetensors", + "model.layers.2.self_attn.v_proj.scales": "model-00002-of-00057.safetensors", + "model.layers.20.input_layernorm.weight": "model-00011-of-00057.safetensors", + "model.layers.20.mlp.down_proj.g_idx": "model-00011-of-00057.safetensors", + "model.layers.20.mlp.down_proj.qweight": "model-00011-of-00057.safetensors", + "model.layers.20.mlp.down_proj.qzeros": "model-00011-of-00057.safetensors", + "model.layers.20.mlp.down_proj.scales": "model-00011-of-00057.safetensors", + "model.layers.20.mlp.gate_proj.g_idx": "model-00010-of-00057.safetensors", + "model.layers.20.mlp.gate_proj.qweight": "model-00010-of-00057.safetensors", + "model.layers.20.mlp.gate_proj.qzeros": "model-00010-of-00057.safetensors", + "model.layers.20.mlp.gate_proj.scales": "model-00010-of-00057.safetensors", + "model.layers.20.mlp.up_proj.g_idx": "model-00010-of-00057.safetensors", + "model.layers.20.mlp.up_proj.qweight": "model-00010-of-00057.safetensors", + "model.layers.20.mlp.up_proj.qzeros": "model-00010-of-00057.safetensors", + "model.layers.20.mlp.up_proj.scales": "model-00010-of-00057.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00011-of-00057.safetensors", + "model.layers.20.self_attn.k_proj.g_idx": "model-00010-of-00057.safetensors", + "model.layers.20.self_attn.k_proj.qweight": "model-00010-of-00057.safetensors", + "model.layers.20.self_attn.k_proj.qzeros": "model-00010-of-00057.safetensors", + "model.layers.20.self_attn.k_proj.scales": "model-00010-of-00057.safetensors", + "model.layers.20.self_attn.o_proj.g_idx": "model-00010-of-00057.safetensors", + "model.layers.20.self_attn.o_proj.qweight": "model-00010-of-00057.safetensors", + "model.layers.20.self_attn.o_proj.qzeros": "model-00010-of-00057.safetensors", + "model.layers.20.self_attn.o_proj.scales": "model-00010-of-00057.safetensors", + "model.layers.20.self_attn.q_proj.g_idx": "model-00010-of-00057.safetensors", + "model.layers.20.self_attn.q_proj.qweight": "model-00010-of-00057.safetensors", + "model.layers.20.self_attn.q_proj.qzeros": "model-00010-of-00057.safetensors", + "model.layers.20.self_attn.q_proj.scales": "model-00010-of-00057.safetensors", + "model.layers.20.self_attn.v_proj.g_idx": "model-00010-of-00057.safetensors", + "model.layers.20.self_attn.v_proj.qweight": "model-00010-of-00057.safetensors", + "model.layers.20.self_attn.v_proj.qzeros": "model-00010-of-00057.safetensors", + "model.layers.20.self_attn.v_proj.scales": "model-00010-of-00057.safetensors", + "model.layers.21.input_layernorm.weight": "model-00011-of-00057.safetensors", + "model.layers.21.mlp.down_proj.g_idx": "model-00011-of-00057.safetensors", + "model.layers.21.mlp.down_proj.qweight": "model-00011-of-00057.safetensors", + "model.layers.21.mlp.down_proj.qzeros": "model-00011-of-00057.safetensors", + "model.layers.21.mlp.down_proj.scales": "model-00011-of-00057.safetensors", + "model.layers.21.mlp.gate_proj.g_idx": "model-00011-of-00057.safetensors", + "model.layers.21.mlp.gate_proj.qweight": "model-00011-of-00057.safetensors", + "model.layers.21.mlp.gate_proj.qzeros": "model-00011-of-00057.safetensors", + "model.layers.21.mlp.gate_proj.scales": "model-00011-of-00057.safetensors", + "model.layers.21.mlp.up_proj.g_idx": "model-00011-of-00057.safetensors", + "model.layers.21.mlp.up_proj.qweight": "model-00011-of-00057.safetensors", + "model.layers.21.mlp.up_proj.qzeros": "model-00011-of-00057.safetensors", + "model.layers.21.mlp.up_proj.scales": "model-00011-of-00057.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00011-of-00057.safetensors", + "model.layers.21.self_attn.k_proj.g_idx": "model-00011-of-00057.safetensors", + "model.layers.21.self_attn.k_proj.qweight": "model-00011-of-00057.safetensors", + "model.layers.21.self_attn.k_proj.qzeros": "model-00011-of-00057.safetensors", + "model.layers.21.self_attn.k_proj.scales": "model-00011-of-00057.safetensors", + "model.layers.21.self_attn.o_proj.g_idx": "model-00011-of-00057.safetensors", + "model.layers.21.self_attn.o_proj.qweight": "model-00011-of-00057.safetensors", + "model.layers.21.self_attn.o_proj.qzeros": "model-00011-of-00057.safetensors", + "model.layers.21.self_attn.o_proj.scales": "model-00011-of-00057.safetensors", + "model.layers.21.self_attn.q_proj.g_idx": "model-00011-of-00057.safetensors", + "model.layers.21.self_attn.q_proj.qweight": "model-00011-of-00057.safetensors", + "model.layers.21.self_attn.q_proj.qzeros": "model-00011-of-00057.safetensors", + "model.layers.21.self_attn.q_proj.scales": "model-00011-of-00057.safetensors", + "model.layers.21.self_attn.v_proj.g_idx": "model-00011-of-00057.safetensors", + "model.layers.21.self_attn.v_proj.qweight": "model-00011-of-00057.safetensors", + "model.layers.21.self_attn.v_proj.qzeros": "model-00011-of-00057.safetensors", + "model.layers.21.self_attn.v_proj.scales": "model-00011-of-00057.safetensors", + "model.layers.22.input_layernorm.weight": "model-00011-of-00057.safetensors", + "model.layers.22.mlp.down_proj.g_idx": "model-00011-of-00057.safetensors", + "model.layers.22.mlp.down_proj.qweight": "model-00011-of-00057.safetensors", + "model.layers.22.mlp.down_proj.qzeros": "model-00011-of-00057.safetensors", + "model.layers.22.mlp.down_proj.scales": "model-00011-of-00057.safetensors", + "model.layers.22.mlp.gate_proj.g_idx": "model-00011-of-00057.safetensors", + "model.layers.22.mlp.gate_proj.qweight": "model-00011-of-00057.safetensors", + "model.layers.22.mlp.gate_proj.qzeros": "model-00011-of-00057.safetensors", + "model.layers.22.mlp.gate_proj.scales": "model-00011-of-00057.safetensors", + "model.layers.22.mlp.up_proj.g_idx": "model-00011-of-00057.safetensors", + "model.layers.22.mlp.up_proj.qweight": "model-00011-of-00057.safetensors", + "model.layers.22.mlp.up_proj.qzeros": "model-00011-of-00057.safetensors", + "model.layers.22.mlp.up_proj.scales": "model-00011-of-00057.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00011-of-00057.safetensors", + "model.layers.22.self_attn.k_proj.g_idx": "model-00011-of-00057.safetensors", + "model.layers.22.self_attn.k_proj.qweight": "model-00011-of-00057.safetensors", + "model.layers.22.self_attn.k_proj.qzeros": "model-00011-of-00057.safetensors", + "model.layers.22.self_attn.k_proj.scales": "model-00011-of-00057.safetensors", + "model.layers.22.self_attn.o_proj.g_idx": "model-00011-of-00057.safetensors", + "model.layers.22.self_attn.o_proj.qweight": "model-00011-of-00057.safetensors", + "model.layers.22.self_attn.o_proj.qzeros": "model-00011-of-00057.safetensors", + "model.layers.22.self_attn.o_proj.scales": "model-00011-of-00057.safetensors", + "model.layers.22.self_attn.q_proj.g_idx": "model-00011-of-00057.safetensors", + "model.layers.22.self_attn.q_proj.qweight": "model-00011-of-00057.safetensors", + "model.layers.22.self_attn.q_proj.qzeros": "model-00011-of-00057.safetensors", + "model.layers.22.self_attn.q_proj.scales": "model-00011-of-00057.safetensors", + "model.layers.22.self_attn.v_proj.g_idx": "model-00011-of-00057.safetensors", + "model.layers.22.self_attn.v_proj.qweight": "model-00011-of-00057.safetensors", + "model.layers.22.self_attn.v_proj.qzeros": "model-00011-of-00057.safetensors", + "model.layers.22.self_attn.v_proj.scales": "model-00011-of-00057.safetensors", + "model.layers.23.input_layernorm.weight": "model-00012-of-00057.safetensors", + "model.layers.23.mlp.down_proj.g_idx": "model-00012-of-00057.safetensors", + "model.layers.23.mlp.down_proj.qweight": "model-00012-of-00057.safetensors", + "model.layers.23.mlp.down_proj.qzeros": "model-00012-of-00057.safetensors", + "model.layers.23.mlp.down_proj.scales": "model-00012-of-00057.safetensors", + "model.layers.23.mlp.gate_proj.g_idx": "model-00012-of-00057.safetensors", + "model.layers.23.mlp.gate_proj.qweight": "model-00012-of-00057.safetensors", + "model.layers.23.mlp.gate_proj.qzeros": "model-00012-of-00057.safetensors", + "model.layers.23.mlp.gate_proj.scales": "model-00012-of-00057.safetensors", + "model.layers.23.mlp.up_proj.g_idx": "model-00012-of-00057.safetensors", + "model.layers.23.mlp.up_proj.qweight": "model-00012-of-00057.safetensors", + "model.layers.23.mlp.up_proj.qzeros": "model-00012-of-00057.safetensors", + "model.layers.23.mlp.up_proj.scales": "model-00012-of-00057.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00012-of-00057.safetensors", + "model.layers.23.self_attn.k_proj.g_idx": "model-00011-of-00057.safetensors", + "model.layers.23.self_attn.k_proj.qweight": "model-00011-of-00057.safetensors", + "model.layers.23.self_attn.k_proj.qzeros": "model-00011-of-00057.safetensors", + "model.layers.23.self_attn.k_proj.scales": "model-00011-of-00057.safetensors", + "model.layers.23.self_attn.o_proj.g_idx": "model-00012-of-00057.safetensors", + "model.layers.23.self_attn.o_proj.qweight": "model-00012-of-00057.safetensors", + "model.layers.23.self_attn.o_proj.qzeros": "model-00012-of-00057.safetensors", + "model.layers.23.self_attn.o_proj.scales": "model-00012-of-00057.safetensors", + "model.layers.23.self_attn.q_proj.g_idx": "model-00011-of-00057.safetensors", + "model.layers.23.self_attn.q_proj.qweight": "model-00011-of-00057.safetensors", + "model.layers.23.self_attn.q_proj.qzeros": "model-00011-of-00057.safetensors", + "model.layers.23.self_attn.q_proj.scales": "model-00011-of-00057.safetensors", + "model.layers.23.self_attn.v_proj.g_idx": "model-00011-of-00057.safetensors", + "model.layers.23.self_attn.v_proj.qweight": "model-00011-of-00057.safetensors", + "model.layers.23.self_attn.v_proj.qzeros": "model-00011-of-00057.safetensors", + "model.layers.23.self_attn.v_proj.scales": "model-00011-of-00057.safetensors", + "model.layers.24.input_layernorm.weight": "model-00012-of-00057.safetensors", + "model.layers.24.mlp.down_proj.g_idx": "model-00012-of-00057.safetensors", + "model.layers.24.mlp.down_proj.qweight": "model-00012-of-00057.safetensors", + "model.layers.24.mlp.down_proj.qzeros": "model-00012-of-00057.safetensors", + "model.layers.24.mlp.down_proj.scales": "model-00012-of-00057.safetensors", + "model.layers.24.mlp.gate_proj.g_idx": "model-00012-of-00057.safetensors", + "model.layers.24.mlp.gate_proj.qweight": "model-00012-of-00057.safetensors", + "model.layers.24.mlp.gate_proj.qzeros": "model-00012-of-00057.safetensors", + "model.layers.24.mlp.gate_proj.scales": "model-00012-of-00057.safetensors", + "model.layers.24.mlp.up_proj.g_idx": "model-00012-of-00057.safetensors", + "model.layers.24.mlp.up_proj.qweight": "model-00012-of-00057.safetensors", + "model.layers.24.mlp.up_proj.qzeros": "model-00012-of-00057.safetensors", + "model.layers.24.mlp.up_proj.scales": "model-00012-of-00057.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00012-of-00057.safetensors", + "model.layers.24.self_attn.k_proj.g_idx": "model-00012-of-00057.safetensors", + "model.layers.24.self_attn.k_proj.qweight": "model-00012-of-00057.safetensors", + "model.layers.24.self_attn.k_proj.qzeros": "model-00012-of-00057.safetensors", + "model.layers.24.self_attn.k_proj.scales": "model-00012-of-00057.safetensors", + "model.layers.24.self_attn.o_proj.g_idx": "model-00012-of-00057.safetensors", + "model.layers.24.self_attn.o_proj.qweight": "model-00012-of-00057.safetensors", + "model.layers.24.self_attn.o_proj.qzeros": "model-00012-of-00057.safetensors", + "model.layers.24.self_attn.o_proj.scales": "model-00012-of-00057.safetensors", + "model.layers.24.self_attn.q_proj.g_idx": "model-00012-of-00057.safetensors", + "model.layers.24.self_attn.q_proj.qweight": "model-00012-of-00057.safetensors", + "model.layers.24.self_attn.q_proj.qzeros": "model-00012-of-00057.safetensors", + "model.layers.24.self_attn.q_proj.scales": "model-00012-of-00057.safetensors", + "model.layers.24.self_attn.v_proj.g_idx": "model-00012-of-00057.safetensors", + "model.layers.24.self_attn.v_proj.qweight": "model-00012-of-00057.safetensors", + "model.layers.24.self_attn.v_proj.qzeros": "model-00012-of-00057.safetensors", + "model.layers.24.self_attn.v_proj.scales": "model-00012-of-00057.safetensors", + "model.layers.25.input_layernorm.weight": "model-00013-of-00057.safetensors", + "model.layers.25.mlp.down_proj.g_idx": "model-00013-of-00057.safetensors", + "model.layers.25.mlp.down_proj.qweight": "model-00013-of-00057.safetensors", + "model.layers.25.mlp.down_proj.qzeros": "model-00013-of-00057.safetensors", + "model.layers.25.mlp.down_proj.scales": "model-00013-of-00057.safetensors", + "model.layers.25.mlp.gate_proj.g_idx": "model-00012-of-00057.safetensors", + "model.layers.25.mlp.gate_proj.qweight": "model-00012-of-00057.safetensors", + "model.layers.25.mlp.gate_proj.qzeros": "model-00012-of-00057.safetensors", + "model.layers.25.mlp.gate_proj.scales": "model-00012-of-00057.safetensors", + "model.layers.25.mlp.up_proj.g_idx": "model-00013-of-00057.safetensors", + "model.layers.25.mlp.up_proj.qweight": "model-00013-of-00057.safetensors", + "model.layers.25.mlp.up_proj.qzeros": "model-00013-of-00057.safetensors", + "model.layers.25.mlp.up_proj.scales": "model-00013-of-00057.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00013-of-00057.safetensors", + "model.layers.25.self_attn.k_proj.g_idx": "model-00012-of-00057.safetensors", + "model.layers.25.self_attn.k_proj.qweight": "model-00012-of-00057.safetensors", + "model.layers.25.self_attn.k_proj.qzeros": "model-00012-of-00057.safetensors", + "model.layers.25.self_attn.k_proj.scales": "model-00012-of-00057.safetensors", + "model.layers.25.self_attn.o_proj.g_idx": "model-00012-of-00057.safetensors", + "model.layers.25.self_attn.o_proj.qweight": "model-00012-of-00057.safetensors", + "model.layers.25.self_attn.o_proj.qzeros": "model-00012-of-00057.safetensors", + "model.layers.25.self_attn.o_proj.scales": "model-00012-of-00057.safetensors", + "model.layers.25.self_attn.q_proj.g_idx": "model-00012-of-00057.safetensors", + "model.layers.25.self_attn.q_proj.qweight": "model-00012-of-00057.safetensors", + "model.layers.25.self_attn.q_proj.qzeros": "model-00012-of-00057.safetensors", + "model.layers.25.self_attn.q_proj.scales": "model-00012-of-00057.safetensors", + "model.layers.25.self_attn.v_proj.g_idx": "model-00012-of-00057.safetensors", + "model.layers.25.self_attn.v_proj.qweight": "model-00012-of-00057.safetensors", + "model.layers.25.self_attn.v_proj.qzeros": "model-00012-of-00057.safetensors", + "model.layers.25.self_attn.v_proj.scales": "model-00012-of-00057.safetensors", + "model.layers.26.input_layernorm.weight": "model-00013-of-00057.safetensors", + "model.layers.26.mlp.down_proj.g_idx": "model-00013-of-00057.safetensors", + "model.layers.26.mlp.down_proj.qweight": "model-00013-of-00057.safetensors", + "model.layers.26.mlp.down_proj.qzeros": "model-00013-of-00057.safetensors", + "model.layers.26.mlp.down_proj.scales": "model-00013-of-00057.safetensors", + "model.layers.26.mlp.gate_proj.g_idx": "model-00013-of-00057.safetensors", + "model.layers.26.mlp.gate_proj.qweight": "model-00013-of-00057.safetensors", + "model.layers.26.mlp.gate_proj.qzeros": "model-00013-of-00057.safetensors", + "model.layers.26.mlp.gate_proj.scales": "model-00013-of-00057.safetensors", + "model.layers.26.mlp.up_proj.g_idx": "model-00013-of-00057.safetensors", + "model.layers.26.mlp.up_proj.qweight": "model-00013-of-00057.safetensors", + "model.layers.26.mlp.up_proj.qzeros": "model-00013-of-00057.safetensors", + "model.layers.26.mlp.up_proj.scales": "model-00013-of-00057.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00013-of-00057.safetensors", + "model.layers.26.self_attn.k_proj.g_idx": "model-00013-of-00057.safetensors", + "model.layers.26.self_attn.k_proj.qweight": "model-00013-of-00057.safetensors", + "model.layers.26.self_attn.k_proj.qzeros": "model-00013-of-00057.safetensors", + "model.layers.26.self_attn.k_proj.scales": "model-00013-of-00057.safetensors", + "model.layers.26.self_attn.o_proj.g_idx": "model-00013-of-00057.safetensors", + "model.layers.26.self_attn.o_proj.qweight": "model-00013-of-00057.safetensors", + "model.layers.26.self_attn.o_proj.qzeros": "model-00013-of-00057.safetensors", + "model.layers.26.self_attn.o_proj.scales": "model-00013-of-00057.safetensors", + "model.layers.26.self_attn.q_proj.g_idx": "model-00013-of-00057.safetensors", + "model.layers.26.self_attn.q_proj.qweight": "model-00013-of-00057.safetensors", + "model.layers.26.self_attn.q_proj.qzeros": "model-00013-of-00057.safetensors", + "model.layers.26.self_attn.q_proj.scales": "model-00013-of-00057.safetensors", + "model.layers.26.self_attn.v_proj.g_idx": "model-00013-of-00057.safetensors", + "model.layers.26.self_attn.v_proj.qweight": "model-00013-of-00057.safetensors", + "model.layers.26.self_attn.v_proj.qzeros": "model-00013-of-00057.safetensors", + "model.layers.26.self_attn.v_proj.scales": "model-00013-of-00057.safetensors", + "model.layers.27.input_layernorm.weight": "model-00014-of-00057.safetensors", + "model.layers.27.mlp.down_proj.g_idx": "model-00014-of-00057.safetensors", + "model.layers.27.mlp.down_proj.qweight": "model-00014-of-00057.safetensors", + "model.layers.27.mlp.down_proj.qzeros": "model-00014-of-00057.safetensors", + "model.layers.27.mlp.down_proj.scales": "model-00014-of-00057.safetensors", + "model.layers.27.mlp.gate_proj.g_idx": "model-00013-of-00057.safetensors", + "model.layers.27.mlp.gate_proj.qweight": "model-00013-of-00057.safetensors", + "model.layers.27.mlp.gate_proj.qzeros": "model-00013-of-00057.safetensors", + "model.layers.27.mlp.gate_proj.scales": "model-00013-of-00057.safetensors", + "model.layers.27.mlp.up_proj.g_idx": "model-00013-of-00057.safetensors", + "model.layers.27.mlp.up_proj.qweight": "model-00013-of-00057.safetensors", + "model.layers.27.mlp.up_proj.qzeros": "model-00013-of-00057.safetensors", + "model.layers.27.mlp.up_proj.scales": "model-00013-of-00057.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00014-of-00057.safetensors", + "model.layers.27.self_attn.k_proj.g_idx": "model-00013-of-00057.safetensors", + "model.layers.27.self_attn.k_proj.qweight": "model-00013-of-00057.safetensors", + "model.layers.27.self_attn.k_proj.qzeros": "model-00013-of-00057.safetensors", + "model.layers.27.self_attn.k_proj.scales": "model-00013-of-00057.safetensors", + "model.layers.27.self_attn.o_proj.g_idx": "model-00013-of-00057.safetensors", + "model.layers.27.self_attn.o_proj.qweight": "model-00013-of-00057.safetensors", + "model.layers.27.self_attn.o_proj.qzeros": "model-00013-of-00057.safetensors", + "model.layers.27.self_attn.o_proj.scales": "model-00013-of-00057.safetensors", + "model.layers.27.self_attn.q_proj.g_idx": "model-00013-of-00057.safetensors", + "model.layers.27.self_attn.q_proj.qweight": "model-00013-of-00057.safetensors", + "model.layers.27.self_attn.q_proj.qzeros": "model-00013-of-00057.safetensors", + "model.layers.27.self_attn.q_proj.scales": "model-00013-of-00057.safetensors", + "model.layers.27.self_attn.v_proj.g_idx": "model-00013-of-00057.safetensors", + "model.layers.27.self_attn.v_proj.qweight": "model-00013-of-00057.safetensors", + "model.layers.27.self_attn.v_proj.qzeros": "model-00013-of-00057.safetensors", + "model.layers.27.self_attn.v_proj.scales": "model-00013-of-00057.safetensors", + "model.layers.28.input_layernorm.weight": "model-00014-of-00057.safetensors", + "model.layers.28.mlp.down_proj.g_idx": "model-00014-of-00057.safetensors", + "model.layers.28.mlp.down_proj.qweight": "model-00014-of-00057.safetensors", + "model.layers.28.mlp.down_proj.qzeros": "model-00014-of-00057.safetensors", + "model.layers.28.mlp.down_proj.scales": "model-00014-of-00057.safetensors", + "model.layers.28.mlp.gate_proj.g_idx": "model-00014-of-00057.safetensors", + "model.layers.28.mlp.gate_proj.qweight": "model-00014-of-00057.safetensors", + "model.layers.28.mlp.gate_proj.qzeros": "model-00014-of-00057.safetensors", + "model.layers.28.mlp.gate_proj.scales": "model-00014-of-00057.safetensors", + "model.layers.28.mlp.up_proj.g_idx": "model-00014-of-00057.safetensors", + "model.layers.28.mlp.up_proj.qweight": "model-00014-of-00057.safetensors", + "model.layers.28.mlp.up_proj.qzeros": "model-00014-of-00057.safetensors", + "model.layers.28.mlp.up_proj.scales": "model-00014-of-00057.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00014-of-00057.safetensors", + "model.layers.28.self_attn.k_proj.g_idx": "model-00014-of-00057.safetensors", + "model.layers.28.self_attn.k_proj.qweight": "model-00014-of-00057.safetensors", + "model.layers.28.self_attn.k_proj.qzeros": "model-00014-of-00057.safetensors", + "model.layers.28.self_attn.k_proj.scales": "model-00014-of-00057.safetensors", + "model.layers.28.self_attn.o_proj.g_idx": "model-00014-of-00057.safetensors", + "model.layers.28.self_attn.o_proj.qweight": "model-00014-of-00057.safetensors", + "model.layers.28.self_attn.o_proj.qzeros": "model-00014-of-00057.safetensors", + "model.layers.28.self_attn.o_proj.scales": "model-00014-of-00057.safetensors", + "model.layers.28.self_attn.q_proj.g_idx": "model-00014-of-00057.safetensors", + "model.layers.28.self_attn.q_proj.qweight": "model-00014-of-00057.safetensors", + "model.layers.28.self_attn.q_proj.qzeros": "model-00014-of-00057.safetensors", + "model.layers.28.self_attn.q_proj.scales": "model-00014-of-00057.safetensors", + "model.layers.28.self_attn.v_proj.g_idx": "model-00014-of-00057.safetensors", + "model.layers.28.self_attn.v_proj.qweight": "model-00014-of-00057.safetensors", + "model.layers.28.self_attn.v_proj.qzeros": "model-00014-of-00057.safetensors", + "model.layers.28.self_attn.v_proj.scales": "model-00014-of-00057.safetensors", + "model.layers.29.input_layernorm.weight": "model-00014-of-00057.safetensors", + "model.layers.29.mlp.down_proj.g_idx": "model-00014-of-00057.safetensors", + "model.layers.29.mlp.down_proj.qweight": "model-00014-of-00057.safetensors", + "model.layers.29.mlp.down_proj.qzeros": "model-00014-of-00057.safetensors", + "model.layers.29.mlp.down_proj.scales": "model-00014-of-00057.safetensors", + "model.layers.29.mlp.gate_proj.g_idx": "model-00014-of-00057.safetensors", + "model.layers.29.mlp.gate_proj.qweight": "model-00014-of-00057.safetensors", + "model.layers.29.mlp.gate_proj.qzeros": "model-00014-of-00057.safetensors", + "model.layers.29.mlp.gate_proj.scales": "model-00014-of-00057.safetensors", + "model.layers.29.mlp.up_proj.g_idx": "model-00014-of-00057.safetensors", + "model.layers.29.mlp.up_proj.qweight": "model-00014-of-00057.safetensors", + "model.layers.29.mlp.up_proj.qzeros": "model-00014-of-00057.safetensors", + "model.layers.29.mlp.up_proj.scales": "model-00014-of-00057.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00014-of-00057.safetensors", + "model.layers.29.self_attn.k_proj.g_idx": "model-00014-of-00057.safetensors", + "model.layers.29.self_attn.k_proj.qweight": "model-00014-of-00057.safetensors", + "model.layers.29.self_attn.k_proj.qzeros": "model-00014-of-00057.safetensors", + "model.layers.29.self_attn.k_proj.scales": "model-00014-of-00057.safetensors", + "model.layers.29.self_attn.o_proj.g_idx": "model-00014-of-00057.safetensors", + "model.layers.29.self_attn.o_proj.qweight": "model-00014-of-00057.safetensors", + "model.layers.29.self_attn.o_proj.qzeros": "model-00014-of-00057.safetensors", + "model.layers.29.self_attn.o_proj.scales": "model-00014-of-00057.safetensors", + "model.layers.29.self_attn.q_proj.g_idx": "model-00014-of-00057.safetensors", + "model.layers.29.self_attn.q_proj.qweight": "model-00014-of-00057.safetensors", + "model.layers.29.self_attn.q_proj.qzeros": "model-00014-of-00057.safetensors", + "model.layers.29.self_attn.q_proj.scales": "model-00014-of-00057.safetensors", + "model.layers.29.self_attn.v_proj.g_idx": "model-00014-of-00057.safetensors", + "model.layers.29.self_attn.v_proj.qweight": "model-00014-of-00057.safetensors", + "model.layers.29.self_attn.v_proj.qzeros": "model-00014-of-00057.safetensors", + "model.layers.29.self_attn.v_proj.scales": "model-00014-of-00057.safetensors", + "model.layers.3.input_layernorm.weight": "model-00003-of-00057.safetensors", + "model.layers.3.mlp.down_proj.g_idx": "model-00003-of-00057.safetensors", + "model.layers.3.mlp.down_proj.qweight": "model-00003-of-00057.safetensors", + "model.layers.3.mlp.down_proj.qzeros": "model-00003-of-00057.safetensors", + "model.layers.3.mlp.down_proj.scales": "model-00003-of-00057.safetensors", + "model.layers.3.mlp.gate_proj.g_idx": "model-00003-of-00057.safetensors", + "model.layers.3.mlp.gate_proj.qweight": "model-00003-of-00057.safetensors", + "model.layers.3.mlp.gate_proj.qzeros": "model-00003-of-00057.safetensors", + "model.layers.3.mlp.gate_proj.scales": "model-00003-of-00057.safetensors", + "model.layers.3.mlp.up_proj.g_idx": "model-00003-of-00057.safetensors", + "model.layers.3.mlp.up_proj.qweight": "model-00003-of-00057.safetensors", + "model.layers.3.mlp.up_proj.qzeros": "model-00003-of-00057.safetensors", + "model.layers.3.mlp.up_proj.scales": "model-00003-of-00057.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00003-of-00057.safetensors", + "model.layers.3.self_attn.k_proj.g_idx": "model-00003-of-00057.safetensors", + "model.layers.3.self_attn.k_proj.qweight": "model-00003-of-00057.safetensors", + "model.layers.3.self_attn.k_proj.qzeros": "model-00003-of-00057.safetensors", + "model.layers.3.self_attn.k_proj.scales": "model-00003-of-00057.safetensors", + "model.layers.3.self_attn.o_proj.g_idx": "model-00003-of-00057.safetensors", + "model.layers.3.self_attn.o_proj.qweight": "model-00003-of-00057.safetensors", + "model.layers.3.self_attn.o_proj.qzeros": "model-00003-of-00057.safetensors", + "model.layers.3.self_attn.o_proj.scales": "model-00003-of-00057.safetensors", + "model.layers.3.self_attn.q_proj.g_idx": "model-00003-of-00057.safetensors", + "model.layers.3.self_attn.q_proj.qweight": "model-00003-of-00057.safetensors", + "model.layers.3.self_attn.q_proj.qzeros": "model-00003-of-00057.safetensors", + "model.layers.3.self_attn.q_proj.scales": "model-00003-of-00057.safetensors", + "model.layers.3.self_attn.v_proj.g_idx": "model-00003-of-00057.safetensors", + "model.layers.3.self_attn.v_proj.qweight": "model-00003-of-00057.safetensors", + "model.layers.3.self_attn.v_proj.qzeros": "model-00003-of-00057.safetensors", + "model.layers.3.self_attn.v_proj.scales": "model-00003-of-00057.safetensors", + "model.layers.30.input_layernorm.weight": "model-00015-of-00057.safetensors", + "model.layers.30.mlp.down_proj.g_idx": "model-00015-of-00057.safetensors", + "model.layers.30.mlp.down_proj.qweight": "model-00015-of-00057.safetensors", + "model.layers.30.mlp.down_proj.qzeros": "model-00015-of-00057.safetensors", + "model.layers.30.mlp.down_proj.scales": "model-00015-of-00057.safetensors", + "model.layers.30.mlp.gate_proj.g_idx": "model-00015-of-00057.safetensors", + "model.layers.30.mlp.gate_proj.qweight": "model-00015-of-00057.safetensors", + "model.layers.30.mlp.gate_proj.qzeros": "model-00015-of-00057.safetensors", + "model.layers.30.mlp.gate_proj.scales": "model-00015-of-00057.safetensors", + "model.layers.30.mlp.up_proj.g_idx": "model-00015-of-00057.safetensors", + "model.layers.30.mlp.up_proj.qweight": "model-00015-of-00057.safetensors", + "model.layers.30.mlp.up_proj.qzeros": "model-00015-of-00057.safetensors", + "model.layers.30.mlp.up_proj.scales": "model-00015-of-00057.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00015-of-00057.safetensors", + "model.layers.30.self_attn.k_proj.g_idx": "model-00014-of-00057.safetensors", + "model.layers.30.self_attn.k_proj.qweight": "model-00014-of-00057.safetensors", + "model.layers.30.self_attn.k_proj.qzeros": "model-00014-of-00057.safetensors", + "model.layers.30.self_attn.k_proj.scales": "model-00014-of-00057.safetensors", + "model.layers.30.self_attn.o_proj.g_idx": "model-00015-of-00057.safetensors", + "model.layers.30.self_attn.o_proj.qweight": "model-00015-of-00057.safetensors", + "model.layers.30.self_attn.o_proj.qzeros": "model-00015-of-00057.safetensors", + "model.layers.30.self_attn.o_proj.scales": "model-00015-of-00057.safetensors", + "model.layers.30.self_attn.q_proj.g_idx": "model-00014-of-00057.safetensors", + "model.layers.30.self_attn.q_proj.qweight": "model-00014-of-00057.safetensors", + "model.layers.30.self_attn.q_proj.qzeros": "model-00014-of-00057.safetensors", + "model.layers.30.self_attn.q_proj.scales": "model-00014-of-00057.safetensors", + "model.layers.30.self_attn.v_proj.g_idx": "model-00014-of-00057.safetensors", + "model.layers.30.self_attn.v_proj.qweight": "model-00014-of-00057.safetensors", + "model.layers.30.self_attn.v_proj.qzeros": "model-00014-of-00057.safetensors", + "model.layers.30.self_attn.v_proj.scales": "model-00014-of-00057.safetensors", + "model.layers.31.input_layernorm.weight": "model-00015-of-00057.safetensors", + "model.layers.31.mlp.down_proj.g_idx": "model-00015-of-00057.safetensors", + "model.layers.31.mlp.down_proj.qweight": "model-00015-of-00057.safetensors", + "model.layers.31.mlp.down_proj.qzeros": "model-00015-of-00057.safetensors", + "model.layers.31.mlp.down_proj.scales": "model-00015-of-00057.safetensors", + "model.layers.31.mlp.gate_proj.g_idx": "model-00015-of-00057.safetensors", + "model.layers.31.mlp.gate_proj.qweight": "model-00015-of-00057.safetensors", + "model.layers.31.mlp.gate_proj.qzeros": "model-00015-of-00057.safetensors", + "model.layers.31.mlp.gate_proj.scales": "model-00015-of-00057.safetensors", + "model.layers.31.mlp.up_proj.g_idx": "model-00015-of-00057.safetensors", + "model.layers.31.mlp.up_proj.qweight": "model-00015-of-00057.safetensors", + "model.layers.31.mlp.up_proj.qzeros": "model-00015-of-00057.safetensors", + "model.layers.31.mlp.up_proj.scales": "model-00015-of-00057.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00015-of-00057.safetensors", + "model.layers.31.self_attn.k_proj.g_idx": "model-00015-of-00057.safetensors", + "model.layers.31.self_attn.k_proj.qweight": "model-00015-of-00057.safetensors", + "model.layers.31.self_attn.k_proj.qzeros": "model-00015-of-00057.safetensors", + "model.layers.31.self_attn.k_proj.scales": "model-00015-of-00057.safetensors", + "model.layers.31.self_attn.o_proj.g_idx": "model-00015-of-00057.safetensors", + "model.layers.31.self_attn.o_proj.qweight": "model-00015-of-00057.safetensors", + "model.layers.31.self_attn.o_proj.qzeros": "model-00015-of-00057.safetensors", + "model.layers.31.self_attn.o_proj.scales": "model-00015-of-00057.safetensors", + "model.layers.31.self_attn.q_proj.g_idx": "model-00015-of-00057.safetensors", + "model.layers.31.self_attn.q_proj.qweight": "model-00015-of-00057.safetensors", + "model.layers.31.self_attn.q_proj.qzeros": "model-00015-of-00057.safetensors", + "model.layers.31.self_attn.q_proj.scales": "model-00015-of-00057.safetensors", + "model.layers.31.self_attn.v_proj.g_idx": "model-00015-of-00057.safetensors", + "model.layers.31.self_attn.v_proj.qweight": "model-00015-of-00057.safetensors", + "model.layers.31.self_attn.v_proj.qzeros": "model-00015-of-00057.safetensors", + "model.layers.31.self_attn.v_proj.scales": "model-00015-of-00057.safetensors", + "model.layers.32.input_layernorm.weight": "model-00016-of-00057.safetensors", + "model.layers.32.mlp.down_proj.g_idx": "model-00016-of-00057.safetensors", + "model.layers.32.mlp.down_proj.qweight": "model-00016-of-00057.safetensors", + "model.layers.32.mlp.down_proj.qzeros": "model-00016-of-00057.safetensors", + "model.layers.32.mlp.down_proj.scales": "model-00016-of-00057.safetensors", + "model.layers.32.mlp.gate_proj.g_idx": "model-00015-of-00057.safetensors", + "model.layers.32.mlp.gate_proj.qweight": "model-00015-of-00057.safetensors", + "model.layers.32.mlp.gate_proj.qzeros": "model-00015-of-00057.safetensors", + "model.layers.32.mlp.gate_proj.scales": "model-00015-of-00057.safetensors", + "model.layers.32.mlp.up_proj.g_idx": "model-00016-of-00057.safetensors", + "model.layers.32.mlp.up_proj.qweight": "model-00016-of-00057.safetensors", + "model.layers.32.mlp.up_proj.qzeros": "model-00016-of-00057.safetensors", + "model.layers.32.mlp.up_proj.scales": "model-00016-of-00057.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00016-of-00057.safetensors", + "model.layers.32.self_attn.k_proj.g_idx": "model-00015-of-00057.safetensors", + "model.layers.32.self_attn.k_proj.qweight": "model-00015-of-00057.safetensors", + "model.layers.32.self_attn.k_proj.qzeros": "model-00015-of-00057.safetensors", + "model.layers.32.self_attn.k_proj.scales": "model-00015-of-00057.safetensors", + "model.layers.32.self_attn.o_proj.g_idx": "model-00015-of-00057.safetensors", + "model.layers.32.self_attn.o_proj.qweight": "model-00015-of-00057.safetensors", + "model.layers.32.self_attn.o_proj.qzeros": "model-00015-of-00057.safetensors", + "model.layers.32.self_attn.o_proj.scales": "model-00015-of-00057.safetensors", + "model.layers.32.self_attn.q_proj.g_idx": "model-00015-of-00057.safetensors", + "model.layers.32.self_attn.q_proj.qweight": "model-00015-of-00057.safetensors", + "model.layers.32.self_attn.q_proj.qzeros": "model-00015-of-00057.safetensors", + "model.layers.32.self_attn.q_proj.scales": "model-00015-of-00057.safetensors", + "model.layers.32.self_attn.v_proj.g_idx": "model-00015-of-00057.safetensors", + "model.layers.32.self_attn.v_proj.qweight": "model-00015-of-00057.safetensors", + "model.layers.32.self_attn.v_proj.qzeros": "model-00015-of-00057.safetensors", + "model.layers.32.self_attn.v_proj.scales": "model-00015-of-00057.safetensors", + "model.layers.33.input_layernorm.weight": "model-00016-of-00057.safetensors", + "model.layers.33.mlp.down_proj.g_idx": "model-00016-of-00057.safetensors", + "model.layers.33.mlp.down_proj.qweight": "model-00016-of-00057.safetensors", + "model.layers.33.mlp.down_proj.qzeros": "model-00016-of-00057.safetensors", + "model.layers.33.mlp.down_proj.scales": "model-00016-of-00057.safetensors", + "model.layers.33.mlp.gate_proj.g_idx": "model-00016-of-00057.safetensors", + "model.layers.33.mlp.gate_proj.qweight": "model-00016-of-00057.safetensors", + "model.layers.33.mlp.gate_proj.qzeros": "model-00016-of-00057.safetensors", + "model.layers.33.mlp.gate_proj.scales": "model-00016-of-00057.safetensors", + "model.layers.33.mlp.up_proj.g_idx": "model-00016-of-00057.safetensors", + "model.layers.33.mlp.up_proj.qweight": "model-00016-of-00057.safetensors", + "model.layers.33.mlp.up_proj.qzeros": "model-00016-of-00057.safetensors", + "model.layers.33.mlp.up_proj.scales": "model-00016-of-00057.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00016-of-00057.safetensors", + "model.layers.33.self_attn.k_proj.g_idx": "model-00016-of-00057.safetensors", + "model.layers.33.self_attn.k_proj.qweight": "model-00016-of-00057.safetensors", + "model.layers.33.self_attn.k_proj.qzeros": "model-00016-of-00057.safetensors", + "model.layers.33.self_attn.k_proj.scales": "model-00016-of-00057.safetensors", + "model.layers.33.self_attn.o_proj.g_idx": "model-00016-of-00057.safetensors", + "model.layers.33.self_attn.o_proj.qweight": "model-00016-of-00057.safetensors", + "model.layers.33.self_attn.o_proj.qzeros": "model-00016-of-00057.safetensors", + "model.layers.33.self_attn.o_proj.scales": "model-00016-of-00057.safetensors", + "model.layers.33.self_attn.q_proj.g_idx": "model-00016-of-00057.safetensors", + "model.layers.33.self_attn.q_proj.qweight": "model-00016-of-00057.safetensors", + "model.layers.33.self_attn.q_proj.qzeros": "model-00016-of-00057.safetensors", + "model.layers.33.self_attn.q_proj.scales": "model-00016-of-00057.safetensors", + "model.layers.33.self_attn.v_proj.g_idx": "model-00016-of-00057.safetensors", + "model.layers.33.self_attn.v_proj.qweight": "model-00016-of-00057.safetensors", + "model.layers.33.self_attn.v_proj.qzeros": "model-00016-of-00057.safetensors", + "model.layers.33.self_attn.v_proj.scales": "model-00016-of-00057.safetensors", + "model.layers.34.input_layernorm.weight": "model-00017-of-00057.safetensors", + "model.layers.34.mlp.down_proj.g_idx": "model-00017-of-00057.safetensors", + "model.layers.34.mlp.down_proj.qweight": "model-00017-of-00057.safetensors", + "model.layers.34.mlp.down_proj.qzeros": "model-00017-of-00057.safetensors", + "model.layers.34.mlp.down_proj.scales": "model-00017-of-00057.safetensors", + "model.layers.34.mlp.gate_proj.g_idx": "model-00016-of-00057.safetensors", + "model.layers.34.mlp.gate_proj.qweight": "model-00016-of-00057.safetensors", + "model.layers.34.mlp.gate_proj.qzeros": "model-00016-of-00057.safetensors", + "model.layers.34.mlp.gate_proj.scales": "model-00016-of-00057.safetensors", + "model.layers.34.mlp.up_proj.g_idx": "model-00016-of-00057.safetensors", + "model.layers.34.mlp.up_proj.qweight": "model-00016-of-00057.safetensors", + "model.layers.34.mlp.up_proj.qzeros": "model-00016-of-00057.safetensors", + "model.layers.34.mlp.up_proj.scales": "model-00016-of-00057.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00017-of-00057.safetensors", + "model.layers.34.self_attn.k_proj.g_idx": "model-00016-of-00057.safetensors", + "model.layers.34.self_attn.k_proj.qweight": "model-00016-of-00057.safetensors", + "model.layers.34.self_attn.k_proj.qzeros": "model-00016-of-00057.safetensors", + "model.layers.34.self_attn.k_proj.scales": "model-00016-of-00057.safetensors", + "model.layers.34.self_attn.o_proj.g_idx": "model-00016-of-00057.safetensors", + "model.layers.34.self_attn.o_proj.qweight": "model-00016-of-00057.safetensors", + "model.layers.34.self_attn.o_proj.qzeros": "model-00016-of-00057.safetensors", + "model.layers.34.self_attn.o_proj.scales": "model-00016-of-00057.safetensors", + "model.layers.34.self_attn.q_proj.g_idx": "model-00016-of-00057.safetensors", + "model.layers.34.self_attn.q_proj.qweight": "model-00016-of-00057.safetensors", + "model.layers.34.self_attn.q_proj.qzeros": "model-00016-of-00057.safetensors", + "model.layers.34.self_attn.q_proj.scales": "model-00016-of-00057.safetensors", + "model.layers.34.self_attn.v_proj.g_idx": "model-00016-of-00057.safetensors", + "model.layers.34.self_attn.v_proj.qweight": "model-00016-of-00057.safetensors", + "model.layers.34.self_attn.v_proj.qzeros": "model-00016-of-00057.safetensors", + "model.layers.34.self_attn.v_proj.scales": "model-00016-of-00057.safetensors", + "model.layers.35.input_layernorm.weight": "model-00017-of-00057.safetensors", + "model.layers.35.mlp.down_proj.g_idx": "model-00017-of-00057.safetensors", + "model.layers.35.mlp.down_proj.qweight": "model-00017-of-00057.safetensors", + "model.layers.35.mlp.down_proj.qzeros": "model-00017-of-00057.safetensors", + "model.layers.35.mlp.down_proj.scales": "model-00017-of-00057.safetensors", + "model.layers.35.mlp.gate_proj.g_idx": "model-00017-of-00057.safetensors", + "model.layers.35.mlp.gate_proj.qweight": "model-00017-of-00057.safetensors", + "model.layers.35.mlp.gate_proj.qzeros": "model-00017-of-00057.safetensors", + "model.layers.35.mlp.gate_proj.scales": "model-00017-of-00057.safetensors", + "model.layers.35.mlp.up_proj.g_idx": "model-00017-of-00057.safetensors", + "model.layers.35.mlp.up_proj.qweight": "model-00017-of-00057.safetensors", + "model.layers.35.mlp.up_proj.qzeros": "model-00017-of-00057.safetensors", + "model.layers.35.mlp.up_proj.scales": "model-00017-of-00057.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00017-of-00057.safetensors", + "model.layers.35.self_attn.k_proj.g_idx": "model-00017-of-00057.safetensors", + "model.layers.35.self_attn.k_proj.qweight": "model-00017-of-00057.safetensors", + "model.layers.35.self_attn.k_proj.qzeros": "model-00017-of-00057.safetensors", + "model.layers.35.self_attn.k_proj.scales": "model-00017-of-00057.safetensors", + "model.layers.35.self_attn.o_proj.g_idx": "model-00017-of-00057.safetensors", + "model.layers.35.self_attn.o_proj.qweight": "model-00017-of-00057.safetensors", + "model.layers.35.self_attn.o_proj.qzeros": "model-00017-of-00057.safetensors", + "model.layers.35.self_attn.o_proj.scales": "model-00017-of-00057.safetensors", + "model.layers.35.self_attn.q_proj.g_idx": "model-00017-of-00057.safetensors", + "model.layers.35.self_attn.q_proj.qweight": "model-00017-of-00057.safetensors", + "model.layers.35.self_attn.q_proj.qzeros": "model-00017-of-00057.safetensors", + "model.layers.35.self_attn.q_proj.scales": "model-00017-of-00057.safetensors", + "model.layers.35.self_attn.v_proj.g_idx": "model-00017-of-00057.safetensors", + "model.layers.35.self_attn.v_proj.qweight": "model-00017-of-00057.safetensors", + "model.layers.35.self_attn.v_proj.qzeros": "model-00017-of-00057.safetensors", + "model.layers.35.self_attn.v_proj.scales": "model-00017-of-00057.safetensors", + "model.layers.36.input_layernorm.weight": "model-00017-of-00057.safetensors", + "model.layers.36.mlp.down_proj.g_idx": "model-00017-of-00057.safetensors", + "model.layers.36.mlp.down_proj.qweight": "model-00017-of-00057.safetensors", + "model.layers.36.mlp.down_proj.qzeros": "model-00017-of-00057.safetensors", + "model.layers.36.mlp.down_proj.scales": "model-00017-of-00057.safetensors", + "model.layers.36.mlp.gate_proj.g_idx": "model-00017-of-00057.safetensors", + "model.layers.36.mlp.gate_proj.qweight": "model-00017-of-00057.safetensors", + "model.layers.36.mlp.gate_proj.qzeros": "model-00017-of-00057.safetensors", + "model.layers.36.mlp.gate_proj.scales": "model-00017-of-00057.safetensors", + "model.layers.36.mlp.up_proj.g_idx": "model-00017-of-00057.safetensors", + "model.layers.36.mlp.up_proj.qweight": "model-00017-of-00057.safetensors", + "model.layers.36.mlp.up_proj.qzeros": "model-00017-of-00057.safetensors", + "model.layers.36.mlp.up_proj.scales": "model-00017-of-00057.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00017-of-00057.safetensors", + "model.layers.36.self_attn.k_proj.g_idx": "model-00017-of-00057.safetensors", + "model.layers.36.self_attn.k_proj.qweight": "model-00017-of-00057.safetensors", + "model.layers.36.self_attn.k_proj.qzeros": "model-00017-of-00057.safetensors", + "model.layers.36.self_attn.k_proj.scales": "model-00017-of-00057.safetensors", + "model.layers.36.self_attn.o_proj.g_idx": "model-00017-of-00057.safetensors", + "model.layers.36.self_attn.o_proj.qweight": "model-00017-of-00057.safetensors", + "model.layers.36.self_attn.o_proj.qzeros": "model-00017-of-00057.safetensors", + "model.layers.36.self_attn.o_proj.scales": "model-00017-of-00057.safetensors", + "model.layers.36.self_attn.q_proj.g_idx": "model-00017-of-00057.safetensors", + "model.layers.36.self_attn.q_proj.qweight": "model-00017-of-00057.safetensors", + "model.layers.36.self_attn.q_proj.qzeros": "model-00017-of-00057.safetensors", + "model.layers.36.self_attn.q_proj.scales": "model-00017-of-00057.safetensors", + "model.layers.36.self_attn.v_proj.g_idx": "model-00017-of-00057.safetensors", + "model.layers.36.self_attn.v_proj.qweight": "model-00017-of-00057.safetensors", + "model.layers.36.self_attn.v_proj.qzeros": "model-00017-of-00057.safetensors", + "model.layers.36.self_attn.v_proj.scales": "model-00017-of-00057.safetensors", + "model.layers.37.input_layernorm.weight": "model-00018-of-00057.safetensors", + "model.layers.37.mlp.down_proj.g_idx": "model-00018-of-00057.safetensors", + "model.layers.37.mlp.down_proj.qweight": "model-00018-of-00057.safetensors", + "model.layers.37.mlp.down_proj.qzeros": "model-00018-of-00057.safetensors", + "model.layers.37.mlp.down_proj.scales": "model-00018-of-00057.safetensors", + "model.layers.37.mlp.gate_proj.g_idx": "model-00018-of-00057.safetensors", + "model.layers.37.mlp.gate_proj.qweight": "model-00018-of-00057.safetensors", + "model.layers.37.mlp.gate_proj.qzeros": "model-00018-of-00057.safetensors", + "model.layers.37.mlp.gate_proj.scales": "model-00018-of-00057.safetensors", + "model.layers.37.mlp.up_proj.g_idx": "model-00018-of-00057.safetensors", + "model.layers.37.mlp.up_proj.qweight": "model-00018-of-00057.safetensors", + "model.layers.37.mlp.up_proj.qzeros": "model-00018-of-00057.safetensors", + "model.layers.37.mlp.up_proj.scales": "model-00018-of-00057.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00018-of-00057.safetensors", + "model.layers.37.self_attn.k_proj.g_idx": "model-00017-of-00057.safetensors", + "model.layers.37.self_attn.k_proj.qweight": "model-00017-of-00057.safetensors", + "model.layers.37.self_attn.k_proj.qzeros": "model-00017-of-00057.safetensors", + "model.layers.37.self_attn.k_proj.scales": "model-00017-of-00057.safetensors", + "model.layers.37.self_attn.o_proj.g_idx": "model-00018-of-00057.safetensors", + "model.layers.37.self_attn.o_proj.qweight": "model-00018-of-00057.safetensors", + "model.layers.37.self_attn.o_proj.qzeros": "model-00018-of-00057.safetensors", + "model.layers.37.self_attn.o_proj.scales": "model-00018-of-00057.safetensors", + "model.layers.37.self_attn.q_proj.g_idx": "model-00017-of-00057.safetensors", + "model.layers.37.self_attn.q_proj.qweight": "model-00017-of-00057.safetensors", + "model.layers.37.self_attn.q_proj.qzeros": "model-00017-of-00057.safetensors", + "model.layers.37.self_attn.q_proj.scales": "model-00017-of-00057.safetensors", + "model.layers.37.self_attn.v_proj.g_idx": "model-00017-of-00057.safetensors", + "model.layers.37.self_attn.v_proj.qweight": "model-00017-of-00057.safetensors", + "model.layers.37.self_attn.v_proj.qzeros": "model-00017-of-00057.safetensors", + "model.layers.37.self_attn.v_proj.scales": "model-00017-of-00057.safetensors", + "model.layers.38.input_layernorm.weight": "model-00018-of-00057.safetensors", + "model.layers.38.mlp.down_proj.g_idx": "model-00018-of-00057.safetensors", + "model.layers.38.mlp.down_proj.qweight": "model-00018-of-00057.safetensors", + "model.layers.38.mlp.down_proj.qzeros": "model-00018-of-00057.safetensors", + "model.layers.38.mlp.down_proj.scales": "model-00018-of-00057.safetensors", + "model.layers.38.mlp.gate_proj.g_idx": "model-00018-of-00057.safetensors", + "model.layers.38.mlp.gate_proj.qweight": "model-00018-of-00057.safetensors", + "model.layers.38.mlp.gate_proj.qzeros": "model-00018-of-00057.safetensors", + "model.layers.38.mlp.gate_proj.scales": "model-00018-of-00057.safetensors", + "model.layers.38.mlp.up_proj.g_idx": "model-00018-of-00057.safetensors", + "model.layers.38.mlp.up_proj.qweight": "model-00018-of-00057.safetensors", + "model.layers.38.mlp.up_proj.qzeros": "model-00018-of-00057.safetensors", + "model.layers.38.mlp.up_proj.scales": "model-00018-of-00057.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00018-of-00057.safetensors", + "model.layers.38.self_attn.k_proj.g_idx": "model-00018-of-00057.safetensors", + "model.layers.38.self_attn.k_proj.qweight": "model-00018-of-00057.safetensors", + "model.layers.38.self_attn.k_proj.qzeros": "model-00018-of-00057.safetensors", + "model.layers.38.self_attn.k_proj.scales": "model-00018-of-00057.safetensors", + "model.layers.38.self_attn.o_proj.g_idx": "model-00018-of-00057.safetensors", + "model.layers.38.self_attn.o_proj.qweight": "model-00018-of-00057.safetensors", + "model.layers.38.self_attn.o_proj.qzeros": "model-00018-of-00057.safetensors", + "model.layers.38.self_attn.o_proj.scales": "model-00018-of-00057.safetensors", + "model.layers.38.self_attn.q_proj.g_idx": "model-00018-of-00057.safetensors", + "model.layers.38.self_attn.q_proj.qweight": "model-00018-of-00057.safetensors", + "model.layers.38.self_attn.q_proj.qzeros": "model-00018-of-00057.safetensors", + "model.layers.38.self_attn.q_proj.scales": "model-00018-of-00057.safetensors", + "model.layers.38.self_attn.v_proj.g_idx": "model-00018-of-00057.safetensors", + "model.layers.38.self_attn.v_proj.qweight": "model-00018-of-00057.safetensors", + "model.layers.38.self_attn.v_proj.qzeros": "model-00018-of-00057.safetensors", + "model.layers.38.self_attn.v_proj.scales": "model-00018-of-00057.safetensors", + "model.layers.39.input_layernorm.weight": "model-00019-of-00057.safetensors", + "model.layers.39.mlp.down_proj.g_idx": "model-00019-of-00057.safetensors", + "model.layers.39.mlp.down_proj.qweight": "model-00019-of-00057.safetensors", + "model.layers.39.mlp.down_proj.qzeros": "model-00019-of-00057.safetensors", + "model.layers.39.mlp.down_proj.scales": "model-00019-of-00057.safetensors", + "model.layers.39.mlp.gate_proj.g_idx": "model-00018-of-00057.safetensors", + "model.layers.39.mlp.gate_proj.qweight": "model-00018-of-00057.safetensors", + "model.layers.39.mlp.gate_proj.qzeros": "model-00018-of-00057.safetensors", + "model.layers.39.mlp.gate_proj.scales": "model-00018-of-00057.safetensors", + "model.layers.39.mlp.up_proj.g_idx": "model-00019-of-00057.safetensors", + "model.layers.39.mlp.up_proj.qweight": "model-00019-of-00057.safetensors", + "model.layers.39.mlp.up_proj.qzeros": "model-00019-of-00057.safetensors", + "model.layers.39.mlp.up_proj.scales": "model-00019-of-00057.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00019-of-00057.safetensors", + "model.layers.39.self_attn.k_proj.g_idx": "model-00018-of-00057.safetensors", + "model.layers.39.self_attn.k_proj.qweight": "model-00018-of-00057.safetensors", + "model.layers.39.self_attn.k_proj.qzeros": "model-00018-of-00057.safetensors", + "model.layers.39.self_attn.k_proj.scales": "model-00018-of-00057.safetensors", + "model.layers.39.self_attn.o_proj.g_idx": "model-00018-of-00057.safetensors", + "model.layers.39.self_attn.o_proj.qweight": "model-00018-of-00057.safetensors", + "model.layers.39.self_attn.o_proj.qzeros": "model-00018-of-00057.safetensors", + "model.layers.39.self_attn.o_proj.scales": "model-00018-of-00057.safetensors", + "model.layers.39.self_attn.q_proj.g_idx": "model-00018-of-00057.safetensors", + "model.layers.39.self_attn.q_proj.qweight": "model-00018-of-00057.safetensors", + "model.layers.39.self_attn.q_proj.qzeros": "model-00018-of-00057.safetensors", + "model.layers.39.self_attn.q_proj.scales": "model-00018-of-00057.safetensors", + "model.layers.39.self_attn.v_proj.g_idx": "model-00018-of-00057.safetensors", + "model.layers.39.self_attn.v_proj.qweight": "model-00018-of-00057.safetensors", + "model.layers.39.self_attn.v_proj.qzeros": "model-00018-of-00057.safetensors", + "model.layers.39.self_attn.v_proj.scales": "model-00018-of-00057.safetensors", + "model.layers.4.input_layernorm.weight": "model-00004-of-00057.safetensors", + "model.layers.4.mlp.down_proj.g_idx": "model-00004-of-00057.safetensors", + "model.layers.4.mlp.down_proj.qweight": "model-00004-of-00057.safetensors", + "model.layers.4.mlp.down_proj.qzeros": "model-00004-of-00057.safetensors", + "model.layers.4.mlp.down_proj.scales": "model-00004-of-00057.safetensors", + "model.layers.4.mlp.gate_proj.g_idx": "model-00003-of-00057.safetensors", + "model.layers.4.mlp.gate_proj.qweight": "model-00003-of-00057.safetensors", + "model.layers.4.mlp.gate_proj.qzeros": "model-00003-of-00057.safetensors", + "model.layers.4.mlp.gate_proj.scales": "model-00003-of-00057.safetensors", + "model.layers.4.mlp.up_proj.g_idx": "model-00004-of-00057.safetensors", + "model.layers.4.mlp.up_proj.qweight": "model-00004-of-00057.safetensors", + "model.layers.4.mlp.up_proj.qzeros": "model-00004-of-00057.safetensors", + "model.layers.4.mlp.up_proj.scales": "model-00004-of-00057.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00004-of-00057.safetensors", + "model.layers.4.self_attn.k_proj.g_idx": "model-00003-of-00057.safetensors", + "model.layers.4.self_attn.k_proj.qweight": "model-00003-of-00057.safetensors", + "model.layers.4.self_attn.k_proj.qzeros": "model-00003-of-00057.safetensors", + "model.layers.4.self_attn.k_proj.scales": "model-00003-of-00057.safetensors", + "model.layers.4.self_attn.o_proj.g_idx": "model-00003-of-00057.safetensors", + "model.layers.4.self_attn.o_proj.qweight": "model-00003-of-00057.safetensors", + "model.layers.4.self_attn.o_proj.qzeros": "model-00003-of-00057.safetensors", + "model.layers.4.self_attn.o_proj.scales": "model-00003-of-00057.safetensors", + "model.layers.4.self_attn.q_proj.g_idx": "model-00003-of-00057.safetensors", + "model.layers.4.self_attn.q_proj.qweight": "model-00003-of-00057.safetensors", + "model.layers.4.self_attn.q_proj.qzeros": "model-00003-of-00057.safetensors", + "model.layers.4.self_attn.q_proj.scales": "model-00003-of-00057.safetensors", + "model.layers.4.self_attn.v_proj.g_idx": "model-00003-of-00057.safetensors", + "model.layers.4.self_attn.v_proj.qweight": "model-00003-of-00057.safetensors", + "model.layers.4.self_attn.v_proj.qzeros": "model-00003-of-00057.safetensors", + "model.layers.4.self_attn.v_proj.scales": "model-00003-of-00057.safetensors", + "model.layers.40.input_layernorm.weight": "model-00019-of-00057.safetensors", + "model.layers.40.mlp.down_proj.g_idx": "model-00019-of-00057.safetensors", + "model.layers.40.mlp.down_proj.qweight": "model-00019-of-00057.safetensors", + "model.layers.40.mlp.down_proj.qzeros": "model-00019-of-00057.safetensors", + "model.layers.40.mlp.down_proj.scales": "model-00019-of-00057.safetensors", + "model.layers.40.mlp.gate_proj.g_idx": "model-00019-of-00057.safetensors", + "model.layers.40.mlp.gate_proj.qweight": "model-00019-of-00057.safetensors", + "model.layers.40.mlp.gate_proj.qzeros": "model-00019-of-00057.safetensors", + "model.layers.40.mlp.gate_proj.scales": "model-00019-of-00057.safetensors", + "model.layers.40.mlp.up_proj.g_idx": "model-00019-of-00057.safetensors", + "model.layers.40.mlp.up_proj.qweight": "model-00019-of-00057.safetensors", + "model.layers.40.mlp.up_proj.qzeros": "model-00019-of-00057.safetensors", + "model.layers.40.mlp.up_proj.scales": "model-00019-of-00057.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00019-of-00057.safetensors", + "model.layers.40.self_attn.k_proj.g_idx": "model-00019-of-00057.safetensors", + "model.layers.40.self_attn.k_proj.qweight": "model-00019-of-00057.safetensors", + "model.layers.40.self_attn.k_proj.qzeros": "model-00019-of-00057.safetensors", + "model.layers.40.self_attn.k_proj.scales": "model-00019-of-00057.safetensors", + "model.layers.40.self_attn.o_proj.g_idx": "model-00019-of-00057.safetensors", + "model.layers.40.self_attn.o_proj.qweight": "model-00019-of-00057.safetensors", + "model.layers.40.self_attn.o_proj.qzeros": "model-00019-of-00057.safetensors", + "model.layers.40.self_attn.o_proj.scales": "model-00019-of-00057.safetensors", + "model.layers.40.self_attn.q_proj.g_idx": "model-00019-of-00057.safetensors", + "model.layers.40.self_attn.q_proj.qweight": "model-00019-of-00057.safetensors", + "model.layers.40.self_attn.q_proj.qzeros": "model-00019-of-00057.safetensors", + "model.layers.40.self_attn.q_proj.scales": "model-00019-of-00057.safetensors", + "model.layers.40.self_attn.v_proj.g_idx": "model-00019-of-00057.safetensors", + "model.layers.40.self_attn.v_proj.qweight": "model-00019-of-00057.safetensors", + "model.layers.40.self_attn.v_proj.qzeros": "model-00019-of-00057.safetensors", + "model.layers.40.self_attn.v_proj.scales": "model-00019-of-00057.safetensors", + "model.layers.41.input_layernorm.weight": "model-00020-of-00057.safetensors", + "model.layers.41.mlp.down_proj.g_idx": "model-00020-of-00057.safetensors", + "model.layers.41.mlp.down_proj.qweight": "model-00020-of-00057.safetensors", + "model.layers.41.mlp.down_proj.qzeros": "model-00020-of-00057.safetensors", + "model.layers.41.mlp.down_proj.scales": "model-00020-of-00057.safetensors", + "model.layers.41.mlp.gate_proj.g_idx": "model-00019-of-00057.safetensors", + "model.layers.41.mlp.gate_proj.qweight": "model-00019-of-00057.safetensors", + "model.layers.41.mlp.gate_proj.qzeros": "model-00019-of-00057.safetensors", + "model.layers.41.mlp.gate_proj.scales": "model-00019-of-00057.safetensors", + "model.layers.41.mlp.up_proj.g_idx": "model-00019-of-00057.safetensors", + "model.layers.41.mlp.up_proj.qweight": "model-00019-of-00057.safetensors", + "model.layers.41.mlp.up_proj.qzeros": "model-00019-of-00057.safetensors", + "model.layers.41.mlp.up_proj.scales": "model-00019-of-00057.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00020-of-00057.safetensors", + "model.layers.41.self_attn.k_proj.g_idx": "model-00019-of-00057.safetensors", + "model.layers.41.self_attn.k_proj.qweight": "model-00019-of-00057.safetensors", + "model.layers.41.self_attn.k_proj.qzeros": "model-00019-of-00057.safetensors", + "model.layers.41.self_attn.k_proj.scales": "model-00019-of-00057.safetensors", + "model.layers.41.self_attn.o_proj.g_idx": "model-00019-of-00057.safetensors", + "model.layers.41.self_attn.o_proj.qweight": "model-00019-of-00057.safetensors", + "model.layers.41.self_attn.o_proj.qzeros": "model-00019-of-00057.safetensors", + "model.layers.41.self_attn.o_proj.scales": "model-00019-of-00057.safetensors", + "model.layers.41.self_attn.q_proj.g_idx": "model-00019-of-00057.safetensors", + "model.layers.41.self_attn.q_proj.qweight": "model-00019-of-00057.safetensors", + "model.layers.41.self_attn.q_proj.qzeros": "model-00019-of-00057.safetensors", + "model.layers.41.self_attn.q_proj.scales": "model-00019-of-00057.safetensors", + "model.layers.41.self_attn.v_proj.g_idx": "model-00019-of-00057.safetensors", + "model.layers.41.self_attn.v_proj.qweight": "model-00019-of-00057.safetensors", + "model.layers.41.self_attn.v_proj.qzeros": "model-00019-of-00057.safetensors", + "model.layers.41.self_attn.v_proj.scales": "model-00019-of-00057.safetensors", + "model.layers.42.input_layernorm.weight": "model-00020-of-00057.safetensors", + "model.layers.42.mlp.down_proj.g_idx": "model-00020-of-00057.safetensors", + "model.layers.42.mlp.down_proj.qweight": "model-00020-of-00057.safetensors", + "model.layers.42.mlp.down_proj.qzeros": "model-00020-of-00057.safetensors", + "model.layers.42.mlp.down_proj.scales": "model-00020-of-00057.safetensors", + "model.layers.42.mlp.gate_proj.g_idx": "model-00020-of-00057.safetensors", + "model.layers.42.mlp.gate_proj.qweight": "model-00020-of-00057.safetensors", + "model.layers.42.mlp.gate_proj.qzeros": "model-00020-of-00057.safetensors", + "model.layers.42.mlp.gate_proj.scales": "model-00020-of-00057.safetensors", + "model.layers.42.mlp.up_proj.g_idx": "model-00020-of-00057.safetensors", + "model.layers.42.mlp.up_proj.qweight": "model-00020-of-00057.safetensors", + "model.layers.42.mlp.up_proj.qzeros": "model-00020-of-00057.safetensors", + "model.layers.42.mlp.up_proj.scales": "model-00020-of-00057.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00020-of-00057.safetensors", + "model.layers.42.self_attn.k_proj.g_idx": "model-00020-of-00057.safetensors", + "model.layers.42.self_attn.k_proj.qweight": "model-00020-of-00057.safetensors", + "model.layers.42.self_attn.k_proj.qzeros": "model-00020-of-00057.safetensors", + "model.layers.42.self_attn.k_proj.scales": "model-00020-of-00057.safetensors", + "model.layers.42.self_attn.o_proj.g_idx": "model-00020-of-00057.safetensors", + "model.layers.42.self_attn.o_proj.qweight": "model-00020-of-00057.safetensors", + "model.layers.42.self_attn.o_proj.qzeros": "model-00020-of-00057.safetensors", + "model.layers.42.self_attn.o_proj.scales": "model-00020-of-00057.safetensors", + "model.layers.42.self_attn.q_proj.g_idx": "model-00020-of-00057.safetensors", + "model.layers.42.self_attn.q_proj.qweight": "model-00020-of-00057.safetensors", + "model.layers.42.self_attn.q_proj.qzeros": "model-00020-of-00057.safetensors", + "model.layers.42.self_attn.q_proj.scales": "model-00020-of-00057.safetensors", + "model.layers.42.self_attn.v_proj.g_idx": "model-00020-of-00057.safetensors", + "model.layers.42.self_attn.v_proj.qweight": "model-00020-of-00057.safetensors", + "model.layers.42.self_attn.v_proj.qzeros": "model-00020-of-00057.safetensors", + "model.layers.42.self_attn.v_proj.scales": "model-00020-of-00057.safetensors", + "model.layers.43.input_layernorm.weight": "model-00020-of-00057.safetensors", + "model.layers.43.mlp.down_proj.g_idx": "model-00020-of-00057.safetensors", + "model.layers.43.mlp.down_proj.qweight": "model-00020-of-00057.safetensors", + "model.layers.43.mlp.down_proj.qzeros": "model-00020-of-00057.safetensors", + "model.layers.43.mlp.down_proj.scales": "model-00020-of-00057.safetensors", + "model.layers.43.mlp.gate_proj.g_idx": "model-00020-of-00057.safetensors", + "model.layers.43.mlp.gate_proj.qweight": "model-00020-of-00057.safetensors", + "model.layers.43.mlp.gate_proj.qzeros": "model-00020-of-00057.safetensors", + "model.layers.43.mlp.gate_proj.scales": "model-00020-of-00057.safetensors", + "model.layers.43.mlp.up_proj.g_idx": "model-00020-of-00057.safetensors", + "model.layers.43.mlp.up_proj.qweight": "model-00020-of-00057.safetensors", + "model.layers.43.mlp.up_proj.qzeros": "model-00020-of-00057.safetensors", + "model.layers.43.mlp.up_proj.scales": "model-00020-of-00057.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00020-of-00057.safetensors", + "model.layers.43.self_attn.k_proj.g_idx": "model-00020-of-00057.safetensors", + "model.layers.43.self_attn.k_proj.qweight": "model-00020-of-00057.safetensors", + "model.layers.43.self_attn.k_proj.qzeros": "model-00020-of-00057.safetensors", + "model.layers.43.self_attn.k_proj.scales": "model-00020-of-00057.safetensors", + "model.layers.43.self_attn.o_proj.g_idx": "model-00020-of-00057.safetensors", + "model.layers.43.self_attn.o_proj.qweight": "model-00020-of-00057.safetensors", + "model.layers.43.self_attn.o_proj.qzeros": "model-00020-of-00057.safetensors", + "model.layers.43.self_attn.o_proj.scales": "model-00020-of-00057.safetensors", + "model.layers.43.self_attn.q_proj.g_idx": "model-00020-of-00057.safetensors", + "model.layers.43.self_attn.q_proj.qweight": "model-00020-of-00057.safetensors", + "model.layers.43.self_attn.q_proj.qzeros": "model-00020-of-00057.safetensors", + "model.layers.43.self_attn.q_proj.scales": "model-00020-of-00057.safetensors", + "model.layers.43.self_attn.v_proj.g_idx": "model-00020-of-00057.safetensors", + "model.layers.43.self_attn.v_proj.qweight": "model-00020-of-00057.safetensors", + "model.layers.43.self_attn.v_proj.qzeros": "model-00020-of-00057.safetensors", + "model.layers.43.self_attn.v_proj.scales": "model-00020-of-00057.safetensors", + "model.layers.44.input_layernorm.weight": "model-00021-of-00057.safetensors", + "model.layers.44.mlp.down_proj.g_idx": "model-00021-of-00057.safetensors", + "model.layers.44.mlp.down_proj.qweight": "model-00021-of-00057.safetensors", + "model.layers.44.mlp.down_proj.qzeros": "model-00021-of-00057.safetensors", + "model.layers.44.mlp.down_proj.scales": "model-00021-of-00057.safetensors", + "model.layers.44.mlp.gate_proj.g_idx": "model-00021-of-00057.safetensors", + "model.layers.44.mlp.gate_proj.qweight": "model-00021-of-00057.safetensors", + "model.layers.44.mlp.gate_proj.qzeros": "model-00021-of-00057.safetensors", + "model.layers.44.mlp.gate_proj.scales": "model-00021-of-00057.safetensors", + "model.layers.44.mlp.up_proj.g_idx": "model-00021-of-00057.safetensors", + "model.layers.44.mlp.up_proj.qweight": "model-00021-of-00057.safetensors", + "model.layers.44.mlp.up_proj.qzeros": "model-00021-of-00057.safetensors", + "model.layers.44.mlp.up_proj.scales": "model-00021-of-00057.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00021-of-00057.safetensors", + "model.layers.44.self_attn.k_proj.g_idx": "model-00020-of-00057.safetensors", + "model.layers.44.self_attn.k_proj.qweight": "model-00020-of-00057.safetensors", + "model.layers.44.self_attn.k_proj.qzeros": "model-00020-of-00057.safetensors", + "model.layers.44.self_attn.k_proj.scales": "model-00020-of-00057.safetensors", + "model.layers.44.self_attn.o_proj.g_idx": "model-00021-of-00057.safetensors", + "model.layers.44.self_attn.o_proj.qweight": "model-00021-of-00057.safetensors", + "model.layers.44.self_attn.o_proj.qzeros": "model-00021-of-00057.safetensors", + "model.layers.44.self_attn.o_proj.scales": "model-00021-of-00057.safetensors", + "model.layers.44.self_attn.q_proj.g_idx": "model-00020-of-00057.safetensors", + "model.layers.44.self_attn.q_proj.qweight": "model-00020-of-00057.safetensors", + "model.layers.44.self_attn.q_proj.qzeros": "model-00020-of-00057.safetensors", + "model.layers.44.self_attn.q_proj.scales": "model-00020-of-00057.safetensors", + "model.layers.44.self_attn.v_proj.g_idx": "model-00020-of-00057.safetensors", + "model.layers.44.self_attn.v_proj.qweight": "model-00020-of-00057.safetensors", + "model.layers.44.self_attn.v_proj.qzeros": "model-00020-of-00057.safetensors", + "model.layers.44.self_attn.v_proj.scales": "model-00020-of-00057.safetensors", + "model.layers.45.input_layernorm.weight": "model-00021-of-00057.safetensors", + "model.layers.45.mlp.down_proj.g_idx": "model-00021-of-00057.safetensors", + "model.layers.45.mlp.down_proj.qweight": "model-00021-of-00057.safetensors", + "model.layers.45.mlp.down_proj.qzeros": "model-00021-of-00057.safetensors", + "model.layers.45.mlp.down_proj.scales": "model-00021-of-00057.safetensors", + "model.layers.45.mlp.gate_proj.g_idx": "model-00021-of-00057.safetensors", + "model.layers.45.mlp.gate_proj.qweight": "model-00021-of-00057.safetensors", + "model.layers.45.mlp.gate_proj.qzeros": "model-00021-of-00057.safetensors", + "model.layers.45.mlp.gate_proj.scales": "model-00021-of-00057.safetensors", + "model.layers.45.mlp.up_proj.g_idx": "model-00021-of-00057.safetensors", + "model.layers.45.mlp.up_proj.qweight": "model-00021-of-00057.safetensors", + "model.layers.45.mlp.up_proj.qzeros": "model-00021-of-00057.safetensors", + "model.layers.45.mlp.up_proj.scales": "model-00021-of-00057.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00021-of-00057.safetensors", + "model.layers.45.self_attn.k_proj.g_idx": "model-00021-of-00057.safetensors", + "model.layers.45.self_attn.k_proj.qweight": "model-00021-of-00057.safetensors", + "model.layers.45.self_attn.k_proj.qzeros": "model-00021-of-00057.safetensors", + "model.layers.45.self_attn.k_proj.scales": "model-00021-of-00057.safetensors", + "model.layers.45.self_attn.o_proj.g_idx": "model-00021-of-00057.safetensors", + "model.layers.45.self_attn.o_proj.qweight": "model-00021-of-00057.safetensors", + "model.layers.45.self_attn.o_proj.qzeros": "model-00021-of-00057.safetensors", + "model.layers.45.self_attn.o_proj.scales": "model-00021-of-00057.safetensors", + "model.layers.45.self_attn.q_proj.g_idx": "model-00021-of-00057.safetensors", + "model.layers.45.self_attn.q_proj.qweight": "model-00021-of-00057.safetensors", + "model.layers.45.self_attn.q_proj.qzeros": "model-00021-of-00057.safetensors", + "model.layers.45.self_attn.q_proj.scales": "model-00021-of-00057.safetensors", + "model.layers.45.self_attn.v_proj.g_idx": "model-00021-of-00057.safetensors", + "model.layers.45.self_attn.v_proj.qweight": "model-00021-of-00057.safetensors", + "model.layers.45.self_attn.v_proj.qzeros": "model-00021-of-00057.safetensors", + "model.layers.45.self_attn.v_proj.scales": "model-00021-of-00057.safetensors", + "model.layers.46.input_layernorm.weight": "model-00022-of-00057.safetensors", + "model.layers.46.mlp.down_proj.g_idx": "model-00022-of-00057.safetensors", + "model.layers.46.mlp.down_proj.qweight": "model-00022-of-00057.safetensors", + "model.layers.46.mlp.down_proj.qzeros": "model-00022-of-00057.safetensors", + "model.layers.46.mlp.down_proj.scales": "model-00022-of-00057.safetensors", + "model.layers.46.mlp.gate_proj.g_idx": "model-00021-of-00057.safetensors", + "model.layers.46.mlp.gate_proj.qweight": "model-00021-of-00057.safetensors", + "model.layers.46.mlp.gate_proj.qzeros": "model-00021-of-00057.safetensors", + "model.layers.46.mlp.gate_proj.scales": "model-00021-of-00057.safetensors", + "model.layers.46.mlp.up_proj.g_idx": "model-00022-of-00057.safetensors", + "model.layers.46.mlp.up_proj.qweight": "model-00022-of-00057.safetensors", + "model.layers.46.mlp.up_proj.qzeros": "model-00022-of-00057.safetensors", + "model.layers.46.mlp.up_proj.scales": "model-00022-of-00057.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00022-of-00057.safetensors", + "model.layers.46.self_attn.k_proj.g_idx": "model-00021-of-00057.safetensors", + "model.layers.46.self_attn.k_proj.qweight": "model-00021-of-00057.safetensors", + "model.layers.46.self_attn.k_proj.qzeros": "model-00021-of-00057.safetensors", + "model.layers.46.self_attn.k_proj.scales": "model-00021-of-00057.safetensors", + "model.layers.46.self_attn.o_proj.g_idx": "model-00021-of-00057.safetensors", + "model.layers.46.self_attn.o_proj.qweight": "model-00021-of-00057.safetensors", + "model.layers.46.self_attn.o_proj.qzeros": "model-00021-of-00057.safetensors", + "model.layers.46.self_attn.o_proj.scales": "model-00021-of-00057.safetensors", + "model.layers.46.self_attn.q_proj.g_idx": "model-00021-of-00057.safetensors", + "model.layers.46.self_attn.q_proj.qweight": "model-00021-of-00057.safetensors", + "model.layers.46.self_attn.q_proj.qzeros": "model-00021-of-00057.safetensors", + "model.layers.46.self_attn.q_proj.scales": "model-00021-of-00057.safetensors", + "model.layers.46.self_attn.v_proj.g_idx": "model-00021-of-00057.safetensors", + "model.layers.46.self_attn.v_proj.qweight": "model-00021-of-00057.safetensors", + "model.layers.46.self_attn.v_proj.qzeros": "model-00021-of-00057.safetensors", + "model.layers.46.self_attn.v_proj.scales": "model-00021-of-00057.safetensors", + "model.layers.47.input_layernorm.weight": "model-00022-of-00057.safetensors", + "model.layers.47.mlp.down_proj.g_idx": "model-00022-of-00057.safetensors", + "model.layers.47.mlp.down_proj.qweight": "model-00022-of-00057.safetensors", + "model.layers.47.mlp.down_proj.qzeros": "model-00022-of-00057.safetensors", + "model.layers.47.mlp.down_proj.scales": "model-00022-of-00057.safetensors", + "model.layers.47.mlp.gate_proj.g_idx": "model-00022-of-00057.safetensors", + "model.layers.47.mlp.gate_proj.qweight": "model-00022-of-00057.safetensors", + "model.layers.47.mlp.gate_proj.qzeros": "model-00022-of-00057.safetensors", + "model.layers.47.mlp.gate_proj.scales": "model-00022-of-00057.safetensors", + "model.layers.47.mlp.up_proj.g_idx": "model-00022-of-00057.safetensors", + "model.layers.47.mlp.up_proj.qweight": "model-00022-of-00057.safetensors", + "model.layers.47.mlp.up_proj.qzeros": "model-00022-of-00057.safetensors", + "model.layers.47.mlp.up_proj.scales": "model-00022-of-00057.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00022-of-00057.safetensors", + "model.layers.47.self_attn.k_proj.g_idx": "model-00022-of-00057.safetensors", + "model.layers.47.self_attn.k_proj.qweight": "model-00022-of-00057.safetensors", + "model.layers.47.self_attn.k_proj.qzeros": "model-00022-of-00057.safetensors", + "model.layers.47.self_attn.k_proj.scales": "model-00022-of-00057.safetensors", + "model.layers.47.self_attn.o_proj.g_idx": "model-00022-of-00057.safetensors", + "model.layers.47.self_attn.o_proj.qweight": "model-00022-of-00057.safetensors", + "model.layers.47.self_attn.o_proj.qzeros": "model-00022-of-00057.safetensors", + "model.layers.47.self_attn.o_proj.scales": "model-00022-of-00057.safetensors", + "model.layers.47.self_attn.q_proj.g_idx": "model-00022-of-00057.safetensors", + "model.layers.47.self_attn.q_proj.qweight": "model-00022-of-00057.safetensors", + "model.layers.47.self_attn.q_proj.qzeros": "model-00022-of-00057.safetensors", + "model.layers.47.self_attn.q_proj.scales": "model-00022-of-00057.safetensors", + "model.layers.47.self_attn.v_proj.g_idx": "model-00022-of-00057.safetensors", + "model.layers.47.self_attn.v_proj.qweight": "model-00022-of-00057.safetensors", + "model.layers.47.self_attn.v_proj.qzeros": "model-00022-of-00057.safetensors", + "model.layers.47.self_attn.v_proj.scales": "model-00022-of-00057.safetensors", + "model.layers.48.input_layernorm.weight": "model-00023-of-00057.safetensors", + "model.layers.48.mlp.down_proj.g_idx": "model-00023-of-00057.safetensors", + "model.layers.48.mlp.down_proj.qweight": "model-00023-of-00057.safetensors", + "model.layers.48.mlp.down_proj.qzeros": "model-00023-of-00057.safetensors", + "model.layers.48.mlp.down_proj.scales": "model-00023-of-00057.safetensors", + "model.layers.48.mlp.gate_proj.g_idx": "model-00022-of-00057.safetensors", + "model.layers.48.mlp.gate_proj.qweight": "model-00022-of-00057.safetensors", + "model.layers.48.mlp.gate_proj.qzeros": "model-00022-of-00057.safetensors", + "model.layers.48.mlp.gate_proj.scales": "model-00022-of-00057.safetensors", + "model.layers.48.mlp.up_proj.g_idx": "model-00022-of-00057.safetensors", + "model.layers.48.mlp.up_proj.qweight": "model-00022-of-00057.safetensors", + "model.layers.48.mlp.up_proj.qzeros": "model-00022-of-00057.safetensors", + "model.layers.48.mlp.up_proj.scales": "model-00022-of-00057.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00023-of-00057.safetensors", + "model.layers.48.self_attn.k_proj.g_idx": "model-00022-of-00057.safetensors", + "model.layers.48.self_attn.k_proj.qweight": "model-00022-of-00057.safetensors", + "model.layers.48.self_attn.k_proj.qzeros": "model-00022-of-00057.safetensors", + "model.layers.48.self_attn.k_proj.scales": "model-00022-of-00057.safetensors", + "model.layers.48.self_attn.o_proj.g_idx": "model-00022-of-00057.safetensors", + "model.layers.48.self_attn.o_proj.qweight": "model-00022-of-00057.safetensors", + "model.layers.48.self_attn.o_proj.qzeros": "model-00022-of-00057.safetensors", + "model.layers.48.self_attn.o_proj.scales": "model-00022-of-00057.safetensors", + "model.layers.48.self_attn.q_proj.g_idx": "model-00022-of-00057.safetensors", + "model.layers.48.self_attn.q_proj.qweight": "model-00022-of-00057.safetensors", + "model.layers.48.self_attn.q_proj.qzeros": "model-00022-of-00057.safetensors", + "model.layers.48.self_attn.q_proj.scales": "model-00022-of-00057.safetensors", + "model.layers.48.self_attn.v_proj.g_idx": "model-00022-of-00057.safetensors", + "model.layers.48.self_attn.v_proj.qweight": "model-00022-of-00057.safetensors", + "model.layers.48.self_attn.v_proj.qzeros": "model-00022-of-00057.safetensors", + "model.layers.48.self_attn.v_proj.scales": "model-00022-of-00057.safetensors", + "model.layers.49.input_layernorm.weight": "model-00023-of-00057.safetensors", + "model.layers.49.mlp.down_proj.g_idx": "model-00023-of-00057.safetensors", + "model.layers.49.mlp.down_proj.qweight": "model-00023-of-00057.safetensors", + "model.layers.49.mlp.down_proj.qzeros": "model-00023-of-00057.safetensors", + "model.layers.49.mlp.down_proj.scales": "model-00023-of-00057.safetensors", + "model.layers.49.mlp.gate_proj.g_idx": "model-00023-of-00057.safetensors", + "model.layers.49.mlp.gate_proj.qweight": "model-00023-of-00057.safetensors", + "model.layers.49.mlp.gate_proj.qzeros": "model-00023-of-00057.safetensors", + "model.layers.49.mlp.gate_proj.scales": "model-00023-of-00057.safetensors", + "model.layers.49.mlp.up_proj.g_idx": "model-00023-of-00057.safetensors", + "model.layers.49.mlp.up_proj.qweight": "model-00023-of-00057.safetensors", + "model.layers.49.mlp.up_proj.qzeros": "model-00023-of-00057.safetensors", + "model.layers.49.mlp.up_proj.scales": "model-00023-of-00057.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00023-of-00057.safetensors", + "model.layers.49.self_attn.k_proj.g_idx": "model-00023-of-00057.safetensors", + "model.layers.49.self_attn.k_proj.qweight": "model-00023-of-00057.safetensors", + "model.layers.49.self_attn.k_proj.qzeros": "model-00023-of-00057.safetensors", + "model.layers.49.self_attn.k_proj.scales": "model-00023-of-00057.safetensors", + "model.layers.49.self_attn.o_proj.g_idx": "model-00023-of-00057.safetensors", + "model.layers.49.self_attn.o_proj.qweight": "model-00023-of-00057.safetensors", + "model.layers.49.self_attn.o_proj.qzeros": "model-00023-of-00057.safetensors", + "model.layers.49.self_attn.o_proj.scales": "model-00023-of-00057.safetensors", + "model.layers.49.self_attn.q_proj.g_idx": "model-00023-of-00057.safetensors", + "model.layers.49.self_attn.q_proj.qweight": "model-00023-of-00057.safetensors", + "model.layers.49.self_attn.q_proj.qzeros": "model-00023-of-00057.safetensors", + "model.layers.49.self_attn.q_proj.scales": "model-00023-of-00057.safetensors", + "model.layers.49.self_attn.v_proj.g_idx": "model-00023-of-00057.safetensors", + "model.layers.49.self_attn.v_proj.qweight": "model-00023-of-00057.safetensors", + "model.layers.49.self_attn.v_proj.qzeros": "model-00023-of-00057.safetensors", + "model.layers.49.self_attn.v_proj.scales": "model-00023-of-00057.safetensors", + "model.layers.5.input_layernorm.weight": "model-00004-of-00057.safetensors", + "model.layers.5.mlp.down_proj.g_idx": "model-00004-of-00057.safetensors", + "model.layers.5.mlp.down_proj.qweight": "model-00004-of-00057.safetensors", + "model.layers.5.mlp.down_proj.qzeros": "model-00004-of-00057.safetensors", + "model.layers.5.mlp.down_proj.scales": "model-00004-of-00057.safetensors", + "model.layers.5.mlp.gate_proj.g_idx": "model-00004-of-00057.safetensors", + "model.layers.5.mlp.gate_proj.qweight": "model-00004-of-00057.safetensors", + "model.layers.5.mlp.gate_proj.qzeros": "model-00004-of-00057.safetensors", + "model.layers.5.mlp.gate_proj.scales": "model-00004-of-00057.safetensors", + "model.layers.5.mlp.up_proj.g_idx": "model-00004-of-00057.safetensors", + "model.layers.5.mlp.up_proj.qweight": "model-00004-of-00057.safetensors", + "model.layers.5.mlp.up_proj.qzeros": "model-00004-of-00057.safetensors", + "model.layers.5.mlp.up_proj.scales": "model-00004-of-00057.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00004-of-00057.safetensors", + "model.layers.5.self_attn.k_proj.g_idx": "model-00004-of-00057.safetensors", + "model.layers.5.self_attn.k_proj.qweight": "model-00004-of-00057.safetensors", + "model.layers.5.self_attn.k_proj.qzeros": "model-00004-of-00057.safetensors", + "model.layers.5.self_attn.k_proj.scales": "model-00004-of-00057.safetensors", + "model.layers.5.self_attn.o_proj.g_idx": "model-00004-of-00057.safetensors", + "model.layers.5.self_attn.o_proj.qweight": "model-00004-of-00057.safetensors", + "model.layers.5.self_attn.o_proj.qzeros": "model-00004-of-00057.safetensors", + "model.layers.5.self_attn.o_proj.scales": "model-00004-of-00057.safetensors", + "model.layers.5.self_attn.q_proj.g_idx": "model-00004-of-00057.safetensors", + "model.layers.5.self_attn.q_proj.qweight": "model-00004-of-00057.safetensors", + "model.layers.5.self_attn.q_proj.qzeros": "model-00004-of-00057.safetensors", + "model.layers.5.self_attn.q_proj.scales": "model-00004-of-00057.safetensors", + "model.layers.5.self_attn.v_proj.g_idx": "model-00004-of-00057.safetensors", + "model.layers.5.self_attn.v_proj.qweight": "model-00004-of-00057.safetensors", + "model.layers.5.self_attn.v_proj.qzeros": "model-00004-of-00057.safetensors", + "model.layers.5.self_attn.v_proj.scales": "model-00004-of-00057.safetensors", + "model.layers.50.input_layernorm.weight": "model-00023-of-00057.safetensors", + "model.layers.50.mlp.down_proj.g_idx": "model-00023-of-00057.safetensors", + "model.layers.50.mlp.down_proj.qweight": "model-00023-of-00057.safetensors", + "model.layers.50.mlp.down_proj.qzeros": "model-00023-of-00057.safetensors", + "model.layers.50.mlp.down_proj.scales": "model-00023-of-00057.safetensors", + "model.layers.50.mlp.gate_proj.g_idx": "model-00023-of-00057.safetensors", + "model.layers.50.mlp.gate_proj.qweight": "model-00023-of-00057.safetensors", + "model.layers.50.mlp.gate_proj.qzeros": "model-00023-of-00057.safetensors", + "model.layers.50.mlp.gate_proj.scales": "model-00023-of-00057.safetensors", + "model.layers.50.mlp.up_proj.g_idx": "model-00023-of-00057.safetensors", + "model.layers.50.mlp.up_proj.qweight": "model-00023-of-00057.safetensors", + "model.layers.50.mlp.up_proj.qzeros": "model-00023-of-00057.safetensors", + "model.layers.50.mlp.up_proj.scales": "model-00023-of-00057.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00023-of-00057.safetensors", + "model.layers.50.self_attn.k_proj.g_idx": "model-00023-of-00057.safetensors", + "model.layers.50.self_attn.k_proj.qweight": "model-00023-of-00057.safetensors", + "model.layers.50.self_attn.k_proj.qzeros": "model-00023-of-00057.safetensors", + "model.layers.50.self_attn.k_proj.scales": "model-00023-of-00057.safetensors", + "model.layers.50.self_attn.o_proj.g_idx": "model-00023-of-00057.safetensors", + "model.layers.50.self_attn.o_proj.qweight": "model-00023-of-00057.safetensors", + "model.layers.50.self_attn.o_proj.qzeros": "model-00023-of-00057.safetensors", + "model.layers.50.self_attn.o_proj.scales": "model-00023-of-00057.safetensors", + "model.layers.50.self_attn.q_proj.g_idx": "model-00023-of-00057.safetensors", + "model.layers.50.self_attn.q_proj.qweight": "model-00023-of-00057.safetensors", + "model.layers.50.self_attn.q_proj.qzeros": "model-00023-of-00057.safetensors", + "model.layers.50.self_attn.q_proj.scales": "model-00023-of-00057.safetensors", + "model.layers.50.self_attn.v_proj.g_idx": "model-00023-of-00057.safetensors", + "model.layers.50.self_attn.v_proj.qweight": "model-00023-of-00057.safetensors", + "model.layers.50.self_attn.v_proj.qzeros": "model-00023-of-00057.safetensors", + "model.layers.50.self_attn.v_proj.scales": "model-00023-of-00057.safetensors", + "model.layers.51.input_layernorm.weight": "model-00024-of-00057.safetensors", + "model.layers.51.mlp.down_proj.g_idx": "model-00024-of-00057.safetensors", + "model.layers.51.mlp.down_proj.qweight": "model-00024-of-00057.safetensors", + "model.layers.51.mlp.down_proj.qzeros": "model-00024-of-00057.safetensors", + "model.layers.51.mlp.down_proj.scales": "model-00024-of-00057.safetensors", + "model.layers.51.mlp.gate_proj.g_idx": "model-00024-of-00057.safetensors", + "model.layers.51.mlp.gate_proj.qweight": "model-00024-of-00057.safetensors", + "model.layers.51.mlp.gate_proj.qzeros": "model-00024-of-00057.safetensors", + "model.layers.51.mlp.gate_proj.scales": "model-00024-of-00057.safetensors", + "model.layers.51.mlp.up_proj.g_idx": "model-00024-of-00057.safetensors", + "model.layers.51.mlp.up_proj.qweight": "model-00024-of-00057.safetensors", + "model.layers.51.mlp.up_proj.qzeros": "model-00024-of-00057.safetensors", + "model.layers.51.mlp.up_proj.scales": "model-00024-of-00057.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00024-of-00057.safetensors", + "model.layers.51.self_attn.k_proj.g_idx": "model-00023-of-00057.safetensors", + "model.layers.51.self_attn.k_proj.qweight": "model-00023-of-00057.safetensors", + "model.layers.51.self_attn.k_proj.qzeros": "model-00023-of-00057.safetensors", + "model.layers.51.self_attn.k_proj.scales": "model-00023-of-00057.safetensors", + "model.layers.51.self_attn.o_proj.g_idx": "model-00024-of-00057.safetensors", + "model.layers.51.self_attn.o_proj.qweight": "model-00024-of-00057.safetensors", + "model.layers.51.self_attn.o_proj.qzeros": "model-00024-of-00057.safetensors", + "model.layers.51.self_attn.o_proj.scales": "model-00024-of-00057.safetensors", + "model.layers.51.self_attn.q_proj.g_idx": "model-00023-of-00057.safetensors", + "model.layers.51.self_attn.q_proj.qweight": "model-00023-of-00057.safetensors", + "model.layers.51.self_attn.q_proj.qzeros": "model-00023-of-00057.safetensors", + "model.layers.51.self_attn.q_proj.scales": "model-00023-of-00057.safetensors", + "model.layers.51.self_attn.v_proj.g_idx": "model-00023-of-00057.safetensors", + "model.layers.51.self_attn.v_proj.qweight": "model-00023-of-00057.safetensors", + "model.layers.51.self_attn.v_proj.qzeros": "model-00023-of-00057.safetensors", + "model.layers.51.self_attn.v_proj.scales": "model-00023-of-00057.safetensors", + "model.layers.52.input_layernorm.weight": "model-00024-of-00057.safetensors", + "model.layers.52.mlp.down_proj.g_idx": "model-00024-of-00057.safetensors", + "model.layers.52.mlp.down_proj.qweight": "model-00024-of-00057.safetensors", + "model.layers.52.mlp.down_proj.qzeros": "model-00024-of-00057.safetensors", + "model.layers.52.mlp.down_proj.scales": "model-00024-of-00057.safetensors", + "model.layers.52.mlp.gate_proj.g_idx": "model-00024-of-00057.safetensors", + "model.layers.52.mlp.gate_proj.qweight": "model-00024-of-00057.safetensors", + "model.layers.52.mlp.gate_proj.qzeros": "model-00024-of-00057.safetensors", + "model.layers.52.mlp.gate_proj.scales": "model-00024-of-00057.safetensors", + "model.layers.52.mlp.up_proj.g_idx": "model-00024-of-00057.safetensors", + "model.layers.52.mlp.up_proj.qweight": "model-00024-of-00057.safetensors", + "model.layers.52.mlp.up_proj.qzeros": "model-00024-of-00057.safetensors", + "model.layers.52.mlp.up_proj.scales": "model-00024-of-00057.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00024-of-00057.safetensors", + "model.layers.52.self_attn.k_proj.g_idx": "model-00024-of-00057.safetensors", + "model.layers.52.self_attn.k_proj.qweight": "model-00024-of-00057.safetensors", + "model.layers.52.self_attn.k_proj.qzeros": "model-00024-of-00057.safetensors", + "model.layers.52.self_attn.k_proj.scales": "model-00024-of-00057.safetensors", + "model.layers.52.self_attn.o_proj.g_idx": "model-00024-of-00057.safetensors", + "model.layers.52.self_attn.o_proj.qweight": "model-00024-of-00057.safetensors", + "model.layers.52.self_attn.o_proj.qzeros": "model-00024-of-00057.safetensors", + "model.layers.52.self_attn.o_proj.scales": "model-00024-of-00057.safetensors", + "model.layers.52.self_attn.q_proj.g_idx": "model-00024-of-00057.safetensors", + "model.layers.52.self_attn.q_proj.qweight": "model-00024-of-00057.safetensors", + "model.layers.52.self_attn.q_proj.qzeros": "model-00024-of-00057.safetensors", + "model.layers.52.self_attn.q_proj.scales": "model-00024-of-00057.safetensors", + "model.layers.52.self_attn.v_proj.g_idx": "model-00024-of-00057.safetensors", + "model.layers.52.self_attn.v_proj.qweight": "model-00024-of-00057.safetensors", + "model.layers.52.self_attn.v_proj.qzeros": "model-00024-of-00057.safetensors", + "model.layers.52.self_attn.v_proj.scales": "model-00024-of-00057.safetensors", + "model.layers.53.input_layernorm.weight": "model-00025-of-00057.safetensors", + "model.layers.53.mlp.down_proj.g_idx": "model-00025-of-00057.safetensors", + "model.layers.53.mlp.down_proj.qweight": "model-00025-of-00057.safetensors", + "model.layers.53.mlp.down_proj.qzeros": "model-00025-of-00057.safetensors", + "model.layers.53.mlp.down_proj.scales": "model-00025-of-00057.safetensors", + "model.layers.53.mlp.gate_proj.g_idx": "model-00024-of-00057.safetensors", + "model.layers.53.mlp.gate_proj.qweight": "model-00024-of-00057.safetensors", + "model.layers.53.mlp.gate_proj.qzeros": "model-00024-of-00057.safetensors", + "model.layers.53.mlp.gate_proj.scales": "model-00024-of-00057.safetensors", + "model.layers.53.mlp.up_proj.g_idx": "model-00025-of-00057.safetensors", + "model.layers.53.mlp.up_proj.qweight": "model-00025-of-00057.safetensors", + "model.layers.53.mlp.up_proj.qzeros": "model-00025-of-00057.safetensors", + "model.layers.53.mlp.up_proj.scales": "model-00025-of-00057.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00025-of-00057.safetensors", + "model.layers.53.self_attn.k_proj.g_idx": "model-00024-of-00057.safetensors", + "model.layers.53.self_attn.k_proj.qweight": "model-00024-of-00057.safetensors", + "model.layers.53.self_attn.k_proj.qzeros": "model-00024-of-00057.safetensors", + "model.layers.53.self_attn.k_proj.scales": "model-00024-of-00057.safetensors", + "model.layers.53.self_attn.o_proj.g_idx": "model-00024-of-00057.safetensors", + "model.layers.53.self_attn.o_proj.qweight": "model-00024-of-00057.safetensors", + "model.layers.53.self_attn.o_proj.qzeros": "model-00024-of-00057.safetensors", + "model.layers.53.self_attn.o_proj.scales": "model-00024-of-00057.safetensors", + "model.layers.53.self_attn.q_proj.g_idx": "model-00024-of-00057.safetensors", + "model.layers.53.self_attn.q_proj.qweight": "model-00024-of-00057.safetensors", + "model.layers.53.self_attn.q_proj.qzeros": "model-00024-of-00057.safetensors", + "model.layers.53.self_attn.q_proj.scales": "model-00024-of-00057.safetensors", + "model.layers.53.self_attn.v_proj.g_idx": "model-00024-of-00057.safetensors", + "model.layers.53.self_attn.v_proj.qweight": "model-00024-of-00057.safetensors", + "model.layers.53.self_attn.v_proj.qzeros": "model-00024-of-00057.safetensors", + "model.layers.53.self_attn.v_proj.scales": "model-00024-of-00057.safetensors", + "model.layers.54.input_layernorm.weight": "model-00025-of-00057.safetensors", + "model.layers.54.mlp.down_proj.g_idx": "model-00025-of-00057.safetensors", + "model.layers.54.mlp.down_proj.qweight": "model-00025-of-00057.safetensors", + "model.layers.54.mlp.down_proj.qzeros": "model-00025-of-00057.safetensors", + "model.layers.54.mlp.down_proj.scales": "model-00025-of-00057.safetensors", + "model.layers.54.mlp.gate_proj.g_idx": "model-00025-of-00057.safetensors", + "model.layers.54.mlp.gate_proj.qweight": "model-00025-of-00057.safetensors", + "model.layers.54.mlp.gate_proj.qzeros": "model-00025-of-00057.safetensors", + "model.layers.54.mlp.gate_proj.scales": "model-00025-of-00057.safetensors", + "model.layers.54.mlp.up_proj.g_idx": "model-00025-of-00057.safetensors", + "model.layers.54.mlp.up_proj.qweight": "model-00025-of-00057.safetensors", + "model.layers.54.mlp.up_proj.qzeros": "model-00025-of-00057.safetensors", + "model.layers.54.mlp.up_proj.scales": "model-00025-of-00057.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00025-of-00057.safetensors", + "model.layers.54.self_attn.k_proj.g_idx": "model-00025-of-00057.safetensors", + "model.layers.54.self_attn.k_proj.qweight": "model-00025-of-00057.safetensors", + "model.layers.54.self_attn.k_proj.qzeros": "model-00025-of-00057.safetensors", + "model.layers.54.self_attn.k_proj.scales": "model-00025-of-00057.safetensors", + "model.layers.54.self_attn.o_proj.g_idx": "model-00025-of-00057.safetensors", + "model.layers.54.self_attn.o_proj.qweight": "model-00025-of-00057.safetensors", + "model.layers.54.self_attn.o_proj.qzeros": "model-00025-of-00057.safetensors", + "model.layers.54.self_attn.o_proj.scales": "model-00025-of-00057.safetensors", + "model.layers.54.self_attn.q_proj.g_idx": "model-00025-of-00057.safetensors", + "model.layers.54.self_attn.q_proj.qweight": "model-00025-of-00057.safetensors", + "model.layers.54.self_attn.q_proj.qzeros": "model-00025-of-00057.safetensors", + "model.layers.54.self_attn.q_proj.scales": "model-00025-of-00057.safetensors", + "model.layers.54.self_attn.v_proj.g_idx": "model-00025-of-00057.safetensors", + "model.layers.54.self_attn.v_proj.qweight": "model-00025-of-00057.safetensors", + "model.layers.54.self_attn.v_proj.qzeros": "model-00025-of-00057.safetensors", + "model.layers.54.self_attn.v_proj.scales": "model-00025-of-00057.safetensors", + "model.layers.55.input_layernorm.weight": "model-00026-of-00057.safetensors", + "model.layers.55.mlp.down_proj.g_idx": "model-00026-of-00057.safetensors", + "model.layers.55.mlp.down_proj.qweight": "model-00026-of-00057.safetensors", + "model.layers.55.mlp.down_proj.qzeros": "model-00026-of-00057.safetensors", + "model.layers.55.mlp.down_proj.scales": "model-00026-of-00057.safetensors", + "model.layers.55.mlp.gate_proj.g_idx": "model-00025-of-00057.safetensors", + "model.layers.55.mlp.gate_proj.qweight": "model-00025-of-00057.safetensors", + "model.layers.55.mlp.gate_proj.qzeros": "model-00025-of-00057.safetensors", + "model.layers.55.mlp.gate_proj.scales": "model-00025-of-00057.safetensors", + "model.layers.55.mlp.up_proj.g_idx": "model-00025-of-00057.safetensors", + "model.layers.55.mlp.up_proj.qweight": "model-00025-of-00057.safetensors", + "model.layers.55.mlp.up_proj.qzeros": "model-00025-of-00057.safetensors", + "model.layers.55.mlp.up_proj.scales": "model-00025-of-00057.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00026-of-00057.safetensors", + "model.layers.55.self_attn.k_proj.g_idx": "model-00025-of-00057.safetensors", + "model.layers.55.self_attn.k_proj.qweight": "model-00025-of-00057.safetensors", + "model.layers.55.self_attn.k_proj.qzeros": "model-00025-of-00057.safetensors", + "model.layers.55.self_attn.k_proj.scales": "model-00025-of-00057.safetensors", + "model.layers.55.self_attn.o_proj.g_idx": "model-00025-of-00057.safetensors", + "model.layers.55.self_attn.o_proj.qweight": "model-00025-of-00057.safetensors", + "model.layers.55.self_attn.o_proj.qzeros": "model-00025-of-00057.safetensors", + "model.layers.55.self_attn.o_proj.scales": "model-00025-of-00057.safetensors", + "model.layers.55.self_attn.q_proj.g_idx": "model-00025-of-00057.safetensors", + "model.layers.55.self_attn.q_proj.qweight": "model-00025-of-00057.safetensors", + "model.layers.55.self_attn.q_proj.qzeros": "model-00025-of-00057.safetensors", + "model.layers.55.self_attn.q_proj.scales": "model-00025-of-00057.safetensors", + "model.layers.55.self_attn.v_proj.g_idx": "model-00025-of-00057.safetensors", + "model.layers.55.self_attn.v_proj.qweight": "model-00025-of-00057.safetensors", + "model.layers.55.self_attn.v_proj.qzeros": "model-00025-of-00057.safetensors", + "model.layers.55.self_attn.v_proj.scales": "model-00025-of-00057.safetensors", + "model.layers.56.input_layernorm.weight": "model-00026-of-00057.safetensors", + "model.layers.56.mlp.down_proj.g_idx": "model-00026-of-00057.safetensors", + "model.layers.56.mlp.down_proj.qweight": "model-00026-of-00057.safetensors", + "model.layers.56.mlp.down_proj.qzeros": "model-00026-of-00057.safetensors", + "model.layers.56.mlp.down_proj.scales": "model-00026-of-00057.safetensors", + "model.layers.56.mlp.gate_proj.g_idx": "model-00026-of-00057.safetensors", + "model.layers.56.mlp.gate_proj.qweight": "model-00026-of-00057.safetensors", + "model.layers.56.mlp.gate_proj.qzeros": "model-00026-of-00057.safetensors", + "model.layers.56.mlp.gate_proj.scales": "model-00026-of-00057.safetensors", + "model.layers.56.mlp.up_proj.g_idx": "model-00026-of-00057.safetensors", + "model.layers.56.mlp.up_proj.qweight": "model-00026-of-00057.safetensors", + "model.layers.56.mlp.up_proj.qzeros": "model-00026-of-00057.safetensors", + "model.layers.56.mlp.up_proj.scales": "model-00026-of-00057.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00026-of-00057.safetensors", + "model.layers.56.self_attn.k_proj.g_idx": "model-00026-of-00057.safetensors", + "model.layers.56.self_attn.k_proj.qweight": "model-00026-of-00057.safetensors", + "model.layers.56.self_attn.k_proj.qzeros": "model-00026-of-00057.safetensors", + "model.layers.56.self_attn.k_proj.scales": "model-00026-of-00057.safetensors", + "model.layers.56.self_attn.o_proj.g_idx": "model-00026-of-00057.safetensors", + "model.layers.56.self_attn.o_proj.qweight": "model-00026-of-00057.safetensors", + "model.layers.56.self_attn.o_proj.qzeros": "model-00026-of-00057.safetensors", + "model.layers.56.self_attn.o_proj.scales": "model-00026-of-00057.safetensors", + "model.layers.56.self_attn.q_proj.g_idx": "model-00026-of-00057.safetensors", + "model.layers.56.self_attn.q_proj.qweight": "model-00026-of-00057.safetensors", + "model.layers.56.self_attn.q_proj.qzeros": "model-00026-of-00057.safetensors", + "model.layers.56.self_attn.q_proj.scales": "model-00026-of-00057.safetensors", + "model.layers.56.self_attn.v_proj.g_idx": "model-00026-of-00057.safetensors", + "model.layers.56.self_attn.v_proj.qweight": "model-00026-of-00057.safetensors", + "model.layers.56.self_attn.v_proj.qzeros": "model-00026-of-00057.safetensors", + "model.layers.56.self_attn.v_proj.scales": "model-00026-of-00057.safetensors", + "model.layers.57.input_layernorm.weight": "model-00026-of-00057.safetensors", + "model.layers.57.mlp.down_proj.g_idx": "model-00026-of-00057.safetensors", + "model.layers.57.mlp.down_proj.qweight": "model-00026-of-00057.safetensors", + "model.layers.57.mlp.down_proj.qzeros": "model-00026-of-00057.safetensors", + "model.layers.57.mlp.down_proj.scales": "model-00026-of-00057.safetensors", + "model.layers.57.mlp.gate_proj.g_idx": "model-00026-of-00057.safetensors", + "model.layers.57.mlp.gate_proj.qweight": "model-00026-of-00057.safetensors", + "model.layers.57.mlp.gate_proj.qzeros": "model-00026-of-00057.safetensors", + "model.layers.57.mlp.gate_proj.scales": "model-00026-of-00057.safetensors", + "model.layers.57.mlp.up_proj.g_idx": "model-00026-of-00057.safetensors", + "model.layers.57.mlp.up_proj.qweight": "model-00026-of-00057.safetensors", + "model.layers.57.mlp.up_proj.qzeros": "model-00026-of-00057.safetensors", + "model.layers.57.mlp.up_proj.scales": "model-00026-of-00057.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00026-of-00057.safetensors", + "model.layers.57.self_attn.k_proj.g_idx": "model-00026-of-00057.safetensors", + "model.layers.57.self_attn.k_proj.qweight": "model-00026-of-00057.safetensors", + "model.layers.57.self_attn.k_proj.qzeros": "model-00026-of-00057.safetensors", + "model.layers.57.self_attn.k_proj.scales": "model-00026-of-00057.safetensors", + "model.layers.57.self_attn.o_proj.g_idx": "model-00026-of-00057.safetensors", + "model.layers.57.self_attn.o_proj.qweight": "model-00026-of-00057.safetensors", + "model.layers.57.self_attn.o_proj.qzeros": "model-00026-of-00057.safetensors", + "model.layers.57.self_attn.o_proj.scales": "model-00026-of-00057.safetensors", + "model.layers.57.self_attn.q_proj.g_idx": "model-00026-of-00057.safetensors", + "model.layers.57.self_attn.q_proj.qweight": "model-00026-of-00057.safetensors", + "model.layers.57.self_attn.q_proj.qzeros": "model-00026-of-00057.safetensors", + "model.layers.57.self_attn.q_proj.scales": "model-00026-of-00057.safetensors", + "model.layers.57.self_attn.v_proj.g_idx": "model-00026-of-00057.safetensors", + "model.layers.57.self_attn.v_proj.qweight": "model-00026-of-00057.safetensors", + "model.layers.57.self_attn.v_proj.qzeros": "model-00026-of-00057.safetensors", + "model.layers.57.self_attn.v_proj.scales": "model-00026-of-00057.safetensors", + "model.layers.58.input_layernorm.weight": "model-00027-of-00057.safetensors", + "model.layers.58.mlp.down_proj.g_idx": "model-00027-of-00057.safetensors", + "model.layers.58.mlp.down_proj.qweight": "model-00027-of-00057.safetensors", + "model.layers.58.mlp.down_proj.qzeros": "model-00027-of-00057.safetensors", + "model.layers.58.mlp.down_proj.scales": "model-00027-of-00057.safetensors", + "model.layers.58.mlp.gate_proj.g_idx": "model-00027-of-00057.safetensors", + "model.layers.58.mlp.gate_proj.qweight": "model-00027-of-00057.safetensors", + "model.layers.58.mlp.gate_proj.qzeros": "model-00027-of-00057.safetensors", + "model.layers.58.mlp.gate_proj.scales": "model-00027-of-00057.safetensors", + "model.layers.58.mlp.up_proj.g_idx": "model-00027-of-00057.safetensors", + "model.layers.58.mlp.up_proj.qweight": "model-00027-of-00057.safetensors", + "model.layers.58.mlp.up_proj.qzeros": "model-00027-of-00057.safetensors", + "model.layers.58.mlp.up_proj.scales": "model-00027-of-00057.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00027-of-00057.safetensors", + "model.layers.58.self_attn.k_proj.g_idx": "model-00026-of-00057.safetensors", + "model.layers.58.self_attn.k_proj.qweight": "model-00026-of-00057.safetensors", + "model.layers.58.self_attn.k_proj.qzeros": "model-00026-of-00057.safetensors", + "model.layers.58.self_attn.k_proj.scales": "model-00026-of-00057.safetensors", + "model.layers.58.self_attn.o_proj.g_idx": "model-00027-of-00057.safetensors", + "model.layers.58.self_attn.o_proj.qweight": "model-00027-of-00057.safetensors", + "model.layers.58.self_attn.o_proj.qzeros": "model-00027-of-00057.safetensors", + "model.layers.58.self_attn.o_proj.scales": "model-00027-of-00057.safetensors", + "model.layers.58.self_attn.q_proj.g_idx": "model-00026-of-00057.safetensors", + "model.layers.58.self_attn.q_proj.qweight": "model-00026-of-00057.safetensors", + "model.layers.58.self_attn.q_proj.qzeros": "model-00026-of-00057.safetensors", + "model.layers.58.self_attn.q_proj.scales": "model-00026-of-00057.safetensors", + "model.layers.58.self_attn.v_proj.g_idx": "model-00026-of-00057.safetensors", + "model.layers.58.self_attn.v_proj.qweight": "model-00026-of-00057.safetensors", + "model.layers.58.self_attn.v_proj.qzeros": "model-00026-of-00057.safetensors", + "model.layers.58.self_attn.v_proj.scales": "model-00026-of-00057.safetensors", + "model.layers.59.input_layernorm.weight": "model-00027-of-00057.safetensors", + "model.layers.59.mlp.down_proj.g_idx": "model-00027-of-00057.safetensors", + "model.layers.59.mlp.down_proj.qweight": "model-00027-of-00057.safetensors", + "model.layers.59.mlp.down_proj.qzeros": "model-00027-of-00057.safetensors", + "model.layers.59.mlp.down_proj.scales": "model-00027-of-00057.safetensors", + "model.layers.59.mlp.gate_proj.g_idx": "model-00027-of-00057.safetensors", + "model.layers.59.mlp.gate_proj.qweight": "model-00027-of-00057.safetensors", + "model.layers.59.mlp.gate_proj.qzeros": "model-00027-of-00057.safetensors", + "model.layers.59.mlp.gate_proj.scales": "model-00027-of-00057.safetensors", + "model.layers.59.mlp.up_proj.g_idx": "model-00027-of-00057.safetensors", + "model.layers.59.mlp.up_proj.qweight": "model-00027-of-00057.safetensors", + "model.layers.59.mlp.up_proj.qzeros": "model-00027-of-00057.safetensors", + "model.layers.59.mlp.up_proj.scales": "model-00027-of-00057.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00027-of-00057.safetensors", + "model.layers.59.self_attn.k_proj.g_idx": "model-00027-of-00057.safetensors", + "model.layers.59.self_attn.k_proj.qweight": "model-00027-of-00057.safetensors", + "model.layers.59.self_attn.k_proj.qzeros": "model-00027-of-00057.safetensors", + "model.layers.59.self_attn.k_proj.scales": "model-00027-of-00057.safetensors", + "model.layers.59.self_attn.o_proj.g_idx": "model-00027-of-00057.safetensors", + "model.layers.59.self_attn.o_proj.qweight": "model-00027-of-00057.safetensors", + "model.layers.59.self_attn.o_proj.qzeros": "model-00027-of-00057.safetensors", + "model.layers.59.self_attn.o_proj.scales": "model-00027-of-00057.safetensors", + "model.layers.59.self_attn.q_proj.g_idx": "model-00027-of-00057.safetensors", + "model.layers.59.self_attn.q_proj.qweight": "model-00027-of-00057.safetensors", + "model.layers.59.self_attn.q_proj.qzeros": "model-00027-of-00057.safetensors", + "model.layers.59.self_attn.q_proj.scales": "model-00027-of-00057.safetensors", + "model.layers.59.self_attn.v_proj.g_idx": "model-00027-of-00057.safetensors", + "model.layers.59.self_attn.v_proj.qweight": "model-00027-of-00057.safetensors", + "model.layers.59.self_attn.v_proj.qzeros": "model-00027-of-00057.safetensors", + "model.layers.59.self_attn.v_proj.scales": "model-00027-of-00057.safetensors", + "model.layers.6.input_layernorm.weight": "model-00005-of-00057.safetensors", + "model.layers.6.mlp.down_proj.g_idx": "model-00005-of-00057.safetensors", + "model.layers.6.mlp.down_proj.qweight": "model-00005-of-00057.safetensors", + "model.layers.6.mlp.down_proj.qzeros": "model-00005-of-00057.safetensors", + "model.layers.6.mlp.down_proj.scales": "model-00005-of-00057.safetensors", + "model.layers.6.mlp.gate_proj.g_idx": "model-00004-of-00057.safetensors", + "model.layers.6.mlp.gate_proj.qweight": "model-00004-of-00057.safetensors", + "model.layers.6.mlp.gate_proj.qzeros": "model-00004-of-00057.safetensors", + "model.layers.6.mlp.gate_proj.scales": "model-00004-of-00057.safetensors", + "model.layers.6.mlp.up_proj.g_idx": "model-00004-of-00057.safetensors", + "model.layers.6.mlp.up_proj.qweight": "model-00004-of-00057.safetensors", + "model.layers.6.mlp.up_proj.qzeros": "model-00004-of-00057.safetensors", + "model.layers.6.mlp.up_proj.scales": "model-00004-of-00057.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00005-of-00057.safetensors", + "model.layers.6.self_attn.k_proj.g_idx": "model-00004-of-00057.safetensors", + "model.layers.6.self_attn.k_proj.qweight": "model-00004-of-00057.safetensors", + "model.layers.6.self_attn.k_proj.qzeros": "model-00004-of-00057.safetensors", + "model.layers.6.self_attn.k_proj.scales": "model-00004-of-00057.safetensors", + "model.layers.6.self_attn.o_proj.g_idx": "model-00004-of-00057.safetensors", + "model.layers.6.self_attn.o_proj.qweight": "model-00004-of-00057.safetensors", + "model.layers.6.self_attn.o_proj.qzeros": "model-00004-of-00057.safetensors", + "model.layers.6.self_attn.o_proj.scales": "model-00004-of-00057.safetensors", + "model.layers.6.self_attn.q_proj.g_idx": "model-00004-of-00057.safetensors", + "model.layers.6.self_attn.q_proj.qweight": "model-00004-of-00057.safetensors", + "model.layers.6.self_attn.q_proj.qzeros": "model-00004-of-00057.safetensors", + "model.layers.6.self_attn.q_proj.scales": "model-00004-of-00057.safetensors", + "model.layers.6.self_attn.v_proj.g_idx": "model-00004-of-00057.safetensors", + "model.layers.6.self_attn.v_proj.qweight": "model-00004-of-00057.safetensors", + "model.layers.6.self_attn.v_proj.qzeros": "model-00004-of-00057.safetensors", + "model.layers.6.self_attn.v_proj.scales": "model-00004-of-00057.safetensors", + "model.layers.60.input_layernorm.weight": "model-00028-of-00057.safetensors", + "model.layers.60.mlp.down_proj.g_idx": "model-00028-of-00057.safetensors", + "model.layers.60.mlp.down_proj.qweight": "model-00028-of-00057.safetensors", + "model.layers.60.mlp.down_proj.qzeros": "model-00028-of-00057.safetensors", + "model.layers.60.mlp.down_proj.scales": "model-00028-of-00057.safetensors", + "model.layers.60.mlp.gate_proj.g_idx": "model-00027-of-00057.safetensors", + "model.layers.60.mlp.gate_proj.qweight": "model-00027-of-00057.safetensors", + "model.layers.60.mlp.gate_proj.qzeros": "model-00027-of-00057.safetensors", + "model.layers.60.mlp.gate_proj.scales": "model-00027-of-00057.safetensors", + "model.layers.60.mlp.up_proj.g_idx": "model-00028-of-00057.safetensors", + "model.layers.60.mlp.up_proj.qweight": "model-00028-of-00057.safetensors", + "model.layers.60.mlp.up_proj.qzeros": "model-00028-of-00057.safetensors", + "model.layers.60.mlp.up_proj.scales": "model-00028-of-00057.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00028-of-00057.safetensors", + "model.layers.60.self_attn.k_proj.g_idx": "model-00027-of-00057.safetensors", + "model.layers.60.self_attn.k_proj.qweight": "model-00027-of-00057.safetensors", + "model.layers.60.self_attn.k_proj.qzeros": "model-00027-of-00057.safetensors", + "model.layers.60.self_attn.k_proj.scales": "model-00027-of-00057.safetensors", + "model.layers.60.self_attn.o_proj.g_idx": "model-00027-of-00057.safetensors", + "model.layers.60.self_attn.o_proj.qweight": "model-00027-of-00057.safetensors", + "model.layers.60.self_attn.o_proj.qzeros": "model-00027-of-00057.safetensors", + "model.layers.60.self_attn.o_proj.scales": "model-00027-of-00057.safetensors", + "model.layers.60.self_attn.q_proj.g_idx": "model-00027-of-00057.safetensors", + "model.layers.60.self_attn.q_proj.qweight": "model-00027-of-00057.safetensors", + "model.layers.60.self_attn.q_proj.qzeros": "model-00027-of-00057.safetensors", + "model.layers.60.self_attn.q_proj.scales": "model-00027-of-00057.safetensors", + "model.layers.60.self_attn.v_proj.g_idx": "model-00027-of-00057.safetensors", + "model.layers.60.self_attn.v_proj.qweight": "model-00027-of-00057.safetensors", + "model.layers.60.self_attn.v_proj.qzeros": "model-00027-of-00057.safetensors", + "model.layers.60.self_attn.v_proj.scales": "model-00027-of-00057.safetensors", + "model.layers.61.input_layernorm.weight": "model-00028-of-00057.safetensors", + "model.layers.61.mlp.down_proj.g_idx": "model-00028-of-00057.safetensors", + "model.layers.61.mlp.down_proj.qweight": "model-00028-of-00057.safetensors", + "model.layers.61.mlp.down_proj.qzeros": "model-00028-of-00057.safetensors", + "model.layers.61.mlp.down_proj.scales": "model-00028-of-00057.safetensors", + "model.layers.61.mlp.gate_proj.g_idx": "model-00028-of-00057.safetensors", + "model.layers.61.mlp.gate_proj.qweight": "model-00028-of-00057.safetensors", + "model.layers.61.mlp.gate_proj.qzeros": "model-00028-of-00057.safetensors", + "model.layers.61.mlp.gate_proj.scales": "model-00028-of-00057.safetensors", + "model.layers.61.mlp.up_proj.g_idx": "model-00028-of-00057.safetensors", + "model.layers.61.mlp.up_proj.qweight": "model-00028-of-00057.safetensors", + "model.layers.61.mlp.up_proj.qzeros": "model-00028-of-00057.safetensors", + "model.layers.61.mlp.up_proj.scales": "model-00028-of-00057.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00028-of-00057.safetensors", + "model.layers.61.self_attn.k_proj.g_idx": "model-00028-of-00057.safetensors", + "model.layers.61.self_attn.k_proj.qweight": "model-00028-of-00057.safetensors", + "model.layers.61.self_attn.k_proj.qzeros": "model-00028-of-00057.safetensors", + "model.layers.61.self_attn.k_proj.scales": "model-00028-of-00057.safetensors", + "model.layers.61.self_attn.o_proj.g_idx": "model-00028-of-00057.safetensors", + "model.layers.61.self_attn.o_proj.qweight": "model-00028-of-00057.safetensors", + "model.layers.61.self_attn.o_proj.qzeros": "model-00028-of-00057.safetensors", + "model.layers.61.self_attn.o_proj.scales": "model-00028-of-00057.safetensors", + "model.layers.61.self_attn.q_proj.g_idx": "model-00028-of-00057.safetensors", + "model.layers.61.self_attn.q_proj.qweight": "model-00028-of-00057.safetensors", + "model.layers.61.self_attn.q_proj.qzeros": "model-00028-of-00057.safetensors", + "model.layers.61.self_attn.q_proj.scales": "model-00028-of-00057.safetensors", + "model.layers.61.self_attn.v_proj.g_idx": "model-00028-of-00057.safetensors", + "model.layers.61.self_attn.v_proj.qweight": "model-00028-of-00057.safetensors", + "model.layers.61.self_attn.v_proj.qzeros": "model-00028-of-00057.safetensors", + "model.layers.61.self_attn.v_proj.scales": "model-00028-of-00057.safetensors", + "model.layers.62.input_layernorm.weight": "model-00029-of-00057.safetensors", + "model.layers.62.mlp.down_proj.g_idx": "model-00029-of-00057.safetensors", + "model.layers.62.mlp.down_proj.qweight": "model-00029-of-00057.safetensors", + "model.layers.62.mlp.down_proj.qzeros": "model-00029-of-00057.safetensors", + "model.layers.62.mlp.down_proj.scales": "model-00029-of-00057.safetensors", + "model.layers.62.mlp.gate_proj.g_idx": "model-00028-of-00057.safetensors", + "model.layers.62.mlp.gate_proj.qweight": "model-00028-of-00057.safetensors", + "model.layers.62.mlp.gate_proj.qzeros": "model-00028-of-00057.safetensors", + "model.layers.62.mlp.gate_proj.scales": "model-00028-of-00057.safetensors", + "model.layers.62.mlp.up_proj.g_idx": "model-00028-of-00057.safetensors", + "model.layers.62.mlp.up_proj.qweight": "model-00028-of-00057.safetensors", + "model.layers.62.mlp.up_proj.qzeros": "model-00028-of-00057.safetensors", + "model.layers.62.mlp.up_proj.scales": "model-00028-of-00057.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00029-of-00057.safetensors", + "model.layers.62.self_attn.k_proj.g_idx": "model-00028-of-00057.safetensors", + "model.layers.62.self_attn.k_proj.qweight": "model-00028-of-00057.safetensors", + "model.layers.62.self_attn.k_proj.qzeros": "model-00028-of-00057.safetensors", + "model.layers.62.self_attn.k_proj.scales": "model-00028-of-00057.safetensors", + "model.layers.62.self_attn.o_proj.g_idx": "model-00028-of-00057.safetensors", + "model.layers.62.self_attn.o_proj.qweight": "model-00028-of-00057.safetensors", + "model.layers.62.self_attn.o_proj.qzeros": "model-00028-of-00057.safetensors", + "model.layers.62.self_attn.o_proj.scales": "model-00028-of-00057.safetensors", + "model.layers.62.self_attn.q_proj.g_idx": "model-00028-of-00057.safetensors", + "model.layers.62.self_attn.q_proj.qweight": "model-00028-of-00057.safetensors", + "model.layers.62.self_attn.q_proj.qzeros": "model-00028-of-00057.safetensors", + "model.layers.62.self_attn.q_proj.scales": "model-00028-of-00057.safetensors", + "model.layers.62.self_attn.v_proj.g_idx": "model-00028-of-00057.safetensors", + "model.layers.62.self_attn.v_proj.qweight": "model-00028-of-00057.safetensors", + "model.layers.62.self_attn.v_proj.qzeros": "model-00028-of-00057.safetensors", + "model.layers.62.self_attn.v_proj.scales": "model-00028-of-00057.safetensors", + "model.layers.63.input_layernorm.weight": "model-00029-of-00057.safetensors", + "model.layers.63.mlp.down_proj.g_idx": "model-00029-of-00057.safetensors", + "model.layers.63.mlp.down_proj.qweight": "model-00029-of-00057.safetensors", + "model.layers.63.mlp.down_proj.qzeros": "model-00029-of-00057.safetensors", + "model.layers.63.mlp.down_proj.scales": "model-00029-of-00057.safetensors", + "model.layers.63.mlp.gate_proj.g_idx": "model-00029-of-00057.safetensors", + "model.layers.63.mlp.gate_proj.qweight": "model-00029-of-00057.safetensors", + "model.layers.63.mlp.gate_proj.qzeros": "model-00029-of-00057.safetensors", + "model.layers.63.mlp.gate_proj.scales": "model-00029-of-00057.safetensors", + "model.layers.63.mlp.up_proj.g_idx": "model-00029-of-00057.safetensors", + "model.layers.63.mlp.up_proj.qweight": "model-00029-of-00057.safetensors", + "model.layers.63.mlp.up_proj.qzeros": "model-00029-of-00057.safetensors", + "model.layers.63.mlp.up_proj.scales": "model-00029-of-00057.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00029-of-00057.safetensors", + "model.layers.63.self_attn.k_proj.g_idx": "model-00029-of-00057.safetensors", + "model.layers.63.self_attn.k_proj.qweight": "model-00029-of-00057.safetensors", + "model.layers.63.self_attn.k_proj.qzeros": "model-00029-of-00057.safetensors", + "model.layers.63.self_attn.k_proj.scales": "model-00029-of-00057.safetensors", + "model.layers.63.self_attn.o_proj.g_idx": "model-00029-of-00057.safetensors", + "model.layers.63.self_attn.o_proj.qweight": "model-00029-of-00057.safetensors", + "model.layers.63.self_attn.o_proj.qzeros": "model-00029-of-00057.safetensors", + "model.layers.63.self_attn.o_proj.scales": "model-00029-of-00057.safetensors", + "model.layers.63.self_attn.q_proj.g_idx": "model-00029-of-00057.safetensors", + "model.layers.63.self_attn.q_proj.qweight": "model-00029-of-00057.safetensors", + "model.layers.63.self_attn.q_proj.qzeros": "model-00029-of-00057.safetensors", + "model.layers.63.self_attn.q_proj.scales": "model-00029-of-00057.safetensors", + "model.layers.63.self_attn.v_proj.g_idx": "model-00029-of-00057.safetensors", + "model.layers.63.self_attn.v_proj.qweight": "model-00029-of-00057.safetensors", + "model.layers.63.self_attn.v_proj.qzeros": "model-00029-of-00057.safetensors", + "model.layers.63.self_attn.v_proj.scales": "model-00029-of-00057.safetensors", + "model.layers.64.input_layernorm.weight": "model-00029-of-00057.safetensors", + "model.layers.64.mlp.down_proj.g_idx": "model-00029-of-00057.safetensors", + "model.layers.64.mlp.down_proj.qweight": "model-00029-of-00057.safetensors", + "model.layers.64.mlp.down_proj.qzeros": "model-00029-of-00057.safetensors", + "model.layers.64.mlp.down_proj.scales": "model-00029-of-00057.safetensors", + "model.layers.64.mlp.gate_proj.g_idx": "model-00029-of-00057.safetensors", + "model.layers.64.mlp.gate_proj.qweight": "model-00029-of-00057.safetensors", + "model.layers.64.mlp.gate_proj.qzeros": "model-00029-of-00057.safetensors", + "model.layers.64.mlp.gate_proj.scales": "model-00029-of-00057.safetensors", + "model.layers.64.mlp.up_proj.g_idx": "model-00029-of-00057.safetensors", + "model.layers.64.mlp.up_proj.qweight": "model-00029-of-00057.safetensors", + "model.layers.64.mlp.up_proj.qzeros": "model-00029-of-00057.safetensors", + "model.layers.64.mlp.up_proj.scales": "model-00029-of-00057.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00029-of-00057.safetensors", + "model.layers.64.self_attn.k_proj.g_idx": "model-00029-of-00057.safetensors", + "model.layers.64.self_attn.k_proj.qweight": "model-00029-of-00057.safetensors", + "model.layers.64.self_attn.k_proj.qzeros": "model-00029-of-00057.safetensors", + "model.layers.64.self_attn.k_proj.scales": "model-00029-of-00057.safetensors", + "model.layers.64.self_attn.o_proj.g_idx": "model-00029-of-00057.safetensors", + "model.layers.64.self_attn.o_proj.qweight": "model-00029-of-00057.safetensors", + "model.layers.64.self_attn.o_proj.qzeros": "model-00029-of-00057.safetensors", + "model.layers.64.self_attn.o_proj.scales": "model-00029-of-00057.safetensors", + "model.layers.64.self_attn.q_proj.g_idx": "model-00029-of-00057.safetensors", + "model.layers.64.self_attn.q_proj.qweight": "model-00029-of-00057.safetensors", + "model.layers.64.self_attn.q_proj.qzeros": "model-00029-of-00057.safetensors", + "model.layers.64.self_attn.q_proj.scales": "model-00029-of-00057.safetensors", + "model.layers.64.self_attn.v_proj.g_idx": "model-00029-of-00057.safetensors", + "model.layers.64.self_attn.v_proj.qweight": "model-00029-of-00057.safetensors", + "model.layers.64.self_attn.v_proj.qzeros": "model-00029-of-00057.safetensors", + "model.layers.64.self_attn.v_proj.scales": "model-00029-of-00057.safetensors", + "model.layers.65.input_layernorm.weight": "model-00030-of-00057.safetensors", + "model.layers.65.mlp.down_proj.g_idx": "model-00030-of-00057.safetensors", + "model.layers.65.mlp.down_proj.qweight": "model-00030-of-00057.safetensors", + "model.layers.65.mlp.down_proj.qzeros": "model-00030-of-00057.safetensors", + "model.layers.65.mlp.down_proj.scales": "model-00030-of-00057.safetensors", + "model.layers.65.mlp.gate_proj.g_idx": "model-00030-of-00057.safetensors", + "model.layers.65.mlp.gate_proj.qweight": "model-00030-of-00057.safetensors", + "model.layers.65.mlp.gate_proj.qzeros": "model-00030-of-00057.safetensors", + "model.layers.65.mlp.gate_proj.scales": "model-00030-of-00057.safetensors", + "model.layers.65.mlp.up_proj.g_idx": "model-00030-of-00057.safetensors", + "model.layers.65.mlp.up_proj.qweight": "model-00030-of-00057.safetensors", + "model.layers.65.mlp.up_proj.qzeros": "model-00030-of-00057.safetensors", + "model.layers.65.mlp.up_proj.scales": "model-00030-of-00057.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00030-of-00057.safetensors", + "model.layers.65.self_attn.k_proj.g_idx": "model-00029-of-00057.safetensors", + "model.layers.65.self_attn.k_proj.qweight": "model-00029-of-00057.safetensors", + "model.layers.65.self_attn.k_proj.qzeros": "model-00029-of-00057.safetensors", + "model.layers.65.self_attn.k_proj.scales": "model-00029-of-00057.safetensors", + "model.layers.65.self_attn.o_proj.g_idx": "model-00030-of-00057.safetensors", + "model.layers.65.self_attn.o_proj.qweight": "model-00030-of-00057.safetensors", + "model.layers.65.self_attn.o_proj.qzeros": "model-00030-of-00057.safetensors", + "model.layers.65.self_attn.o_proj.scales": "model-00030-of-00057.safetensors", + "model.layers.65.self_attn.q_proj.g_idx": "model-00029-of-00057.safetensors", + "model.layers.65.self_attn.q_proj.qweight": "model-00029-of-00057.safetensors", + "model.layers.65.self_attn.q_proj.qzeros": "model-00029-of-00057.safetensors", + "model.layers.65.self_attn.q_proj.scales": "model-00029-of-00057.safetensors", + "model.layers.65.self_attn.v_proj.g_idx": "model-00029-of-00057.safetensors", + "model.layers.65.self_attn.v_proj.qweight": "model-00029-of-00057.safetensors", + "model.layers.65.self_attn.v_proj.qzeros": "model-00029-of-00057.safetensors", + "model.layers.65.self_attn.v_proj.scales": "model-00029-of-00057.safetensors", + "model.layers.66.input_layernorm.weight": "model-00030-of-00057.safetensors", + "model.layers.66.mlp.down_proj.g_idx": "model-00030-of-00057.safetensors", + "model.layers.66.mlp.down_proj.qweight": "model-00030-of-00057.safetensors", + "model.layers.66.mlp.down_proj.qzeros": "model-00030-of-00057.safetensors", + "model.layers.66.mlp.down_proj.scales": "model-00030-of-00057.safetensors", + "model.layers.66.mlp.gate_proj.g_idx": "model-00030-of-00057.safetensors", + "model.layers.66.mlp.gate_proj.qweight": "model-00030-of-00057.safetensors", + "model.layers.66.mlp.gate_proj.qzeros": "model-00030-of-00057.safetensors", + "model.layers.66.mlp.gate_proj.scales": "model-00030-of-00057.safetensors", + "model.layers.66.mlp.up_proj.g_idx": "model-00030-of-00057.safetensors", + "model.layers.66.mlp.up_proj.qweight": "model-00030-of-00057.safetensors", + "model.layers.66.mlp.up_proj.qzeros": "model-00030-of-00057.safetensors", + "model.layers.66.mlp.up_proj.scales": "model-00030-of-00057.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00030-of-00057.safetensors", + "model.layers.66.self_attn.k_proj.g_idx": "model-00030-of-00057.safetensors", + "model.layers.66.self_attn.k_proj.qweight": "model-00030-of-00057.safetensors", + "model.layers.66.self_attn.k_proj.qzeros": "model-00030-of-00057.safetensors", + "model.layers.66.self_attn.k_proj.scales": "model-00030-of-00057.safetensors", + "model.layers.66.self_attn.o_proj.g_idx": "model-00030-of-00057.safetensors", + "model.layers.66.self_attn.o_proj.qweight": "model-00030-of-00057.safetensors", + "model.layers.66.self_attn.o_proj.qzeros": "model-00030-of-00057.safetensors", + "model.layers.66.self_attn.o_proj.scales": "model-00030-of-00057.safetensors", + "model.layers.66.self_attn.q_proj.g_idx": "model-00030-of-00057.safetensors", + "model.layers.66.self_attn.q_proj.qweight": "model-00030-of-00057.safetensors", + "model.layers.66.self_attn.q_proj.qzeros": "model-00030-of-00057.safetensors", + "model.layers.66.self_attn.q_proj.scales": "model-00030-of-00057.safetensors", + "model.layers.66.self_attn.v_proj.g_idx": "model-00030-of-00057.safetensors", + "model.layers.66.self_attn.v_proj.qweight": "model-00030-of-00057.safetensors", + "model.layers.66.self_attn.v_proj.qzeros": "model-00030-of-00057.safetensors", + "model.layers.66.self_attn.v_proj.scales": "model-00030-of-00057.safetensors", + "model.layers.67.input_layernorm.weight": "model-00031-of-00057.safetensors", + "model.layers.67.mlp.down_proj.g_idx": "model-00031-of-00057.safetensors", + "model.layers.67.mlp.down_proj.qweight": "model-00031-of-00057.safetensors", + "model.layers.67.mlp.down_proj.qzeros": "model-00031-of-00057.safetensors", + "model.layers.67.mlp.down_proj.scales": "model-00031-of-00057.safetensors", + "model.layers.67.mlp.gate_proj.g_idx": "model-00030-of-00057.safetensors", + "model.layers.67.mlp.gate_proj.qweight": "model-00030-of-00057.safetensors", + "model.layers.67.mlp.gate_proj.qzeros": "model-00030-of-00057.safetensors", + "model.layers.67.mlp.gate_proj.scales": "model-00030-of-00057.safetensors", + "model.layers.67.mlp.up_proj.g_idx": "model-00031-of-00057.safetensors", + "model.layers.67.mlp.up_proj.qweight": "model-00031-of-00057.safetensors", + "model.layers.67.mlp.up_proj.qzeros": "model-00031-of-00057.safetensors", + "model.layers.67.mlp.up_proj.scales": "model-00031-of-00057.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00031-of-00057.safetensors", + "model.layers.67.self_attn.k_proj.g_idx": "model-00030-of-00057.safetensors", + "model.layers.67.self_attn.k_proj.qweight": "model-00030-of-00057.safetensors", + "model.layers.67.self_attn.k_proj.qzeros": "model-00030-of-00057.safetensors", + "model.layers.67.self_attn.k_proj.scales": "model-00030-of-00057.safetensors", + "model.layers.67.self_attn.o_proj.g_idx": "model-00030-of-00057.safetensors", + "model.layers.67.self_attn.o_proj.qweight": "model-00030-of-00057.safetensors", + "model.layers.67.self_attn.o_proj.qzeros": "model-00030-of-00057.safetensors", + "model.layers.67.self_attn.o_proj.scales": "model-00030-of-00057.safetensors", + "model.layers.67.self_attn.q_proj.g_idx": "model-00030-of-00057.safetensors", + "model.layers.67.self_attn.q_proj.qweight": "model-00030-of-00057.safetensors", + "model.layers.67.self_attn.q_proj.qzeros": "model-00030-of-00057.safetensors", + "model.layers.67.self_attn.q_proj.scales": "model-00030-of-00057.safetensors", + "model.layers.67.self_attn.v_proj.g_idx": "model-00030-of-00057.safetensors", + "model.layers.67.self_attn.v_proj.qweight": "model-00030-of-00057.safetensors", + "model.layers.67.self_attn.v_proj.qzeros": "model-00030-of-00057.safetensors", + "model.layers.67.self_attn.v_proj.scales": "model-00030-of-00057.safetensors", + "model.layers.68.input_layernorm.weight": "model-00031-of-00057.safetensors", + "model.layers.68.mlp.down_proj.g_idx": "model-00031-of-00057.safetensors", + "model.layers.68.mlp.down_proj.qweight": "model-00031-of-00057.safetensors", + "model.layers.68.mlp.down_proj.qzeros": "model-00031-of-00057.safetensors", + "model.layers.68.mlp.down_proj.scales": "model-00031-of-00057.safetensors", + "model.layers.68.mlp.gate_proj.g_idx": "model-00031-of-00057.safetensors", + "model.layers.68.mlp.gate_proj.qweight": "model-00031-of-00057.safetensors", + "model.layers.68.mlp.gate_proj.qzeros": "model-00031-of-00057.safetensors", + "model.layers.68.mlp.gate_proj.scales": "model-00031-of-00057.safetensors", + "model.layers.68.mlp.up_proj.g_idx": "model-00031-of-00057.safetensors", + "model.layers.68.mlp.up_proj.qweight": "model-00031-of-00057.safetensors", + "model.layers.68.mlp.up_proj.qzeros": "model-00031-of-00057.safetensors", + "model.layers.68.mlp.up_proj.scales": "model-00031-of-00057.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00031-of-00057.safetensors", + "model.layers.68.self_attn.k_proj.g_idx": "model-00031-of-00057.safetensors", + "model.layers.68.self_attn.k_proj.qweight": "model-00031-of-00057.safetensors", + "model.layers.68.self_attn.k_proj.qzeros": "model-00031-of-00057.safetensors", + "model.layers.68.self_attn.k_proj.scales": "model-00031-of-00057.safetensors", + "model.layers.68.self_attn.o_proj.g_idx": "model-00031-of-00057.safetensors", + "model.layers.68.self_attn.o_proj.qweight": "model-00031-of-00057.safetensors", + "model.layers.68.self_attn.o_proj.qzeros": "model-00031-of-00057.safetensors", + "model.layers.68.self_attn.o_proj.scales": "model-00031-of-00057.safetensors", + "model.layers.68.self_attn.q_proj.g_idx": "model-00031-of-00057.safetensors", + "model.layers.68.self_attn.q_proj.qweight": "model-00031-of-00057.safetensors", + "model.layers.68.self_attn.q_proj.qzeros": "model-00031-of-00057.safetensors", + "model.layers.68.self_attn.q_proj.scales": "model-00031-of-00057.safetensors", + "model.layers.68.self_attn.v_proj.g_idx": "model-00031-of-00057.safetensors", + "model.layers.68.self_attn.v_proj.qweight": "model-00031-of-00057.safetensors", + "model.layers.68.self_attn.v_proj.qzeros": "model-00031-of-00057.safetensors", + "model.layers.68.self_attn.v_proj.scales": "model-00031-of-00057.safetensors", + "model.layers.69.input_layernorm.weight": "model-00032-of-00057.safetensors", + "model.layers.69.mlp.down_proj.g_idx": "model-00032-of-00057.safetensors", + "model.layers.69.mlp.down_proj.qweight": "model-00032-of-00057.safetensors", + "model.layers.69.mlp.down_proj.qzeros": "model-00032-of-00057.safetensors", + "model.layers.69.mlp.down_proj.scales": "model-00032-of-00057.safetensors", + "model.layers.69.mlp.gate_proj.g_idx": "model-00031-of-00057.safetensors", + "model.layers.69.mlp.gate_proj.qweight": "model-00031-of-00057.safetensors", + "model.layers.69.mlp.gate_proj.qzeros": "model-00031-of-00057.safetensors", + "model.layers.69.mlp.gate_proj.scales": "model-00031-of-00057.safetensors", + "model.layers.69.mlp.up_proj.g_idx": "model-00031-of-00057.safetensors", + "model.layers.69.mlp.up_proj.qweight": "model-00031-of-00057.safetensors", + "model.layers.69.mlp.up_proj.qzeros": "model-00031-of-00057.safetensors", + "model.layers.69.mlp.up_proj.scales": "model-00031-of-00057.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00032-of-00057.safetensors", + "model.layers.69.self_attn.k_proj.g_idx": "model-00031-of-00057.safetensors", + "model.layers.69.self_attn.k_proj.qweight": "model-00031-of-00057.safetensors", + "model.layers.69.self_attn.k_proj.qzeros": "model-00031-of-00057.safetensors", + "model.layers.69.self_attn.k_proj.scales": "model-00031-of-00057.safetensors", + "model.layers.69.self_attn.o_proj.g_idx": "model-00031-of-00057.safetensors", + "model.layers.69.self_attn.o_proj.qweight": "model-00031-of-00057.safetensors", + "model.layers.69.self_attn.o_proj.qzeros": "model-00031-of-00057.safetensors", + "model.layers.69.self_attn.o_proj.scales": "model-00031-of-00057.safetensors", + "model.layers.69.self_attn.q_proj.g_idx": "model-00031-of-00057.safetensors", + "model.layers.69.self_attn.q_proj.qweight": "model-00031-of-00057.safetensors", + "model.layers.69.self_attn.q_proj.qzeros": "model-00031-of-00057.safetensors", + "model.layers.69.self_attn.q_proj.scales": "model-00031-of-00057.safetensors", + "model.layers.69.self_attn.v_proj.g_idx": "model-00031-of-00057.safetensors", + "model.layers.69.self_attn.v_proj.qweight": "model-00031-of-00057.safetensors", + "model.layers.69.self_attn.v_proj.qzeros": "model-00031-of-00057.safetensors", + "model.layers.69.self_attn.v_proj.scales": "model-00031-of-00057.safetensors", + "model.layers.7.input_layernorm.weight": "model-00005-of-00057.safetensors", + "model.layers.7.mlp.down_proj.g_idx": "model-00005-of-00057.safetensors", + "model.layers.7.mlp.down_proj.qweight": "model-00005-of-00057.safetensors", + "model.layers.7.mlp.down_proj.qzeros": "model-00005-of-00057.safetensors", + "model.layers.7.mlp.down_proj.scales": "model-00005-of-00057.safetensors", + "model.layers.7.mlp.gate_proj.g_idx": "model-00005-of-00057.safetensors", + "model.layers.7.mlp.gate_proj.qweight": "model-00005-of-00057.safetensors", + "model.layers.7.mlp.gate_proj.qzeros": "model-00005-of-00057.safetensors", + "model.layers.7.mlp.gate_proj.scales": "model-00005-of-00057.safetensors", + "model.layers.7.mlp.up_proj.g_idx": "model-00005-of-00057.safetensors", + "model.layers.7.mlp.up_proj.qweight": "model-00005-of-00057.safetensors", + "model.layers.7.mlp.up_proj.qzeros": "model-00005-of-00057.safetensors", + "model.layers.7.mlp.up_proj.scales": "model-00005-of-00057.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00005-of-00057.safetensors", + "model.layers.7.self_attn.k_proj.g_idx": "model-00005-of-00057.safetensors", + "model.layers.7.self_attn.k_proj.qweight": "model-00005-of-00057.safetensors", + "model.layers.7.self_attn.k_proj.qzeros": "model-00005-of-00057.safetensors", + "model.layers.7.self_attn.k_proj.scales": "model-00005-of-00057.safetensors", + "model.layers.7.self_attn.o_proj.g_idx": "model-00005-of-00057.safetensors", + "model.layers.7.self_attn.o_proj.qweight": "model-00005-of-00057.safetensors", + "model.layers.7.self_attn.o_proj.qzeros": "model-00005-of-00057.safetensors", + "model.layers.7.self_attn.o_proj.scales": "model-00005-of-00057.safetensors", + "model.layers.7.self_attn.q_proj.g_idx": "model-00005-of-00057.safetensors", + "model.layers.7.self_attn.q_proj.qweight": "model-00005-of-00057.safetensors", + "model.layers.7.self_attn.q_proj.qzeros": "model-00005-of-00057.safetensors", + "model.layers.7.self_attn.q_proj.scales": "model-00005-of-00057.safetensors", + "model.layers.7.self_attn.v_proj.g_idx": "model-00005-of-00057.safetensors", + "model.layers.7.self_attn.v_proj.qweight": "model-00005-of-00057.safetensors", + "model.layers.7.self_attn.v_proj.qzeros": "model-00005-of-00057.safetensors", + "model.layers.7.self_attn.v_proj.scales": "model-00005-of-00057.safetensors", + "model.layers.70.input_layernorm.weight": "model-00032-of-00057.safetensors", + "model.layers.70.mlp.down_proj.g_idx": "model-00032-of-00057.safetensors", + "model.layers.70.mlp.down_proj.qweight": "model-00032-of-00057.safetensors", + "model.layers.70.mlp.down_proj.qzeros": "model-00032-of-00057.safetensors", + "model.layers.70.mlp.down_proj.scales": "model-00032-of-00057.safetensors", + "model.layers.70.mlp.gate_proj.g_idx": "model-00032-of-00057.safetensors", + "model.layers.70.mlp.gate_proj.qweight": "model-00032-of-00057.safetensors", + "model.layers.70.mlp.gate_proj.qzeros": "model-00032-of-00057.safetensors", + "model.layers.70.mlp.gate_proj.scales": "model-00032-of-00057.safetensors", + "model.layers.70.mlp.up_proj.g_idx": "model-00032-of-00057.safetensors", + "model.layers.70.mlp.up_proj.qweight": "model-00032-of-00057.safetensors", + "model.layers.70.mlp.up_proj.qzeros": "model-00032-of-00057.safetensors", + "model.layers.70.mlp.up_proj.scales": "model-00032-of-00057.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00032-of-00057.safetensors", + "model.layers.70.self_attn.k_proj.g_idx": "model-00032-of-00057.safetensors", + "model.layers.70.self_attn.k_proj.qweight": "model-00032-of-00057.safetensors", + "model.layers.70.self_attn.k_proj.qzeros": "model-00032-of-00057.safetensors", + "model.layers.70.self_attn.k_proj.scales": "model-00032-of-00057.safetensors", + "model.layers.70.self_attn.o_proj.g_idx": "model-00032-of-00057.safetensors", + "model.layers.70.self_attn.o_proj.qweight": "model-00032-of-00057.safetensors", + "model.layers.70.self_attn.o_proj.qzeros": "model-00032-of-00057.safetensors", + "model.layers.70.self_attn.o_proj.scales": "model-00032-of-00057.safetensors", + "model.layers.70.self_attn.q_proj.g_idx": "model-00032-of-00057.safetensors", + "model.layers.70.self_attn.q_proj.qweight": "model-00032-of-00057.safetensors", + "model.layers.70.self_attn.q_proj.qzeros": "model-00032-of-00057.safetensors", + "model.layers.70.self_attn.q_proj.scales": "model-00032-of-00057.safetensors", + "model.layers.70.self_attn.v_proj.g_idx": "model-00032-of-00057.safetensors", + "model.layers.70.self_attn.v_proj.qweight": "model-00032-of-00057.safetensors", + "model.layers.70.self_attn.v_proj.qzeros": "model-00032-of-00057.safetensors", + "model.layers.70.self_attn.v_proj.scales": "model-00032-of-00057.safetensors", + "model.layers.71.input_layernorm.weight": "model-00032-of-00057.safetensors", + "model.layers.71.mlp.down_proj.g_idx": "model-00032-of-00057.safetensors", + "model.layers.71.mlp.down_proj.qweight": "model-00032-of-00057.safetensors", + "model.layers.71.mlp.down_proj.qzeros": "model-00032-of-00057.safetensors", + "model.layers.71.mlp.down_proj.scales": "model-00032-of-00057.safetensors", + "model.layers.71.mlp.gate_proj.g_idx": "model-00032-of-00057.safetensors", + "model.layers.71.mlp.gate_proj.qweight": "model-00032-of-00057.safetensors", + "model.layers.71.mlp.gate_proj.qzeros": "model-00032-of-00057.safetensors", + "model.layers.71.mlp.gate_proj.scales": "model-00032-of-00057.safetensors", + "model.layers.71.mlp.up_proj.g_idx": "model-00032-of-00057.safetensors", + "model.layers.71.mlp.up_proj.qweight": "model-00032-of-00057.safetensors", + "model.layers.71.mlp.up_proj.qzeros": "model-00032-of-00057.safetensors", + "model.layers.71.mlp.up_proj.scales": "model-00032-of-00057.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00032-of-00057.safetensors", + "model.layers.71.self_attn.k_proj.g_idx": "model-00032-of-00057.safetensors", + "model.layers.71.self_attn.k_proj.qweight": "model-00032-of-00057.safetensors", + "model.layers.71.self_attn.k_proj.qzeros": "model-00032-of-00057.safetensors", + "model.layers.71.self_attn.k_proj.scales": "model-00032-of-00057.safetensors", + "model.layers.71.self_attn.o_proj.g_idx": "model-00032-of-00057.safetensors", + "model.layers.71.self_attn.o_proj.qweight": "model-00032-of-00057.safetensors", + "model.layers.71.self_attn.o_proj.qzeros": "model-00032-of-00057.safetensors", + "model.layers.71.self_attn.o_proj.scales": "model-00032-of-00057.safetensors", + "model.layers.71.self_attn.q_proj.g_idx": "model-00032-of-00057.safetensors", + "model.layers.71.self_attn.q_proj.qweight": "model-00032-of-00057.safetensors", + "model.layers.71.self_attn.q_proj.qzeros": "model-00032-of-00057.safetensors", + "model.layers.71.self_attn.q_proj.scales": "model-00032-of-00057.safetensors", + "model.layers.71.self_attn.v_proj.g_idx": "model-00032-of-00057.safetensors", + "model.layers.71.self_attn.v_proj.qweight": "model-00032-of-00057.safetensors", + "model.layers.71.self_attn.v_proj.qzeros": "model-00032-of-00057.safetensors", + "model.layers.71.self_attn.v_proj.scales": "model-00032-of-00057.safetensors", + "model.layers.72.input_layernorm.weight": "model-00033-of-00057.safetensors", + "model.layers.72.mlp.down_proj.g_idx": "model-00033-of-00057.safetensors", + "model.layers.72.mlp.down_proj.qweight": "model-00033-of-00057.safetensors", + "model.layers.72.mlp.down_proj.qzeros": "model-00033-of-00057.safetensors", + "model.layers.72.mlp.down_proj.scales": "model-00033-of-00057.safetensors", + "model.layers.72.mlp.gate_proj.g_idx": "model-00033-of-00057.safetensors", + "model.layers.72.mlp.gate_proj.qweight": "model-00033-of-00057.safetensors", + "model.layers.72.mlp.gate_proj.qzeros": "model-00033-of-00057.safetensors", + "model.layers.72.mlp.gate_proj.scales": "model-00033-of-00057.safetensors", + "model.layers.72.mlp.up_proj.g_idx": "model-00033-of-00057.safetensors", + "model.layers.72.mlp.up_proj.qweight": "model-00033-of-00057.safetensors", + "model.layers.72.mlp.up_proj.qzeros": "model-00033-of-00057.safetensors", + "model.layers.72.mlp.up_proj.scales": "model-00033-of-00057.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00033-of-00057.safetensors", + "model.layers.72.self_attn.k_proj.g_idx": "model-00032-of-00057.safetensors", + "model.layers.72.self_attn.k_proj.qweight": "model-00032-of-00057.safetensors", + "model.layers.72.self_attn.k_proj.qzeros": "model-00032-of-00057.safetensors", + "model.layers.72.self_attn.k_proj.scales": "model-00032-of-00057.safetensors", + "model.layers.72.self_attn.o_proj.g_idx": "model-00033-of-00057.safetensors", + "model.layers.72.self_attn.o_proj.qweight": "model-00033-of-00057.safetensors", + "model.layers.72.self_attn.o_proj.qzeros": "model-00033-of-00057.safetensors", + "model.layers.72.self_attn.o_proj.scales": "model-00033-of-00057.safetensors", + "model.layers.72.self_attn.q_proj.g_idx": "model-00032-of-00057.safetensors", + "model.layers.72.self_attn.q_proj.qweight": "model-00032-of-00057.safetensors", + "model.layers.72.self_attn.q_proj.qzeros": "model-00032-of-00057.safetensors", + "model.layers.72.self_attn.q_proj.scales": "model-00032-of-00057.safetensors", + "model.layers.72.self_attn.v_proj.g_idx": "model-00032-of-00057.safetensors", + "model.layers.72.self_attn.v_proj.qweight": "model-00032-of-00057.safetensors", + "model.layers.72.self_attn.v_proj.qzeros": "model-00032-of-00057.safetensors", + "model.layers.72.self_attn.v_proj.scales": "model-00032-of-00057.safetensors", + "model.layers.73.input_layernorm.weight": "model-00033-of-00057.safetensors", + "model.layers.73.mlp.down_proj.g_idx": "model-00033-of-00057.safetensors", + "model.layers.73.mlp.down_proj.qweight": "model-00033-of-00057.safetensors", + "model.layers.73.mlp.down_proj.qzeros": "model-00033-of-00057.safetensors", + "model.layers.73.mlp.down_proj.scales": "model-00033-of-00057.safetensors", + "model.layers.73.mlp.gate_proj.g_idx": "model-00033-of-00057.safetensors", + "model.layers.73.mlp.gate_proj.qweight": "model-00033-of-00057.safetensors", + "model.layers.73.mlp.gate_proj.qzeros": "model-00033-of-00057.safetensors", + "model.layers.73.mlp.gate_proj.scales": "model-00033-of-00057.safetensors", + "model.layers.73.mlp.up_proj.g_idx": "model-00033-of-00057.safetensors", + "model.layers.73.mlp.up_proj.qweight": "model-00033-of-00057.safetensors", + "model.layers.73.mlp.up_proj.qzeros": "model-00033-of-00057.safetensors", + "model.layers.73.mlp.up_proj.scales": "model-00033-of-00057.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00033-of-00057.safetensors", + "model.layers.73.self_attn.k_proj.g_idx": "model-00033-of-00057.safetensors", + "model.layers.73.self_attn.k_proj.qweight": "model-00033-of-00057.safetensors", + "model.layers.73.self_attn.k_proj.qzeros": "model-00033-of-00057.safetensors", + "model.layers.73.self_attn.k_proj.scales": "model-00033-of-00057.safetensors", + "model.layers.73.self_attn.o_proj.g_idx": "model-00033-of-00057.safetensors", + "model.layers.73.self_attn.o_proj.qweight": "model-00033-of-00057.safetensors", + "model.layers.73.self_attn.o_proj.qzeros": "model-00033-of-00057.safetensors", + "model.layers.73.self_attn.o_proj.scales": "model-00033-of-00057.safetensors", + "model.layers.73.self_attn.q_proj.g_idx": "model-00033-of-00057.safetensors", + "model.layers.73.self_attn.q_proj.qweight": "model-00033-of-00057.safetensors", + "model.layers.73.self_attn.q_proj.qzeros": "model-00033-of-00057.safetensors", + "model.layers.73.self_attn.q_proj.scales": "model-00033-of-00057.safetensors", + "model.layers.73.self_attn.v_proj.g_idx": "model-00033-of-00057.safetensors", + "model.layers.73.self_attn.v_proj.qweight": "model-00033-of-00057.safetensors", + "model.layers.73.self_attn.v_proj.qzeros": "model-00033-of-00057.safetensors", + "model.layers.73.self_attn.v_proj.scales": "model-00033-of-00057.safetensors", + "model.layers.74.input_layernorm.weight": "model-00034-of-00057.safetensors", + "model.layers.74.mlp.down_proj.g_idx": "model-00034-of-00057.safetensors", + "model.layers.74.mlp.down_proj.qweight": "model-00034-of-00057.safetensors", + "model.layers.74.mlp.down_proj.qzeros": "model-00034-of-00057.safetensors", + "model.layers.74.mlp.down_proj.scales": "model-00034-of-00057.safetensors", + "model.layers.74.mlp.gate_proj.g_idx": "model-00033-of-00057.safetensors", + "model.layers.74.mlp.gate_proj.qweight": "model-00033-of-00057.safetensors", + "model.layers.74.mlp.gate_proj.qzeros": "model-00033-of-00057.safetensors", + "model.layers.74.mlp.gate_proj.scales": "model-00033-of-00057.safetensors", + "model.layers.74.mlp.up_proj.g_idx": "model-00034-of-00057.safetensors", + "model.layers.74.mlp.up_proj.qweight": "model-00034-of-00057.safetensors", + "model.layers.74.mlp.up_proj.qzeros": "model-00034-of-00057.safetensors", + "model.layers.74.mlp.up_proj.scales": "model-00034-of-00057.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00034-of-00057.safetensors", + "model.layers.74.self_attn.k_proj.g_idx": "model-00033-of-00057.safetensors", + "model.layers.74.self_attn.k_proj.qweight": "model-00033-of-00057.safetensors", + "model.layers.74.self_attn.k_proj.qzeros": "model-00033-of-00057.safetensors", + "model.layers.74.self_attn.k_proj.scales": "model-00033-of-00057.safetensors", + "model.layers.74.self_attn.o_proj.g_idx": "model-00033-of-00057.safetensors", + "model.layers.74.self_attn.o_proj.qweight": "model-00033-of-00057.safetensors", + "model.layers.74.self_attn.o_proj.qzeros": "model-00033-of-00057.safetensors", + "model.layers.74.self_attn.o_proj.scales": "model-00033-of-00057.safetensors", + "model.layers.74.self_attn.q_proj.g_idx": "model-00033-of-00057.safetensors", + "model.layers.74.self_attn.q_proj.qweight": "model-00033-of-00057.safetensors", + "model.layers.74.self_attn.q_proj.qzeros": "model-00033-of-00057.safetensors", + "model.layers.74.self_attn.q_proj.scales": "model-00033-of-00057.safetensors", + "model.layers.74.self_attn.v_proj.g_idx": "model-00033-of-00057.safetensors", + "model.layers.74.self_attn.v_proj.qweight": "model-00033-of-00057.safetensors", + "model.layers.74.self_attn.v_proj.qzeros": "model-00033-of-00057.safetensors", + "model.layers.74.self_attn.v_proj.scales": "model-00033-of-00057.safetensors", + "model.layers.75.input_layernorm.weight": "model-00034-of-00057.safetensors", + "model.layers.75.mlp.down_proj.g_idx": "model-00034-of-00057.safetensors", + "model.layers.75.mlp.down_proj.qweight": "model-00034-of-00057.safetensors", + "model.layers.75.mlp.down_proj.qzeros": "model-00034-of-00057.safetensors", + "model.layers.75.mlp.down_proj.scales": "model-00034-of-00057.safetensors", + "model.layers.75.mlp.gate_proj.g_idx": "model-00034-of-00057.safetensors", + "model.layers.75.mlp.gate_proj.qweight": "model-00034-of-00057.safetensors", + "model.layers.75.mlp.gate_proj.qzeros": "model-00034-of-00057.safetensors", + "model.layers.75.mlp.gate_proj.scales": "model-00034-of-00057.safetensors", + "model.layers.75.mlp.up_proj.g_idx": "model-00034-of-00057.safetensors", + "model.layers.75.mlp.up_proj.qweight": "model-00034-of-00057.safetensors", + "model.layers.75.mlp.up_proj.qzeros": "model-00034-of-00057.safetensors", + "model.layers.75.mlp.up_proj.scales": "model-00034-of-00057.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00034-of-00057.safetensors", + "model.layers.75.self_attn.k_proj.g_idx": "model-00034-of-00057.safetensors", + "model.layers.75.self_attn.k_proj.qweight": "model-00034-of-00057.safetensors", + "model.layers.75.self_attn.k_proj.qzeros": "model-00034-of-00057.safetensors", + "model.layers.75.self_attn.k_proj.scales": "model-00034-of-00057.safetensors", + "model.layers.75.self_attn.o_proj.g_idx": "model-00034-of-00057.safetensors", + "model.layers.75.self_attn.o_proj.qweight": "model-00034-of-00057.safetensors", + "model.layers.75.self_attn.o_proj.qzeros": "model-00034-of-00057.safetensors", + "model.layers.75.self_attn.o_proj.scales": "model-00034-of-00057.safetensors", + "model.layers.75.self_attn.q_proj.g_idx": "model-00034-of-00057.safetensors", + "model.layers.75.self_attn.q_proj.qweight": "model-00034-of-00057.safetensors", + "model.layers.75.self_attn.q_proj.qzeros": "model-00034-of-00057.safetensors", + "model.layers.75.self_attn.q_proj.scales": "model-00034-of-00057.safetensors", + "model.layers.75.self_attn.v_proj.g_idx": "model-00034-of-00057.safetensors", + "model.layers.75.self_attn.v_proj.qweight": "model-00034-of-00057.safetensors", + "model.layers.75.self_attn.v_proj.qzeros": "model-00034-of-00057.safetensors", + "model.layers.75.self_attn.v_proj.scales": "model-00034-of-00057.safetensors", + "model.layers.76.input_layernorm.weight": "model-00035-of-00057.safetensors", + "model.layers.76.mlp.down_proj.g_idx": "model-00035-of-00057.safetensors", + "model.layers.76.mlp.down_proj.qweight": "model-00035-of-00057.safetensors", + "model.layers.76.mlp.down_proj.qzeros": "model-00035-of-00057.safetensors", + "model.layers.76.mlp.down_proj.scales": "model-00035-of-00057.safetensors", + "model.layers.76.mlp.gate_proj.g_idx": "model-00034-of-00057.safetensors", + "model.layers.76.mlp.gate_proj.qweight": "model-00034-of-00057.safetensors", + "model.layers.76.mlp.gate_proj.qzeros": "model-00034-of-00057.safetensors", + "model.layers.76.mlp.gate_proj.scales": "model-00034-of-00057.safetensors", + "model.layers.76.mlp.up_proj.g_idx": "model-00034-of-00057.safetensors", + "model.layers.76.mlp.up_proj.qweight": "model-00034-of-00057.safetensors", + "model.layers.76.mlp.up_proj.qzeros": "model-00034-of-00057.safetensors", + "model.layers.76.mlp.up_proj.scales": "model-00034-of-00057.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00035-of-00057.safetensors", + "model.layers.76.self_attn.k_proj.g_idx": "model-00034-of-00057.safetensors", + "model.layers.76.self_attn.k_proj.qweight": "model-00034-of-00057.safetensors", + "model.layers.76.self_attn.k_proj.qzeros": "model-00034-of-00057.safetensors", + "model.layers.76.self_attn.k_proj.scales": "model-00034-of-00057.safetensors", + "model.layers.76.self_attn.o_proj.g_idx": "model-00034-of-00057.safetensors", + "model.layers.76.self_attn.o_proj.qweight": "model-00034-of-00057.safetensors", + "model.layers.76.self_attn.o_proj.qzeros": "model-00034-of-00057.safetensors", + "model.layers.76.self_attn.o_proj.scales": "model-00034-of-00057.safetensors", + "model.layers.76.self_attn.q_proj.g_idx": "model-00034-of-00057.safetensors", + "model.layers.76.self_attn.q_proj.qweight": "model-00034-of-00057.safetensors", + "model.layers.76.self_attn.q_proj.qzeros": "model-00034-of-00057.safetensors", + "model.layers.76.self_attn.q_proj.scales": "model-00034-of-00057.safetensors", + "model.layers.76.self_attn.v_proj.g_idx": "model-00034-of-00057.safetensors", + "model.layers.76.self_attn.v_proj.qweight": "model-00034-of-00057.safetensors", + "model.layers.76.self_attn.v_proj.qzeros": "model-00034-of-00057.safetensors", + "model.layers.76.self_attn.v_proj.scales": "model-00034-of-00057.safetensors", + "model.layers.77.input_layernorm.weight": "model-00035-of-00057.safetensors", + "model.layers.77.mlp.down_proj.g_idx": "model-00035-of-00057.safetensors", + "model.layers.77.mlp.down_proj.qweight": "model-00035-of-00057.safetensors", + "model.layers.77.mlp.down_proj.qzeros": "model-00035-of-00057.safetensors", + "model.layers.77.mlp.down_proj.scales": "model-00035-of-00057.safetensors", + "model.layers.77.mlp.gate_proj.g_idx": "model-00035-of-00057.safetensors", + "model.layers.77.mlp.gate_proj.qweight": "model-00035-of-00057.safetensors", + "model.layers.77.mlp.gate_proj.qzeros": "model-00035-of-00057.safetensors", + "model.layers.77.mlp.gate_proj.scales": "model-00035-of-00057.safetensors", + "model.layers.77.mlp.up_proj.g_idx": "model-00035-of-00057.safetensors", + "model.layers.77.mlp.up_proj.qweight": "model-00035-of-00057.safetensors", + "model.layers.77.mlp.up_proj.qzeros": "model-00035-of-00057.safetensors", + "model.layers.77.mlp.up_proj.scales": "model-00035-of-00057.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00035-of-00057.safetensors", + "model.layers.77.self_attn.k_proj.g_idx": "model-00035-of-00057.safetensors", + "model.layers.77.self_attn.k_proj.qweight": "model-00035-of-00057.safetensors", + "model.layers.77.self_attn.k_proj.qzeros": "model-00035-of-00057.safetensors", + "model.layers.77.self_attn.k_proj.scales": "model-00035-of-00057.safetensors", + "model.layers.77.self_attn.o_proj.g_idx": "model-00035-of-00057.safetensors", + "model.layers.77.self_attn.o_proj.qweight": "model-00035-of-00057.safetensors", + "model.layers.77.self_attn.o_proj.qzeros": "model-00035-of-00057.safetensors", + "model.layers.77.self_attn.o_proj.scales": "model-00035-of-00057.safetensors", + "model.layers.77.self_attn.q_proj.g_idx": "model-00035-of-00057.safetensors", + "model.layers.77.self_attn.q_proj.qweight": "model-00035-of-00057.safetensors", + "model.layers.77.self_attn.q_proj.qzeros": "model-00035-of-00057.safetensors", + "model.layers.77.self_attn.q_proj.scales": "model-00035-of-00057.safetensors", + "model.layers.77.self_attn.v_proj.g_idx": "model-00035-of-00057.safetensors", + "model.layers.77.self_attn.v_proj.qweight": "model-00035-of-00057.safetensors", + "model.layers.77.self_attn.v_proj.qzeros": "model-00035-of-00057.safetensors", + "model.layers.77.self_attn.v_proj.scales": "model-00035-of-00057.safetensors", + "model.layers.78.input_layernorm.weight": "model-00035-of-00057.safetensors", + "model.layers.78.mlp.down_proj.g_idx": "model-00035-of-00057.safetensors", + "model.layers.78.mlp.down_proj.qweight": "model-00035-of-00057.safetensors", + "model.layers.78.mlp.down_proj.qzeros": "model-00035-of-00057.safetensors", + "model.layers.78.mlp.down_proj.scales": "model-00035-of-00057.safetensors", + "model.layers.78.mlp.gate_proj.g_idx": "model-00035-of-00057.safetensors", + "model.layers.78.mlp.gate_proj.qweight": "model-00035-of-00057.safetensors", + "model.layers.78.mlp.gate_proj.qzeros": "model-00035-of-00057.safetensors", + "model.layers.78.mlp.gate_proj.scales": "model-00035-of-00057.safetensors", + "model.layers.78.mlp.up_proj.g_idx": "model-00035-of-00057.safetensors", + "model.layers.78.mlp.up_proj.qweight": "model-00035-of-00057.safetensors", + "model.layers.78.mlp.up_proj.qzeros": "model-00035-of-00057.safetensors", + "model.layers.78.mlp.up_proj.scales": "model-00035-of-00057.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00035-of-00057.safetensors", + "model.layers.78.self_attn.k_proj.g_idx": "model-00035-of-00057.safetensors", + "model.layers.78.self_attn.k_proj.qweight": "model-00035-of-00057.safetensors", + "model.layers.78.self_attn.k_proj.qzeros": "model-00035-of-00057.safetensors", + "model.layers.78.self_attn.k_proj.scales": "model-00035-of-00057.safetensors", + "model.layers.78.self_attn.o_proj.g_idx": "model-00035-of-00057.safetensors", + "model.layers.78.self_attn.o_proj.qweight": "model-00035-of-00057.safetensors", + "model.layers.78.self_attn.o_proj.qzeros": "model-00035-of-00057.safetensors", + "model.layers.78.self_attn.o_proj.scales": "model-00035-of-00057.safetensors", + "model.layers.78.self_attn.q_proj.g_idx": "model-00035-of-00057.safetensors", + "model.layers.78.self_attn.q_proj.qweight": "model-00035-of-00057.safetensors", + "model.layers.78.self_attn.q_proj.qzeros": "model-00035-of-00057.safetensors", + "model.layers.78.self_attn.q_proj.scales": "model-00035-of-00057.safetensors", + "model.layers.78.self_attn.v_proj.g_idx": "model-00035-of-00057.safetensors", + "model.layers.78.self_attn.v_proj.qweight": "model-00035-of-00057.safetensors", + "model.layers.78.self_attn.v_proj.qzeros": "model-00035-of-00057.safetensors", + "model.layers.78.self_attn.v_proj.scales": "model-00035-of-00057.safetensors", + "model.layers.79.input_layernorm.weight": "model-00036-of-00057.safetensors", + "model.layers.79.mlp.down_proj.g_idx": "model-00036-of-00057.safetensors", + "model.layers.79.mlp.down_proj.qweight": "model-00036-of-00057.safetensors", + "model.layers.79.mlp.down_proj.qzeros": "model-00036-of-00057.safetensors", + "model.layers.79.mlp.down_proj.scales": "model-00036-of-00057.safetensors", + "model.layers.79.mlp.gate_proj.g_idx": "model-00036-of-00057.safetensors", + "model.layers.79.mlp.gate_proj.qweight": "model-00036-of-00057.safetensors", + "model.layers.79.mlp.gate_proj.qzeros": "model-00036-of-00057.safetensors", + "model.layers.79.mlp.gate_proj.scales": "model-00036-of-00057.safetensors", + "model.layers.79.mlp.up_proj.g_idx": "model-00036-of-00057.safetensors", + "model.layers.79.mlp.up_proj.qweight": "model-00036-of-00057.safetensors", + "model.layers.79.mlp.up_proj.qzeros": "model-00036-of-00057.safetensors", + "model.layers.79.mlp.up_proj.scales": "model-00036-of-00057.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00036-of-00057.safetensors", + "model.layers.79.self_attn.k_proj.g_idx": "model-00035-of-00057.safetensors", + "model.layers.79.self_attn.k_proj.qweight": "model-00035-of-00057.safetensors", + "model.layers.79.self_attn.k_proj.qzeros": "model-00035-of-00057.safetensors", + "model.layers.79.self_attn.k_proj.scales": "model-00035-of-00057.safetensors", + "model.layers.79.self_attn.o_proj.g_idx": "model-00036-of-00057.safetensors", + "model.layers.79.self_attn.o_proj.qweight": "model-00036-of-00057.safetensors", + "model.layers.79.self_attn.o_proj.qzeros": "model-00036-of-00057.safetensors", + "model.layers.79.self_attn.o_proj.scales": "model-00036-of-00057.safetensors", + "model.layers.79.self_attn.q_proj.g_idx": "model-00035-of-00057.safetensors", + "model.layers.79.self_attn.q_proj.qweight": "model-00035-of-00057.safetensors", + "model.layers.79.self_attn.q_proj.qzeros": "model-00035-of-00057.safetensors", + "model.layers.79.self_attn.q_proj.scales": "model-00035-of-00057.safetensors", + "model.layers.79.self_attn.v_proj.g_idx": "model-00035-of-00057.safetensors", + "model.layers.79.self_attn.v_proj.qweight": "model-00035-of-00057.safetensors", + "model.layers.79.self_attn.v_proj.qzeros": "model-00035-of-00057.safetensors", + "model.layers.79.self_attn.v_proj.scales": "model-00035-of-00057.safetensors", + "model.layers.8.input_layernorm.weight": "model-00005-of-00057.safetensors", + "model.layers.8.mlp.down_proj.g_idx": "model-00005-of-00057.safetensors", + "model.layers.8.mlp.down_proj.qweight": "model-00005-of-00057.safetensors", + "model.layers.8.mlp.down_proj.qzeros": "model-00005-of-00057.safetensors", + "model.layers.8.mlp.down_proj.scales": "model-00005-of-00057.safetensors", + "model.layers.8.mlp.gate_proj.g_idx": "model-00005-of-00057.safetensors", + "model.layers.8.mlp.gate_proj.qweight": "model-00005-of-00057.safetensors", + "model.layers.8.mlp.gate_proj.qzeros": "model-00005-of-00057.safetensors", + "model.layers.8.mlp.gate_proj.scales": "model-00005-of-00057.safetensors", + "model.layers.8.mlp.up_proj.g_idx": "model-00005-of-00057.safetensors", + "model.layers.8.mlp.up_proj.qweight": "model-00005-of-00057.safetensors", + "model.layers.8.mlp.up_proj.qzeros": "model-00005-of-00057.safetensors", + "model.layers.8.mlp.up_proj.scales": "model-00005-of-00057.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00005-of-00057.safetensors", + "model.layers.8.self_attn.k_proj.g_idx": "model-00005-of-00057.safetensors", + "model.layers.8.self_attn.k_proj.qweight": "model-00005-of-00057.safetensors", + "model.layers.8.self_attn.k_proj.qzeros": "model-00005-of-00057.safetensors", + "model.layers.8.self_attn.k_proj.scales": "model-00005-of-00057.safetensors", + "model.layers.8.self_attn.o_proj.g_idx": "model-00005-of-00057.safetensors", + "model.layers.8.self_attn.o_proj.qweight": "model-00005-of-00057.safetensors", + "model.layers.8.self_attn.o_proj.qzeros": "model-00005-of-00057.safetensors", + "model.layers.8.self_attn.o_proj.scales": "model-00005-of-00057.safetensors", + "model.layers.8.self_attn.q_proj.g_idx": "model-00005-of-00057.safetensors", + "model.layers.8.self_attn.q_proj.qweight": "model-00005-of-00057.safetensors", + "model.layers.8.self_attn.q_proj.qzeros": "model-00005-of-00057.safetensors", + "model.layers.8.self_attn.q_proj.scales": "model-00005-of-00057.safetensors", + "model.layers.8.self_attn.v_proj.g_idx": "model-00005-of-00057.safetensors", + "model.layers.8.self_attn.v_proj.qweight": "model-00005-of-00057.safetensors", + "model.layers.8.self_attn.v_proj.qzeros": "model-00005-of-00057.safetensors", + "model.layers.8.self_attn.v_proj.scales": "model-00005-of-00057.safetensors", + "model.layers.80.input_layernorm.weight": "model-00036-of-00057.safetensors", + "model.layers.80.mlp.down_proj.g_idx": "model-00036-of-00057.safetensors", + "model.layers.80.mlp.down_proj.qweight": "model-00036-of-00057.safetensors", + "model.layers.80.mlp.down_proj.qzeros": "model-00036-of-00057.safetensors", + "model.layers.80.mlp.down_proj.scales": "model-00036-of-00057.safetensors", + "model.layers.80.mlp.gate_proj.g_idx": "model-00036-of-00057.safetensors", + "model.layers.80.mlp.gate_proj.qweight": "model-00036-of-00057.safetensors", + "model.layers.80.mlp.gate_proj.qzeros": "model-00036-of-00057.safetensors", + "model.layers.80.mlp.gate_proj.scales": "model-00036-of-00057.safetensors", + "model.layers.80.mlp.up_proj.g_idx": "model-00036-of-00057.safetensors", + "model.layers.80.mlp.up_proj.qweight": "model-00036-of-00057.safetensors", + "model.layers.80.mlp.up_proj.qzeros": "model-00036-of-00057.safetensors", + "model.layers.80.mlp.up_proj.scales": "model-00036-of-00057.safetensors", + "model.layers.80.post_attention_layernorm.weight": "model-00036-of-00057.safetensors", + "model.layers.80.self_attn.k_proj.g_idx": "model-00036-of-00057.safetensors", + "model.layers.80.self_attn.k_proj.qweight": "model-00036-of-00057.safetensors", + "model.layers.80.self_attn.k_proj.qzeros": "model-00036-of-00057.safetensors", + "model.layers.80.self_attn.k_proj.scales": "model-00036-of-00057.safetensors", + "model.layers.80.self_attn.o_proj.g_idx": "model-00036-of-00057.safetensors", + "model.layers.80.self_attn.o_proj.qweight": "model-00036-of-00057.safetensors", + "model.layers.80.self_attn.o_proj.qzeros": "model-00036-of-00057.safetensors", + "model.layers.80.self_attn.o_proj.scales": "model-00036-of-00057.safetensors", + "model.layers.80.self_attn.q_proj.g_idx": "model-00036-of-00057.safetensors", + "model.layers.80.self_attn.q_proj.qweight": "model-00036-of-00057.safetensors", + "model.layers.80.self_attn.q_proj.qzeros": "model-00036-of-00057.safetensors", + "model.layers.80.self_attn.q_proj.scales": "model-00036-of-00057.safetensors", + "model.layers.80.self_attn.v_proj.g_idx": "model-00036-of-00057.safetensors", + "model.layers.80.self_attn.v_proj.qweight": "model-00036-of-00057.safetensors", + "model.layers.80.self_attn.v_proj.qzeros": "model-00036-of-00057.safetensors", + "model.layers.80.self_attn.v_proj.scales": "model-00036-of-00057.safetensors", + "model.layers.81.input_layernorm.weight": "model-00037-of-00057.safetensors", + "model.layers.81.mlp.down_proj.g_idx": "model-00037-of-00057.safetensors", + "model.layers.81.mlp.down_proj.qweight": "model-00037-of-00057.safetensors", + "model.layers.81.mlp.down_proj.qzeros": "model-00037-of-00057.safetensors", + "model.layers.81.mlp.down_proj.scales": "model-00037-of-00057.safetensors", + "model.layers.81.mlp.gate_proj.g_idx": "model-00036-of-00057.safetensors", + "model.layers.81.mlp.gate_proj.qweight": "model-00036-of-00057.safetensors", + "model.layers.81.mlp.gate_proj.qzeros": "model-00036-of-00057.safetensors", + "model.layers.81.mlp.gate_proj.scales": "model-00036-of-00057.safetensors", + "model.layers.81.mlp.up_proj.g_idx": "model-00037-of-00057.safetensors", + "model.layers.81.mlp.up_proj.qweight": "model-00037-of-00057.safetensors", + "model.layers.81.mlp.up_proj.qzeros": "model-00037-of-00057.safetensors", + "model.layers.81.mlp.up_proj.scales": "model-00037-of-00057.safetensors", + "model.layers.81.post_attention_layernorm.weight": "model-00037-of-00057.safetensors", + "model.layers.81.self_attn.k_proj.g_idx": "model-00036-of-00057.safetensors", + "model.layers.81.self_attn.k_proj.qweight": "model-00036-of-00057.safetensors", + "model.layers.81.self_attn.k_proj.qzeros": "model-00036-of-00057.safetensors", + "model.layers.81.self_attn.k_proj.scales": "model-00036-of-00057.safetensors", + "model.layers.81.self_attn.o_proj.g_idx": "model-00036-of-00057.safetensors", + "model.layers.81.self_attn.o_proj.qweight": "model-00036-of-00057.safetensors", + "model.layers.81.self_attn.o_proj.qzeros": "model-00036-of-00057.safetensors", + "model.layers.81.self_attn.o_proj.scales": "model-00036-of-00057.safetensors", + "model.layers.81.self_attn.q_proj.g_idx": "model-00036-of-00057.safetensors", + "model.layers.81.self_attn.q_proj.qweight": "model-00036-of-00057.safetensors", + "model.layers.81.self_attn.q_proj.qzeros": "model-00036-of-00057.safetensors", + "model.layers.81.self_attn.q_proj.scales": "model-00036-of-00057.safetensors", + "model.layers.81.self_attn.v_proj.g_idx": "model-00036-of-00057.safetensors", + "model.layers.81.self_attn.v_proj.qweight": "model-00036-of-00057.safetensors", + "model.layers.81.self_attn.v_proj.qzeros": "model-00036-of-00057.safetensors", + "model.layers.81.self_attn.v_proj.scales": "model-00036-of-00057.safetensors", + "model.layers.82.input_layernorm.weight": "model-00037-of-00057.safetensors", + "model.layers.82.mlp.down_proj.g_idx": "model-00037-of-00057.safetensors", + "model.layers.82.mlp.down_proj.qweight": "model-00037-of-00057.safetensors", + "model.layers.82.mlp.down_proj.qzeros": "model-00037-of-00057.safetensors", + "model.layers.82.mlp.down_proj.scales": "model-00037-of-00057.safetensors", + "model.layers.82.mlp.gate_proj.g_idx": "model-00037-of-00057.safetensors", + "model.layers.82.mlp.gate_proj.qweight": "model-00037-of-00057.safetensors", + "model.layers.82.mlp.gate_proj.qzeros": "model-00037-of-00057.safetensors", + "model.layers.82.mlp.gate_proj.scales": "model-00037-of-00057.safetensors", + "model.layers.82.mlp.up_proj.g_idx": "model-00037-of-00057.safetensors", + "model.layers.82.mlp.up_proj.qweight": "model-00037-of-00057.safetensors", + "model.layers.82.mlp.up_proj.qzeros": "model-00037-of-00057.safetensors", + "model.layers.82.mlp.up_proj.scales": "model-00037-of-00057.safetensors", + "model.layers.82.post_attention_layernorm.weight": "model-00037-of-00057.safetensors", + "model.layers.82.self_attn.k_proj.g_idx": "model-00037-of-00057.safetensors", + "model.layers.82.self_attn.k_proj.qweight": "model-00037-of-00057.safetensors", + "model.layers.82.self_attn.k_proj.qzeros": "model-00037-of-00057.safetensors", + "model.layers.82.self_attn.k_proj.scales": "model-00037-of-00057.safetensors", + "model.layers.82.self_attn.o_proj.g_idx": "model-00037-of-00057.safetensors", + "model.layers.82.self_attn.o_proj.qweight": "model-00037-of-00057.safetensors", + "model.layers.82.self_attn.o_proj.qzeros": "model-00037-of-00057.safetensors", + "model.layers.82.self_attn.o_proj.scales": "model-00037-of-00057.safetensors", + "model.layers.82.self_attn.q_proj.g_idx": "model-00037-of-00057.safetensors", + "model.layers.82.self_attn.q_proj.qweight": "model-00037-of-00057.safetensors", + "model.layers.82.self_attn.q_proj.qzeros": "model-00037-of-00057.safetensors", + "model.layers.82.self_attn.q_proj.scales": "model-00037-of-00057.safetensors", + "model.layers.82.self_attn.v_proj.g_idx": "model-00037-of-00057.safetensors", + "model.layers.82.self_attn.v_proj.qweight": "model-00037-of-00057.safetensors", + "model.layers.82.self_attn.v_proj.qzeros": "model-00037-of-00057.safetensors", + "model.layers.82.self_attn.v_proj.scales": "model-00037-of-00057.safetensors", + "model.layers.83.input_layernorm.weight": "model-00038-of-00057.safetensors", + "model.layers.83.mlp.down_proj.g_idx": "model-00038-of-00057.safetensors", + "model.layers.83.mlp.down_proj.qweight": "model-00038-of-00057.safetensors", + "model.layers.83.mlp.down_proj.qzeros": "model-00038-of-00057.safetensors", + "model.layers.83.mlp.down_proj.scales": "model-00038-of-00057.safetensors", + "model.layers.83.mlp.gate_proj.g_idx": "model-00037-of-00057.safetensors", + "model.layers.83.mlp.gate_proj.qweight": "model-00037-of-00057.safetensors", + "model.layers.83.mlp.gate_proj.qzeros": "model-00037-of-00057.safetensors", + "model.layers.83.mlp.gate_proj.scales": "model-00037-of-00057.safetensors", + "model.layers.83.mlp.up_proj.g_idx": "model-00037-of-00057.safetensors", + "model.layers.83.mlp.up_proj.qweight": "model-00037-of-00057.safetensors", + "model.layers.83.mlp.up_proj.qzeros": "model-00037-of-00057.safetensors", + "model.layers.83.mlp.up_proj.scales": "model-00037-of-00057.safetensors", + "model.layers.83.post_attention_layernorm.weight": "model-00038-of-00057.safetensors", + "model.layers.83.self_attn.k_proj.g_idx": "model-00037-of-00057.safetensors", + "model.layers.83.self_attn.k_proj.qweight": "model-00037-of-00057.safetensors", + "model.layers.83.self_attn.k_proj.qzeros": "model-00037-of-00057.safetensors", + "model.layers.83.self_attn.k_proj.scales": "model-00037-of-00057.safetensors", + "model.layers.83.self_attn.o_proj.g_idx": "model-00037-of-00057.safetensors", + "model.layers.83.self_attn.o_proj.qweight": "model-00037-of-00057.safetensors", + "model.layers.83.self_attn.o_proj.qzeros": "model-00037-of-00057.safetensors", + "model.layers.83.self_attn.o_proj.scales": "model-00037-of-00057.safetensors", + "model.layers.83.self_attn.q_proj.g_idx": "model-00037-of-00057.safetensors", + "model.layers.83.self_attn.q_proj.qweight": "model-00037-of-00057.safetensors", + "model.layers.83.self_attn.q_proj.qzeros": "model-00037-of-00057.safetensors", + "model.layers.83.self_attn.q_proj.scales": "model-00037-of-00057.safetensors", + "model.layers.83.self_attn.v_proj.g_idx": "model-00037-of-00057.safetensors", + "model.layers.83.self_attn.v_proj.qweight": "model-00037-of-00057.safetensors", + "model.layers.83.self_attn.v_proj.qzeros": "model-00037-of-00057.safetensors", + "model.layers.83.self_attn.v_proj.scales": "model-00037-of-00057.safetensors", + "model.layers.84.input_layernorm.weight": "model-00038-of-00057.safetensors", + "model.layers.84.mlp.down_proj.g_idx": "model-00038-of-00057.safetensors", + "model.layers.84.mlp.down_proj.qweight": "model-00038-of-00057.safetensors", + "model.layers.84.mlp.down_proj.qzeros": "model-00038-of-00057.safetensors", + "model.layers.84.mlp.down_proj.scales": "model-00038-of-00057.safetensors", + "model.layers.84.mlp.gate_proj.g_idx": "model-00038-of-00057.safetensors", + "model.layers.84.mlp.gate_proj.qweight": "model-00038-of-00057.safetensors", + "model.layers.84.mlp.gate_proj.qzeros": "model-00038-of-00057.safetensors", + "model.layers.84.mlp.gate_proj.scales": "model-00038-of-00057.safetensors", + "model.layers.84.mlp.up_proj.g_idx": "model-00038-of-00057.safetensors", + "model.layers.84.mlp.up_proj.qweight": "model-00038-of-00057.safetensors", + "model.layers.84.mlp.up_proj.qzeros": "model-00038-of-00057.safetensors", + "model.layers.84.mlp.up_proj.scales": "model-00038-of-00057.safetensors", + "model.layers.84.post_attention_layernorm.weight": "model-00038-of-00057.safetensors", + "model.layers.84.self_attn.k_proj.g_idx": "model-00038-of-00057.safetensors", + "model.layers.84.self_attn.k_proj.qweight": "model-00038-of-00057.safetensors", + "model.layers.84.self_attn.k_proj.qzeros": "model-00038-of-00057.safetensors", + "model.layers.84.self_attn.k_proj.scales": "model-00038-of-00057.safetensors", + "model.layers.84.self_attn.o_proj.g_idx": "model-00038-of-00057.safetensors", + "model.layers.84.self_attn.o_proj.qweight": "model-00038-of-00057.safetensors", + "model.layers.84.self_attn.o_proj.qzeros": "model-00038-of-00057.safetensors", + "model.layers.84.self_attn.o_proj.scales": "model-00038-of-00057.safetensors", + "model.layers.84.self_attn.q_proj.g_idx": "model-00038-of-00057.safetensors", + "model.layers.84.self_attn.q_proj.qweight": "model-00038-of-00057.safetensors", + "model.layers.84.self_attn.q_proj.qzeros": "model-00038-of-00057.safetensors", + "model.layers.84.self_attn.q_proj.scales": "model-00038-of-00057.safetensors", + "model.layers.84.self_attn.v_proj.g_idx": "model-00038-of-00057.safetensors", + "model.layers.84.self_attn.v_proj.qweight": "model-00038-of-00057.safetensors", + "model.layers.84.self_attn.v_proj.qzeros": "model-00038-of-00057.safetensors", + "model.layers.84.self_attn.v_proj.scales": "model-00038-of-00057.safetensors", + "model.layers.85.input_layernorm.weight": "model-00038-of-00057.safetensors", + "model.layers.85.mlp.down_proj.g_idx": "model-00038-of-00057.safetensors", + "model.layers.85.mlp.down_proj.qweight": "model-00038-of-00057.safetensors", + "model.layers.85.mlp.down_proj.qzeros": "model-00038-of-00057.safetensors", + "model.layers.85.mlp.down_proj.scales": "model-00038-of-00057.safetensors", + "model.layers.85.mlp.gate_proj.g_idx": "model-00038-of-00057.safetensors", + "model.layers.85.mlp.gate_proj.qweight": "model-00038-of-00057.safetensors", + "model.layers.85.mlp.gate_proj.qzeros": "model-00038-of-00057.safetensors", + "model.layers.85.mlp.gate_proj.scales": "model-00038-of-00057.safetensors", + "model.layers.85.mlp.up_proj.g_idx": "model-00038-of-00057.safetensors", + "model.layers.85.mlp.up_proj.qweight": "model-00038-of-00057.safetensors", + "model.layers.85.mlp.up_proj.qzeros": "model-00038-of-00057.safetensors", + "model.layers.85.mlp.up_proj.scales": "model-00038-of-00057.safetensors", + "model.layers.85.post_attention_layernorm.weight": "model-00038-of-00057.safetensors", + "model.layers.85.self_attn.k_proj.g_idx": "model-00038-of-00057.safetensors", + "model.layers.85.self_attn.k_proj.qweight": "model-00038-of-00057.safetensors", + "model.layers.85.self_attn.k_proj.qzeros": "model-00038-of-00057.safetensors", + "model.layers.85.self_attn.k_proj.scales": "model-00038-of-00057.safetensors", + "model.layers.85.self_attn.o_proj.g_idx": "model-00038-of-00057.safetensors", + "model.layers.85.self_attn.o_proj.qweight": "model-00038-of-00057.safetensors", + "model.layers.85.self_attn.o_proj.qzeros": "model-00038-of-00057.safetensors", + "model.layers.85.self_attn.o_proj.scales": "model-00038-of-00057.safetensors", + "model.layers.85.self_attn.q_proj.g_idx": "model-00038-of-00057.safetensors", + "model.layers.85.self_attn.q_proj.qweight": "model-00038-of-00057.safetensors", + "model.layers.85.self_attn.q_proj.qzeros": "model-00038-of-00057.safetensors", + "model.layers.85.self_attn.q_proj.scales": "model-00038-of-00057.safetensors", + "model.layers.85.self_attn.v_proj.g_idx": "model-00038-of-00057.safetensors", + "model.layers.85.self_attn.v_proj.qweight": "model-00038-of-00057.safetensors", + "model.layers.85.self_attn.v_proj.qzeros": "model-00038-of-00057.safetensors", + "model.layers.85.self_attn.v_proj.scales": "model-00038-of-00057.safetensors", + "model.layers.86.input_layernorm.weight": "model-00039-of-00057.safetensors", + "model.layers.86.mlp.down_proj.g_idx": "model-00039-of-00057.safetensors", + "model.layers.86.mlp.down_proj.qweight": "model-00039-of-00057.safetensors", + "model.layers.86.mlp.down_proj.qzeros": "model-00039-of-00057.safetensors", + "model.layers.86.mlp.down_proj.scales": "model-00039-of-00057.safetensors", + "model.layers.86.mlp.gate_proj.g_idx": "model-00039-of-00057.safetensors", + "model.layers.86.mlp.gate_proj.qweight": "model-00039-of-00057.safetensors", + "model.layers.86.mlp.gate_proj.qzeros": "model-00039-of-00057.safetensors", + "model.layers.86.mlp.gate_proj.scales": "model-00039-of-00057.safetensors", + "model.layers.86.mlp.up_proj.g_idx": "model-00039-of-00057.safetensors", + "model.layers.86.mlp.up_proj.qweight": "model-00039-of-00057.safetensors", + "model.layers.86.mlp.up_proj.qzeros": "model-00039-of-00057.safetensors", + "model.layers.86.mlp.up_proj.scales": "model-00039-of-00057.safetensors", + "model.layers.86.post_attention_layernorm.weight": "model-00039-of-00057.safetensors", + "model.layers.86.self_attn.k_proj.g_idx": "model-00038-of-00057.safetensors", + "model.layers.86.self_attn.k_proj.qweight": "model-00038-of-00057.safetensors", + "model.layers.86.self_attn.k_proj.qzeros": "model-00038-of-00057.safetensors", + "model.layers.86.self_attn.k_proj.scales": "model-00038-of-00057.safetensors", + "model.layers.86.self_attn.o_proj.g_idx": "model-00039-of-00057.safetensors", + "model.layers.86.self_attn.o_proj.qweight": "model-00039-of-00057.safetensors", + "model.layers.86.self_attn.o_proj.qzeros": "model-00039-of-00057.safetensors", + "model.layers.86.self_attn.o_proj.scales": "model-00039-of-00057.safetensors", + "model.layers.86.self_attn.q_proj.g_idx": "model-00038-of-00057.safetensors", + "model.layers.86.self_attn.q_proj.qweight": "model-00038-of-00057.safetensors", + "model.layers.86.self_attn.q_proj.qzeros": "model-00038-of-00057.safetensors", + "model.layers.86.self_attn.q_proj.scales": "model-00038-of-00057.safetensors", + "model.layers.86.self_attn.v_proj.g_idx": "model-00038-of-00057.safetensors", + "model.layers.86.self_attn.v_proj.qweight": "model-00038-of-00057.safetensors", + "model.layers.86.self_attn.v_proj.qzeros": "model-00038-of-00057.safetensors", + "model.layers.86.self_attn.v_proj.scales": "model-00038-of-00057.safetensors", + "model.layers.87.input_layernorm.weight": "model-00039-of-00057.safetensors", + "model.layers.87.mlp.down_proj.g_idx": "model-00039-of-00057.safetensors", + "model.layers.87.mlp.down_proj.qweight": "model-00039-of-00057.safetensors", + "model.layers.87.mlp.down_proj.qzeros": "model-00039-of-00057.safetensors", + "model.layers.87.mlp.down_proj.scales": "model-00039-of-00057.safetensors", + "model.layers.87.mlp.gate_proj.g_idx": "model-00039-of-00057.safetensors", + "model.layers.87.mlp.gate_proj.qweight": "model-00039-of-00057.safetensors", + "model.layers.87.mlp.gate_proj.qzeros": "model-00039-of-00057.safetensors", + "model.layers.87.mlp.gate_proj.scales": "model-00039-of-00057.safetensors", + "model.layers.87.mlp.up_proj.g_idx": "model-00039-of-00057.safetensors", + "model.layers.87.mlp.up_proj.qweight": "model-00039-of-00057.safetensors", + "model.layers.87.mlp.up_proj.qzeros": "model-00039-of-00057.safetensors", + "model.layers.87.mlp.up_proj.scales": "model-00039-of-00057.safetensors", + "model.layers.87.post_attention_layernorm.weight": "model-00039-of-00057.safetensors", + "model.layers.87.self_attn.k_proj.g_idx": "model-00039-of-00057.safetensors", + "model.layers.87.self_attn.k_proj.qweight": "model-00039-of-00057.safetensors", + "model.layers.87.self_attn.k_proj.qzeros": "model-00039-of-00057.safetensors", + "model.layers.87.self_attn.k_proj.scales": "model-00039-of-00057.safetensors", + "model.layers.87.self_attn.o_proj.g_idx": "model-00039-of-00057.safetensors", + "model.layers.87.self_attn.o_proj.qweight": "model-00039-of-00057.safetensors", + "model.layers.87.self_attn.o_proj.qzeros": "model-00039-of-00057.safetensors", + "model.layers.87.self_attn.o_proj.scales": "model-00039-of-00057.safetensors", + "model.layers.87.self_attn.q_proj.g_idx": "model-00039-of-00057.safetensors", + "model.layers.87.self_attn.q_proj.qweight": "model-00039-of-00057.safetensors", + "model.layers.87.self_attn.q_proj.qzeros": "model-00039-of-00057.safetensors", + "model.layers.87.self_attn.q_proj.scales": "model-00039-of-00057.safetensors", + "model.layers.87.self_attn.v_proj.g_idx": "model-00039-of-00057.safetensors", + "model.layers.87.self_attn.v_proj.qweight": "model-00039-of-00057.safetensors", + "model.layers.87.self_attn.v_proj.qzeros": "model-00039-of-00057.safetensors", + "model.layers.87.self_attn.v_proj.scales": "model-00039-of-00057.safetensors", + "model.layers.88.input_layernorm.weight": "model-00040-of-00057.safetensors", + "model.layers.88.mlp.down_proj.g_idx": "model-00040-of-00057.safetensors", + "model.layers.88.mlp.down_proj.qweight": "model-00040-of-00057.safetensors", + "model.layers.88.mlp.down_proj.qzeros": "model-00040-of-00057.safetensors", + "model.layers.88.mlp.down_proj.scales": "model-00040-of-00057.safetensors", + "model.layers.88.mlp.gate_proj.g_idx": "model-00039-of-00057.safetensors", + "model.layers.88.mlp.gate_proj.qweight": "model-00039-of-00057.safetensors", + "model.layers.88.mlp.gate_proj.qzeros": "model-00039-of-00057.safetensors", + "model.layers.88.mlp.gate_proj.scales": "model-00039-of-00057.safetensors", + "model.layers.88.mlp.up_proj.g_idx": "model-00040-of-00057.safetensors", + "model.layers.88.mlp.up_proj.qweight": "model-00040-of-00057.safetensors", + "model.layers.88.mlp.up_proj.qzeros": "model-00040-of-00057.safetensors", + "model.layers.88.mlp.up_proj.scales": "model-00040-of-00057.safetensors", + "model.layers.88.post_attention_layernorm.weight": "model-00040-of-00057.safetensors", + "model.layers.88.self_attn.k_proj.g_idx": "model-00039-of-00057.safetensors", + "model.layers.88.self_attn.k_proj.qweight": "model-00039-of-00057.safetensors", + "model.layers.88.self_attn.k_proj.qzeros": "model-00039-of-00057.safetensors", + "model.layers.88.self_attn.k_proj.scales": "model-00039-of-00057.safetensors", + "model.layers.88.self_attn.o_proj.g_idx": "model-00039-of-00057.safetensors", + "model.layers.88.self_attn.o_proj.qweight": "model-00039-of-00057.safetensors", + "model.layers.88.self_attn.o_proj.qzeros": "model-00039-of-00057.safetensors", + "model.layers.88.self_attn.o_proj.scales": "model-00039-of-00057.safetensors", + "model.layers.88.self_attn.q_proj.g_idx": "model-00039-of-00057.safetensors", + "model.layers.88.self_attn.q_proj.qweight": "model-00039-of-00057.safetensors", + "model.layers.88.self_attn.q_proj.qzeros": "model-00039-of-00057.safetensors", + "model.layers.88.self_attn.q_proj.scales": "model-00039-of-00057.safetensors", + "model.layers.88.self_attn.v_proj.g_idx": "model-00039-of-00057.safetensors", + "model.layers.88.self_attn.v_proj.qweight": "model-00039-of-00057.safetensors", + "model.layers.88.self_attn.v_proj.qzeros": "model-00039-of-00057.safetensors", + "model.layers.88.self_attn.v_proj.scales": "model-00039-of-00057.safetensors", + "model.layers.89.input_layernorm.weight": "model-00040-of-00057.safetensors", + "model.layers.89.mlp.down_proj.g_idx": "model-00040-of-00057.safetensors", + "model.layers.89.mlp.down_proj.qweight": "model-00040-of-00057.safetensors", + "model.layers.89.mlp.down_proj.qzeros": "model-00040-of-00057.safetensors", + "model.layers.89.mlp.down_proj.scales": "model-00040-of-00057.safetensors", + "model.layers.89.mlp.gate_proj.g_idx": "model-00040-of-00057.safetensors", + "model.layers.89.mlp.gate_proj.qweight": "model-00040-of-00057.safetensors", + "model.layers.89.mlp.gate_proj.qzeros": "model-00040-of-00057.safetensors", + "model.layers.89.mlp.gate_proj.scales": "model-00040-of-00057.safetensors", + "model.layers.89.mlp.up_proj.g_idx": "model-00040-of-00057.safetensors", + "model.layers.89.mlp.up_proj.qweight": "model-00040-of-00057.safetensors", + "model.layers.89.mlp.up_proj.qzeros": "model-00040-of-00057.safetensors", + "model.layers.89.mlp.up_proj.scales": "model-00040-of-00057.safetensors", + "model.layers.89.post_attention_layernorm.weight": "model-00040-of-00057.safetensors", + "model.layers.89.self_attn.k_proj.g_idx": "model-00040-of-00057.safetensors", + "model.layers.89.self_attn.k_proj.qweight": "model-00040-of-00057.safetensors", + "model.layers.89.self_attn.k_proj.qzeros": "model-00040-of-00057.safetensors", + "model.layers.89.self_attn.k_proj.scales": "model-00040-of-00057.safetensors", + "model.layers.89.self_attn.o_proj.g_idx": "model-00040-of-00057.safetensors", + "model.layers.89.self_attn.o_proj.qweight": "model-00040-of-00057.safetensors", + "model.layers.89.self_attn.o_proj.qzeros": "model-00040-of-00057.safetensors", + "model.layers.89.self_attn.o_proj.scales": "model-00040-of-00057.safetensors", + "model.layers.89.self_attn.q_proj.g_idx": "model-00040-of-00057.safetensors", + "model.layers.89.self_attn.q_proj.qweight": "model-00040-of-00057.safetensors", + "model.layers.89.self_attn.q_proj.qzeros": "model-00040-of-00057.safetensors", + "model.layers.89.self_attn.q_proj.scales": "model-00040-of-00057.safetensors", + "model.layers.89.self_attn.v_proj.g_idx": "model-00040-of-00057.safetensors", + "model.layers.89.self_attn.v_proj.qweight": "model-00040-of-00057.safetensors", + "model.layers.89.self_attn.v_proj.qzeros": "model-00040-of-00057.safetensors", + "model.layers.89.self_attn.v_proj.scales": "model-00040-of-00057.safetensors", + "model.layers.9.input_layernorm.weight": "model-00006-of-00057.safetensors", + "model.layers.9.mlp.down_proj.g_idx": "model-00006-of-00057.safetensors", + "model.layers.9.mlp.down_proj.qweight": "model-00006-of-00057.safetensors", + "model.layers.9.mlp.down_proj.qzeros": "model-00006-of-00057.safetensors", + "model.layers.9.mlp.down_proj.scales": "model-00006-of-00057.safetensors", + "model.layers.9.mlp.gate_proj.g_idx": "model-00006-of-00057.safetensors", + "model.layers.9.mlp.gate_proj.qweight": "model-00006-of-00057.safetensors", + "model.layers.9.mlp.gate_proj.qzeros": "model-00006-of-00057.safetensors", + "model.layers.9.mlp.gate_proj.scales": "model-00006-of-00057.safetensors", + "model.layers.9.mlp.up_proj.g_idx": "model-00006-of-00057.safetensors", + "model.layers.9.mlp.up_proj.qweight": "model-00006-of-00057.safetensors", + "model.layers.9.mlp.up_proj.qzeros": "model-00006-of-00057.safetensors", + "model.layers.9.mlp.up_proj.scales": "model-00006-of-00057.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00006-of-00057.safetensors", + "model.layers.9.self_attn.k_proj.g_idx": "model-00005-of-00057.safetensors", + "model.layers.9.self_attn.k_proj.qweight": "model-00005-of-00057.safetensors", + "model.layers.9.self_attn.k_proj.qzeros": "model-00005-of-00057.safetensors", + "model.layers.9.self_attn.k_proj.scales": "model-00005-of-00057.safetensors", + "model.layers.9.self_attn.o_proj.g_idx": "model-00006-of-00057.safetensors", + "model.layers.9.self_attn.o_proj.qweight": "model-00006-of-00057.safetensors", + "model.layers.9.self_attn.o_proj.qzeros": "model-00006-of-00057.safetensors", + "model.layers.9.self_attn.o_proj.scales": "model-00006-of-00057.safetensors", + "model.layers.9.self_attn.q_proj.g_idx": "model-00005-of-00057.safetensors", + "model.layers.9.self_attn.q_proj.qweight": "model-00005-of-00057.safetensors", + "model.layers.9.self_attn.q_proj.qzeros": "model-00005-of-00057.safetensors", + "model.layers.9.self_attn.q_proj.scales": "model-00005-of-00057.safetensors", + "model.layers.9.self_attn.v_proj.g_idx": "model-00005-of-00057.safetensors", + "model.layers.9.self_attn.v_proj.qweight": "model-00005-of-00057.safetensors", + "model.layers.9.self_attn.v_proj.qzeros": "model-00005-of-00057.safetensors", + "model.layers.9.self_attn.v_proj.scales": "model-00005-of-00057.safetensors", + "model.layers.90.input_layernorm.weight": "model-00041-of-00057.safetensors", + "model.layers.90.mlp.down_proj.g_idx": "model-00041-of-00057.safetensors", + "model.layers.90.mlp.down_proj.qweight": "model-00041-of-00057.safetensors", + "model.layers.90.mlp.down_proj.qzeros": "model-00041-of-00057.safetensors", + "model.layers.90.mlp.down_proj.scales": "model-00041-of-00057.safetensors", + "model.layers.90.mlp.gate_proj.g_idx": "model-00040-of-00057.safetensors", + "model.layers.90.mlp.gate_proj.qweight": "model-00040-of-00057.safetensors", + "model.layers.90.mlp.gate_proj.qzeros": "model-00040-of-00057.safetensors", + "model.layers.90.mlp.gate_proj.scales": "model-00040-of-00057.safetensors", + "model.layers.90.mlp.up_proj.g_idx": "model-00040-of-00057.safetensors", + "model.layers.90.mlp.up_proj.qweight": "model-00040-of-00057.safetensors", + "model.layers.90.mlp.up_proj.qzeros": "model-00040-of-00057.safetensors", + "model.layers.90.mlp.up_proj.scales": "model-00040-of-00057.safetensors", + "model.layers.90.post_attention_layernorm.weight": "model-00041-of-00057.safetensors", + "model.layers.90.self_attn.k_proj.g_idx": "model-00040-of-00057.safetensors", + "model.layers.90.self_attn.k_proj.qweight": "model-00040-of-00057.safetensors", + "model.layers.90.self_attn.k_proj.qzeros": "model-00040-of-00057.safetensors", + "model.layers.90.self_attn.k_proj.scales": "model-00040-of-00057.safetensors", + "model.layers.90.self_attn.o_proj.g_idx": "model-00040-of-00057.safetensors", + "model.layers.90.self_attn.o_proj.qweight": "model-00040-of-00057.safetensors", + "model.layers.90.self_attn.o_proj.qzeros": "model-00040-of-00057.safetensors", + "model.layers.90.self_attn.o_proj.scales": "model-00040-of-00057.safetensors", + "model.layers.90.self_attn.q_proj.g_idx": "model-00040-of-00057.safetensors", + "model.layers.90.self_attn.q_proj.qweight": "model-00040-of-00057.safetensors", + "model.layers.90.self_attn.q_proj.qzeros": "model-00040-of-00057.safetensors", + "model.layers.90.self_attn.q_proj.scales": "model-00040-of-00057.safetensors", + "model.layers.90.self_attn.v_proj.g_idx": "model-00040-of-00057.safetensors", + "model.layers.90.self_attn.v_proj.qweight": "model-00040-of-00057.safetensors", + "model.layers.90.self_attn.v_proj.qzeros": "model-00040-of-00057.safetensors", + "model.layers.90.self_attn.v_proj.scales": "model-00040-of-00057.safetensors", + "model.layers.91.input_layernorm.weight": "model-00041-of-00057.safetensors", + "model.layers.91.mlp.down_proj.g_idx": "model-00041-of-00057.safetensors", + "model.layers.91.mlp.down_proj.qweight": "model-00041-of-00057.safetensors", + "model.layers.91.mlp.down_proj.qzeros": "model-00041-of-00057.safetensors", + "model.layers.91.mlp.down_proj.scales": "model-00041-of-00057.safetensors", + "model.layers.91.mlp.gate_proj.g_idx": "model-00041-of-00057.safetensors", + "model.layers.91.mlp.gate_proj.qweight": "model-00041-of-00057.safetensors", + "model.layers.91.mlp.gate_proj.qzeros": "model-00041-of-00057.safetensors", + "model.layers.91.mlp.gate_proj.scales": "model-00041-of-00057.safetensors", + "model.layers.91.mlp.up_proj.g_idx": "model-00041-of-00057.safetensors", + "model.layers.91.mlp.up_proj.qweight": "model-00041-of-00057.safetensors", + "model.layers.91.mlp.up_proj.qzeros": "model-00041-of-00057.safetensors", + "model.layers.91.mlp.up_proj.scales": "model-00041-of-00057.safetensors", + "model.layers.91.post_attention_layernorm.weight": "model-00041-of-00057.safetensors", + "model.layers.91.self_attn.k_proj.g_idx": "model-00041-of-00057.safetensors", + "model.layers.91.self_attn.k_proj.qweight": "model-00041-of-00057.safetensors", + "model.layers.91.self_attn.k_proj.qzeros": "model-00041-of-00057.safetensors", + "model.layers.91.self_attn.k_proj.scales": "model-00041-of-00057.safetensors", + "model.layers.91.self_attn.o_proj.g_idx": "model-00041-of-00057.safetensors", + "model.layers.91.self_attn.o_proj.qweight": "model-00041-of-00057.safetensors", + "model.layers.91.self_attn.o_proj.qzeros": "model-00041-of-00057.safetensors", + "model.layers.91.self_attn.o_proj.scales": "model-00041-of-00057.safetensors", + "model.layers.91.self_attn.q_proj.g_idx": "model-00041-of-00057.safetensors", + "model.layers.91.self_attn.q_proj.qweight": "model-00041-of-00057.safetensors", + "model.layers.91.self_attn.q_proj.qzeros": "model-00041-of-00057.safetensors", + "model.layers.91.self_attn.q_proj.scales": "model-00041-of-00057.safetensors", + "model.layers.91.self_attn.v_proj.g_idx": "model-00041-of-00057.safetensors", + "model.layers.91.self_attn.v_proj.qweight": "model-00041-of-00057.safetensors", + "model.layers.91.self_attn.v_proj.qzeros": "model-00041-of-00057.safetensors", + "model.layers.91.self_attn.v_proj.scales": "model-00041-of-00057.safetensors", + "model.layers.92.input_layernorm.weight": "model-00041-of-00057.safetensors", + "model.layers.92.mlp.down_proj.g_idx": "model-00041-of-00057.safetensors", + "model.layers.92.mlp.down_proj.qweight": "model-00041-of-00057.safetensors", + "model.layers.92.mlp.down_proj.qzeros": "model-00041-of-00057.safetensors", + "model.layers.92.mlp.down_proj.scales": "model-00041-of-00057.safetensors", + "model.layers.92.mlp.gate_proj.g_idx": "model-00041-of-00057.safetensors", + "model.layers.92.mlp.gate_proj.qweight": "model-00041-of-00057.safetensors", + "model.layers.92.mlp.gate_proj.qzeros": "model-00041-of-00057.safetensors", + "model.layers.92.mlp.gate_proj.scales": "model-00041-of-00057.safetensors", + "model.layers.92.mlp.up_proj.g_idx": "model-00041-of-00057.safetensors", + "model.layers.92.mlp.up_proj.qweight": "model-00041-of-00057.safetensors", + "model.layers.92.mlp.up_proj.qzeros": "model-00041-of-00057.safetensors", + "model.layers.92.mlp.up_proj.scales": "model-00041-of-00057.safetensors", + "model.layers.92.post_attention_layernorm.weight": "model-00041-of-00057.safetensors", + "model.layers.92.self_attn.k_proj.g_idx": "model-00041-of-00057.safetensors", + "model.layers.92.self_attn.k_proj.qweight": "model-00041-of-00057.safetensors", + "model.layers.92.self_attn.k_proj.qzeros": "model-00041-of-00057.safetensors", + "model.layers.92.self_attn.k_proj.scales": "model-00041-of-00057.safetensors", + "model.layers.92.self_attn.o_proj.g_idx": "model-00041-of-00057.safetensors", + "model.layers.92.self_attn.o_proj.qweight": "model-00041-of-00057.safetensors", + "model.layers.92.self_attn.o_proj.qzeros": "model-00041-of-00057.safetensors", + "model.layers.92.self_attn.o_proj.scales": "model-00041-of-00057.safetensors", + "model.layers.92.self_attn.q_proj.g_idx": "model-00041-of-00057.safetensors", + "model.layers.92.self_attn.q_proj.qweight": "model-00041-of-00057.safetensors", + "model.layers.92.self_attn.q_proj.qzeros": "model-00041-of-00057.safetensors", + "model.layers.92.self_attn.q_proj.scales": "model-00041-of-00057.safetensors", + "model.layers.92.self_attn.v_proj.g_idx": "model-00041-of-00057.safetensors", + "model.layers.92.self_attn.v_proj.qweight": "model-00041-of-00057.safetensors", + "model.layers.92.self_attn.v_proj.qzeros": "model-00041-of-00057.safetensors", + "model.layers.92.self_attn.v_proj.scales": "model-00041-of-00057.safetensors", + "model.layers.93.input_layernorm.weight": "model-00042-of-00057.safetensors", + "model.layers.93.mlp.down_proj.g_idx": "model-00042-of-00057.safetensors", + "model.layers.93.mlp.down_proj.qweight": "model-00042-of-00057.safetensors", + "model.layers.93.mlp.down_proj.qzeros": "model-00042-of-00057.safetensors", + "model.layers.93.mlp.down_proj.scales": "model-00042-of-00057.safetensors", + "model.layers.93.mlp.gate_proj.g_idx": "model-00042-of-00057.safetensors", + "model.layers.93.mlp.gate_proj.qweight": "model-00042-of-00057.safetensors", + "model.layers.93.mlp.gate_proj.qzeros": "model-00042-of-00057.safetensors", + "model.layers.93.mlp.gate_proj.scales": "model-00042-of-00057.safetensors", + "model.layers.93.mlp.up_proj.g_idx": "model-00042-of-00057.safetensors", + "model.layers.93.mlp.up_proj.qweight": "model-00042-of-00057.safetensors", + "model.layers.93.mlp.up_proj.qzeros": "model-00042-of-00057.safetensors", + "model.layers.93.mlp.up_proj.scales": "model-00042-of-00057.safetensors", + "model.layers.93.post_attention_layernorm.weight": "model-00042-of-00057.safetensors", + "model.layers.93.self_attn.k_proj.g_idx": "model-00041-of-00057.safetensors", + "model.layers.93.self_attn.k_proj.qweight": "model-00041-of-00057.safetensors", + "model.layers.93.self_attn.k_proj.qzeros": "model-00041-of-00057.safetensors", + "model.layers.93.self_attn.k_proj.scales": "model-00041-of-00057.safetensors", + "model.layers.93.self_attn.o_proj.g_idx": "model-00042-of-00057.safetensors", + "model.layers.93.self_attn.o_proj.qweight": "model-00042-of-00057.safetensors", + "model.layers.93.self_attn.o_proj.qzeros": "model-00042-of-00057.safetensors", + "model.layers.93.self_attn.o_proj.scales": "model-00042-of-00057.safetensors", + "model.layers.93.self_attn.q_proj.g_idx": "model-00041-of-00057.safetensors", + "model.layers.93.self_attn.q_proj.qweight": "model-00041-of-00057.safetensors", + "model.layers.93.self_attn.q_proj.qzeros": "model-00041-of-00057.safetensors", + "model.layers.93.self_attn.q_proj.scales": "model-00041-of-00057.safetensors", + "model.layers.93.self_attn.v_proj.g_idx": "model-00041-of-00057.safetensors", + "model.layers.93.self_attn.v_proj.qweight": "model-00041-of-00057.safetensors", + "model.layers.93.self_attn.v_proj.qzeros": "model-00041-of-00057.safetensors", + "model.layers.93.self_attn.v_proj.scales": "model-00041-of-00057.safetensors", + "model.layers.94.input_layernorm.weight": "model-00042-of-00057.safetensors", + "model.layers.94.mlp.down_proj.g_idx": "model-00042-of-00057.safetensors", + "model.layers.94.mlp.down_proj.qweight": "model-00042-of-00057.safetensors", + "model.layers.94.mlp.down_proj.qzeros": "model-00042-of-00057.safetensors", + "model.layers.94.mlp.down_proj.scales": "model-00042-of-00057.safetensors", + "model.layers.94.mlp.gate_proj.g_idx": "model-00042-of-00057.safetensors", + "model.layers.94.mlp.gate_proj.qweight": "model-00042-of-00057.safetensors", + "model.layers.94.mlp.gate_proj.qzeros": "model-00042-of-00057.safetensors", + "model.layers.94.mlp.gate_proj.scales": "model-00042-of-00057.safetensors", + "model.layers.94.mlp.up_proj.g_idx": "model-00042-of-00057.safetensors", + "model.layers.94.mlp.up_proj.qweight": "model-00042-of-00057.safetensors", + "model.layers.94.mlp.up_proj.qzeros": "model-00042-of-00057.safetensors", + "model.layers.94.mlp.up_proj.scales": "model-00042-of-00057.safetensors", + "model.layers.94.post_attention_layernorm.weight": "model-00042-of-00057.safetensors", + "model.layers.94.self_attn.k_proj.g_idx": "model-00042-of-00057.safetensors", + "model.layers.94.self_attn.k_proj.qweight": "model-00042-of-00057.safetensors", + "model.layers.94.self_attn.k_proj.qzeros": "model-00042-of-00057.safetensors", + "model.layers.94.self_attn.k_proj.scales": "model-00042-of-00057.safetensors", + "model.layers.94.self_attn.o_proj.g_idx": "model-00042-of-00057.safetensors", + "model.layers.94.self_attn.o_proj.qweight": "model-00042-of-00057.safetensors", + "model.layers.94.self_attn.o_proj.qzeros": "model-00042-of-00057.safetensors", + "model.layers.94.self_attn.o_proj.scales": "model-00042-of-00057.safetensors", + "model.layers.94.self_attn.q_proj.g_idx": "model-00042-of-00057.safetensors", + "model.layers.94.self_attn.q_proj.qweight": "model-00042-of-00057.safetensors", + "model.layers.94.self_attn.q_proj.qzeros": "model-00042-of-00057.safetensors", + "model.layers.94.self_attn.q_proj.scales": "model-00042-of-00057.safetensors", + "model.layers.94.self_attn.v_proj.g_idx": "model-00042-of-00057.safetensors", + "model.layers.94.self_attn.v_proj.qweight": "model-00042-of-00057.safetensors", + "model.layers.94.self_attn.v_proj.qzeros": "model-00042-of-00057.safetensors", + "model.layers.94.self_attn.v_proj.scales": "model-00042-of-00057.safetensors", + "model.layers.95.input_layernorm.weight": "model-00043-of-00057.safetensors", + "model.layers.95.mlp.down_proj.g_idx": "model-00043-of-00057.safetensors", + "model.layers.95.mlp.down_proj.qweight": "model-00043-of-00057.safetensors", + "model.layers.95.mlp.down_proj.qzeros": "model-00043-of-00057.safetensors", + "model.layers.95.mlp.down_proj.scales": "model-00043-of-00057.safetensors", + "model.layers.95.mlp.gate_proj.g_idx": "model-00042-of-00057.safetensors", + "model.layers.95.mlp.gate_proj.qweight": "model-00042-of-00057.safetensors", + "model.layers.95.mlp.gate_proj.qzeros": "model-00042-of-00057.safetensors", + "model.layers.95.mlp.gate_proj.scales": "model-00042-of-00057.safetensors", + "model.layers.95.mlp.up_proj.g_idx": "model-00043-of-00057.safetensors", + "model.layers.95.mlp.up_proj.qweight": "model-00043-of-00057.safetensors", + "model.layers.95.mlp.up_proj.qzeros": "model-00043-of-00057.safetensors", + "model.layers.95.mlp.up_proj.scales": "model-00043-of-00057.safetensors", + "model.layers.95.post_attention_layernorm.weight": "model-00043-of-00057.safetensors", + "model.layers.95.self_attn.k_proj.g_idx": "model-00042-of-00057.safetensors", + "model.layers.95.self_attn.k_proj.qweight": "model-00042-of-00057.safetensors", + "model.layers.95.self_attn.k_proj.qzeros": "model-00042-of-00057.safetensors", + "model.layers.95.self_attn.k_proj.scales": "model-00042-of-00057.safetensors", + "model.layers.95.self_attn.o_proj.g_idx": "model-00042-of-00057.safetensors", + "model.layers.95.self_attn.o_proj.qweight": "model-00042-of-00057.safetensors", + "model.layers.95.self_attn.o_proj.qzeros": "model-00042-of-00057.safetensors", + "model.layers.95.self_attn.o_proj.scales": "model-00042-of-00057.safetensors", + "model.layers.95.self_attn.q_proj.g_idx": "model-00042-of-00057.safetensors", + "model.layers.95.self_attn.q_proj.qweight": "model-00042-of-00057.safetensors", + "model.layers.95.self_attn.q_proj.qzeros": "model-00042-of-00057.safetensors", + "model.layers.95.self_attn.q_proj.scales": "model-00042-of-00057.safetensors", + "model.layers.95.self_attn.v_proj.g_idx": "model-00042-of-00057.safetensors", + "model.layers.95.self_attn.v_proj.qweight": "model-00042-of-00057.safetensors", + "model.layers.95.self_attn.v_proj.qzeros": "model-00042-of-00057.safetensors", + "model.layers.95.self_attn.v_proj.scales": "model-00042-of-00057.safetensors", + "model.layers.96.input_layernorm.weight": "model-00043-of-00057.safetensors", + "model.layers.96.mlp.down_proj.g_idx": "model-00043-of-00057.safetensors", + "model.layers.96.mlp.down_proj.qweight": "model-00043-of-00057.safetensors", + "model.layers.96.mlp.down_proj.qzeros": "model-00043-of-00057.safetensors", + "model.layers.96.mlp.down_proj.scales": "model-00043-of-00057.safetensors", + "model.layers.96.mlp.gate_proj.g_idx": "model-00043-of-00057.safetensors", + "model.layers.96.mlp.gate_proj.qweight": "model-00043-of-00057.safetensors", + "model.layers.96.mlp.gate_proj.qzeros": "model-00043-of-00057.safetensors", + "model.layers.96.mlp.gate_proj.scales": "model-00043-of-00057.safetensors", + "model.layers.96.mlp.up_proj.g_idx": "model-00043-of-00057.safetensors", + "model.layers.96.mlp.up_proj.qweight": "model-00043-of-00057.safetensors", + "model.layers.96.mlp.up_proj.qzeros": "model-00043-of-00057.safetensors", + "model.layers.96.mlp.up_proj.scales": "model-00043-of-00057.safetensors", + "model.layers.96.post_attention_layernorm.weight": "model-00043-of-00057.safetensors", + "model.layers.96.self_attn.k_proj.g_idx": "model-00043-of-00057.safetensors", + "model.layers.96.self_attn.k_proj.qweight": "model-00043-of-00057.safetensors", + "model.layers.96.self_attn.k_proj.qzeros": "model-00043-of-00057.safetensors", + "model.layers.96.self_attn.k_proj.scales": "model-00043-of-00057.safetensors", + "model.layers.96.self_attn.o_proj.g_idx": "model-00043-of-00057.safetensors", + "model.layers.96.self_attn.o_proj.qweight": "model-00043-of-00057.safetensors", + "model.layers.96.self_attn.o_proj.qzeros": "model-00043-of-00057.safetensors", + "model.layers.96.self_attn.o_proj.scales": "model-00043-of-00057.safetensors", + "model.layers.96.self_attn.q_proj.g_idx": "model-00043-of-00057.safetensors", + "model.layers.96.self_attn.q_proj.qweight": "model-00043-of-00057.safetensors", + "model.layers.96.self_attn.q_proj.qzeros": "model-00043-of-00057.safetensors", + "model.layers.96.self_attn.q_proj.scales": "model-00043-of-00057.safetensors", + "model.layers.96.self_attn.v_proj.g_idx": "model-00043-of-00057.safetensors", + "model.layers.96.self_attn.v_proj.qweight": "model-00043-of-00057.safetensors", + "model.layers.96.self_attn.v_proj.qzeros": "model-00043-of-00057.safetensors", + "model.layers.96.self_attn.v_proj.scales": "model-00043-of-00057.safetensors", + "model.layers.97.input_layernorm.weight": "model-00044-of-00057.safetensors", + "model.layers.97.mlp.down_proj.g_idx": "model-00044-of-00057.safetensors", + "model.layers.97.mlp.down_proj.qweight": "model-00044-of-00057.safetensors", + "model.layers.97.mlp.down_proj.qzeros": "model-00044-of-00057.safetensors", + "model.layers.97.mlp.down_proj.scales": "model-00044-of-00057.safetensors", + "model.layers.97.mlp.gate_proj.g_idx": "model-00043-of-00057.safetensors", + "model.layers.97.mlp.gate_proj.qweight": "model-00043-of-00057.safetensors", + "model.layers.97.mlp.gate_proj.qzeros": "model-00043-of-00057.safetensors", + "model.layers.97.mlp.gate_proj.scales": "model-00043-of-00057.safetensors", + "model.layers.97.mlp.up_proj.g_idx": "model-00043-of-00057.safetensors", + "model.layers.97.mlp.up_proj.qweight": "model-00043-of-00057.safetensors", + "model.layers.97.mlp.up_proj.qzeros": "model-00043-of-00057.safetensors", + "model.layers.97.mlp.up_proj.scales": "model-00043-of-00057.safetensors", + "model.layers.97.post_attention_layernorm.weight": "model-00044-of-00057.safetensors", + "model.layers.97.self_attn.k_proj.g_idx": "model-00043-of-00057.safetensors", + "model.layers.97.self_attn.k_proj.qweight": "model-00043-of-00057.safetensors", + "model.layers.97.self_attn.k_proj.qzeros": "model-00043-of-00057.safetensors", + "model.layers.97.self_attn.k_proj.scales": "model-00043-of-00057.safetensors", + "model.layers.97.self_attn.o_proj.g_idx": "model-00043-of-00057.safetensors", + "model.layers.97.self_attn.o_proj.qweight": "model-00043-of-00057.safetensors", + "model.layers.97.self_attn.o_proj.qzeros": "model-00043-of-00057.safetensors", + "model.layers.97.self_attn.o_proj.scales": "model-00043-of-00057.safetensors", + "model.layers.97.self_attn.q_proj.g_idx": "model-00043-of-00057.safetensors", + "model.layers.97.self_attn.q_proj.qweight": "model-00043-of-00057.safetensors", + "model.layers.97.self_attn.q_proj.qzeros": "model-00043-of-00057.safetensors", + "model.layers.97.self_attn.q_proj.scales": "model-00043-of-00057.safetensors", + "model.layers.97.self_attn.v_proj.g_idx": "model-00043-of-00057.safetensors", + "model.layers.97.self_attn.v_proj.qweight": "model-00043-of-00057.safetensors", + "model.layers.97.self_attn.v_proj.qzeros": "model-00043-of-00057.safetensors", + "model.layers.97.self_attn.v_proj.scales": "model-00043-of-00057.safetensors", + "model.layers.98.input_layernorm.weight": "model-00044-of-00057.safetensors", + "model.layers.98.mlp.down_proj.g_idx": "model-00044-of-00057.safetensors", + "model.layers.98.mlp.down_proj.qweight": "model-00044-of-00057.safetensors", + "model.layers.98.mlp.down_proj.qzeros": "model-00044-of-00057.safetensors", + "model.layers.98.mlp.down_proj.scales": "model-00044-of-00057.safetensors", + "model.layers.98.mlp.gate_proj.g_idx": "model-00044-of-00057.safetensors", + "model.layers.98.mlp.gate_proj.qweight": "model-00044-of-00057.safetensors", + "model.layers.98.mlp.gate_proj.qzeros": "model-00044-of-00057.safetensors", + "model.layers.98.mlp.gate_proj.scales": "model-00044-of-00057.safetensors", + "model.layers.98.mlp.up_proj.g_idx": "model-00044-of-00057.safetensors", + "model.layers.98.mlp.up_proj.qweight": "model-00044-of-00057.safetensors", + "model.layers.98.mlp.up_proj.qzeros": "model-00044-of-00057.safetensors", + "model.layers.98.mlp.up_proj.scales": "model-00044-of-00057.safetensors", + "model.layers.98.post_attention_layernorm.weight": "model-00044-of-00057.safetensors", + "model.layers.98.self_attn.k_proj.g_idx": "model-00044-of-00057.safetensors", + "model.layers.98.self_attn.k_proj.qweight": "model-00044-of-00057.safetensors", + "model.layers.98.self_attn.k_proj.qzeros": "model-00044-of-00057.safetensors", + "model.layers.98.self_attn.k_proj.scales": "model-00044-of-00057.safetensors", + "model.layers.98.self_attn.o_proj.g_idx": "model-00044-of-00057.safetensors", + "model.layers.98.self_attn.o_proj.qweight": "model-00044-of-00057.safetensors", + "model.layers.98.self_attn.o_proj.qzeros": "model-00044-of-00057.safetensors", + "model.layers.98.self_attn.o_proj.scales": "model-00044-of-00057.safetensors", + "model.layers.98.self_attn.q_proj.g_idx": "model-00044-of-00057.safetensors", + "model.layers.98.self_attn.q_proj.qweight": "model-00044-of-00057.safetensors", + "model.layers.98.self_attn.q_proj.qzeros": "model-00044-of-00057.safetensors", + "model.layers.98.self_attn.q_proj.scales": "model-00044-of-00057.safetensors", + "model.layers.98.self_attn.v_proj.g_idx": "model-00044-of-00057.safetensors", + "model.layers.98.self_attn.v_proj.qweight": "model-00044-of-00057.safetensors", + "model.layers.98.self_attn.v_proj.qzeros": "model-00044-of-00057.safetensors", + "model.layers.98.self_attn.v_proj.scales": "model-00044-of-00057.safetensors", + "model.layers.99.input_layernorm.weight": "model-00044-of-00057.safetensors", + "model.layers.99.mlp.down_proj.g_idx": "model-00044-of-00057.safetensors", + "model.layers.99.mlp.down_proj.qweight": "model-00044-of-00057.safetensors", + "model.layers.99.mlp.down_proj.qzeros": "model-00044-of-00057.safetensors", + "model.layers.99.mlp.down_proj.scales": "model-00044-of-00057.safetensors", + "model.layers.99.mlp.gate_proj.g_idx": "model-00044-of-00057.safetensors", + "model.layers.99.mlp.gate_proj.qweight": "model-00044-of-00057.safetensors", + "model.layers.99.mlp.gate_proj.qzeros": "model-00044-of-00057.safetensors", + "model.layers.99.mlp.gate_proj.scales": "model-00044-of-00057.safetensors", + "model.layers.99.mlp.up_proj.g_idx": "model-00044-of-00057.safetensors", + "model.layers.99.mlp.up_proj.qweight": "model-00044-of-00057.safetensors", + "model.layers.99.mlp.up_proj.qzeros": "model-00044-of-00057.safetensors", + "model.layers.99.mlp.up_proj.scales": "model-00044-of-00057.safetensors", + "model.layers.99.post_attention_layernorm.weight": "model-00044-of-00057.safetensors", + "model.layers.99.self_attn.k_proj.g_idx": "model-00044-of-00057.safetensors", + "model.layers.99.self_attn.k_proj.qweight": "model-00044-of-00057.safetensors", + "model.layers.99.self_attn.k_proj.qzeros": "model-00044-of-00057.safetensors", + "model.layers.99.self_attn.k_proj.scales": "model-00044-of-00057.safetensors", + "model.layers.99.self_attn.o_proj.g_idx": "model-00044-of-00057.safetensors", + "model.layers.99.self_attn.o_proj.qweight": "model-00044-of-00057.safetensors", + "model.layers.99.self_attn.o_proj.qzeros": "model-00044-of-00057.safetensors", + "model.layers.99.self_attn.o_proj.scales": "model-00044-of-00057.safetensors", + "model.layers.99.self_attn.q_proj.g_idx": "model-00044-of-00057.safetensors", + "model.layers.99.self_attn.q_proj.qweight": "model-00044-of-00057.safetensors", + "model.layers.99.self_attn.q_proj.qzeros": "model-00044-of-00057.safetensors", + "model.layers.99.self_attn.q_proj.scales": "model-00044-of-00057.safetensors", + "model.layers.99.self_attn.v_proj.g_idx": "model-00044-of-00057.safetensors", + "model.layers.99.self_attn.v_proj.qweight": "model-00044-of-00057.safetensors", + "model.layers.99.self_attn.v_proj.qzeros": "model-00044-of-00057.safetensors", + "model.layers.99.self_attn.v_proj.scales": "model-00044-of-00057.safetensors", + "model.norm.weight": "model-00057-of-00057.safetensors" + } +} diff --git a/quant_log.csv b/quant_log.csv new file mode 100644 index 0000000000000000000000000000000000000000..ef4f06233315db50fdb3a6b4659cc4515cb08ced --- /dev/null +++ b/quant_log.csv @@ -0,0 +1,883 @@ +layer,module,loss,samples,damp,time +0,self_attn.k_proj,0.01340501,0.01000,3.285 +0,self_attn.v_proj,0.00001147,0.01000,3.020 +0,self_attn.q_proj,0.00805894,0.01000,3.195 +0,self_attn.o_proj,637.13763428,0.01250,3.586 +0,mlp.up_proj,0.00058788,0.01000,3.921 +0,mlp.gate_proj,0.00059290,0.01000,3.949 +0,mlp.down_proj,0.00002978,0.01000,24.164 +1,self_attn.k_proj,0.00865047,0.01000,3.008 +1,self_attn.v_proj,0.00037662,0.01000,2.980 +1,self_attn.q_proj,0.03175554,0.01000,3.140 +1,self_attn.o_proj,0.00000977,0.01000,3.185 +1,mlp.up_proj,0.00248405,0.01000,3.792 +1,mlp.gate_proj,0.00236319,0.01000,3.835 +1,mlp.down_proj,0.00000471,0.01000,24.071 +2,self_attn.k_proj,0.00436026,0.01000,3.061 +2,self_attn.v_proj,0.00048939,0.01000,3.043 +2,self_attn.q_proj,0.01531996,0.01000,3.254 +2,self_attn.o_proj,0.00001726,0.01000,3.213 +2,mlp.up_proj,0.00660630,0.01000,3.797 +2,mlp.gate_proj,0.00644774,0.01000,3.746 +2,mlp.down_proj,0.00002189,0.01000,24.023 +3,self_attn.k_proj,0.00614090,0.01000,3.035 +3,self_attn.v_proj,0.00043810,0.01000,2.991 +3,self_attn.q_proj,0.02046010,0.01000,3.170 +3,self_attn.o_proj,0.00001271,0.01000,3.293 +3,mlp.up_proj,0.01210508,0.01000,3.786 +3,mlp.gate_proj,0.01214893,0.01000,3.755 +3,mlp.down_proj,0.00004038,0.01000,24.034 +4,self_attn.k_proj,0.00427368,0.01000,3.013 +4,self_attn.v_proj,0.00043675,0.01000,2.969 +4,self_attn.q_proj,0.01451525,0.01000,3.151 +4,self_attn.o_proj,0.00002606,0.01000,3.161 +4,mlp.up_proj,0.01569575,0.01000,3.801 +4,mlp.gate_proj,0.01589233,0.01000,3.749 +4,mlp.down_proj,0.00007078,0.01000,24.108 +5,self_attn.k_proj,0.00375222,0.01000,3.017 +5,self_attn.v_proj,0.00027751,0.01000,2.975 +5,self_attn.q_proj,0.01916856,0.01000,3.160 +5,self_attn.o_proj,626.70428467,0.01250,3.562 +5,mlp.up_proj,0.01245417,0.01000,3.786 +5,mlp.gate_proj,0.01285015,0.01000,3.752 +5,mlp.down_proj,0.39059037,0.01250,37.606 +6,self_attn.k_proj,0.00632325,0.01000,3.033 +6,self_attn.v_proj,0.00159276,0.01000,2.970 +6,self_attn.q_proj,0.03263030,0.01000,3.157 +6,self_attn.o_proj,0.00001922,0.01000,3.163 +6,mlp.up_proj,0.01734336,0.01000,3.801 +6,mlp.gate_proj,0.01735633,0.01000,3.758 +6,mlp.down_proj,0.00005175,0.01000,24.049 +7,self_attn.k_proj,0.00885481,0.01000,3.004 +7,self_attn.v_proj,0.00294300,0.01000,2.963 +7,self_attn.q_proj,0.05252545,0.01000,3.153 +7,self_attn.o_proj,1215.37890625,0.01250,3.574 +7,mlp.up_proj,0.01996547,0.01000,3.785 +7,mlp.gate_proj,0.02006237,0.01000,3.742 +7,mlp.down_proj,0.00006615,0.01000,24.008 +8,self_attn.k_proj,0.00615571,0.01000,2.995 +8,self_attn.v_proj,0.00207685,0.01000,2.967 +8,self_attn.q_proj,0.02982869,0.01000,3.149 +8,self_attn.o_proj,1600.84008789,0.01250,3.549 +8,mlp.up_proj,0.02876170,0.01000,3.807 +8,mlp.gate_proj,0.04372563,0.01000,3.864 +8,mlp.down_proj,0.00008887,0.01000,23.941 +9,self_attn.k_proj,0.01052244,0.01000,3.007 +9,self_attn.v_proj,0.00357798,0.01000,2.978 +9,self_attn.q_proj,0.05744779,0.01000,3.292 +9,self_attn.o_proj,0.00003156,0.01000,3.155 +9,mlp.up_proj,0.03711343,0.01000,3.788 +9,mlp.gate_proj,0.08362643,0.01000,3.748 +9,mlp.down_proj,0.00013371,0.01000,23.992 +10,self_attn.k_proj,0.01159614,0.01000,2.988 +10,self_attn.v_proj,0.00338568,0.01000,2.970 +10,self_attn.q_proj,0.05602583,0.01000,3.147 +10,self_attn.o_proj,0.00005510,0.01000,3.291 +10,mlp.up_proj,0.04480598,0.01000,3.798 +10,mlp.gate_proj,0.11546280,0.01000,3.900 +10,mlp.down_proj,0.00018933,0.01000,23.977 +11,self_attn.k_proj,0.01268323,0.01000,3.018 +11,self_attn.v_proj,0.00340545,0.01000,2.964 +11,self_attn.q_proj,0.05780433,0.01000,3.146 +11,self_attn.o_proj,0.00007052,0.01000,3.161 +11,mlp.up_proj,0.05881945,0.01000,3.969 +11,mlp.gate_proj,0.15850842,0.01000,3.794 +11,mlp.down_proj,0.00027917,0.01000,24.200 +12,self_attn.k_proj,0.02048246,0.01000,2.892 +12,self_attn.v_proj,0.00535795,0.01000,2.852 +12,self_attn.q_proj,0.10532396,0.01000,3.043 +12,self_attn.o_proj,0.00015830,0.01000,3.057 +12,mlp.up_proj,0.08990082,0.01000,3.709 +12,mlp.gate_proj,0.25874731,0.01000,3.659 +12,mlp.down_proj,0.00050724,0.01000,23.583 +13,self_attn.k_proj,0.01251122,0.01000,2.974 +13,self_attn.v_proj,0.00355584,0.01000,2.913 +13,self_attn.q_proj,0.06038317,0.01000,3.187 +13,self_attn.o_proj,0.00010240,0.01000,3.046 +13,mlp.up_proj,0.12299667,0.01000,3.704 +13,mlp.gate_proj,0.25547975,0.01000,3.663 +13,mlp.down_proj,0.00068327,0.01000,23.828 +14,self_attn.k_proj,0.01330811,0.01000,2.872 +14,self_attn.v_proj,0.00331716,0.01000,2.897 +14,self_attn.q_proj,0.05191506,0.01000,3.038 +14,self_attn.o_proj,0.00005780,0.01000,3.039 +14,mlp.up_proj,0.14787352,0.01000,3.709 +14,mlp.gate_proj,0.16315928,0.01000,3.665 +14,mlp.down_proj,0.00093976,0.01000,23.720 +15,self_attn.k_proj,0.00641087,0.01000,2.881 +15,self_attn.v_proj,0.00170123,0.01000,2.840 +15,self_attn.q_proj,0.01785612,0.01000,3.032 +15,self_attn.o_proj,0.00008176,0.01000,3.041 +15,mlp.up_proj,0.14926648,0.01000,3.849 +15,mlp.gate_proj,0.15955693,0.01000,3.789 +15,mlp.down_proj,0.00093947,0.01000,23.880 +16,self_attn.k_proj,0.02037925,0.01000,2.864 +16,self_attn.v_proj,0.00525839,0.01000,2.844 +16,self_attn.q_proj,0.09276567,0.01000,3.033 +16,self_attn.o_proj,0.00021556,0.01000,3.041 +16,mlp.up_proj,0.12637484,0.01000,3.727 +16,mlp.gate_proj,0.13565056,0.01000,3.667 +16,mlp.down_proj,0.00086744,0.01000,23.613 +17,self_attn.k_proj,0.01697640,0.01000,2.877 +17,self_attn.v_proj,0.00485271,0.01000,2.854 +17,self_attn.q_proj,0.08043237,0.01000,3.040 +17,self_attn.o_proj,0.00045735,0.01000,3.163 +17,mlp.up_proj,0.12374384,0.01000,3.717 +17,mlp.gate_proj,0.13359484,0.01000,3.662 +17,mlp.down_proj,0.00098586,0.01000,23.612 +18,self_attn.k_proj,0.04538812,0.01000,2.871 +18,self_attn.v_proj,0.01017945,0.01000,2.918 +18,self_attn.q_proj,0.21874514,0.01000,3.043 +18,self_attn.o_proj,0.00068680,0.01000,3.097 +18,mlp.up_proj,0.13400060,0.01000,3.701 +18,mlp.gate_proj,0.14101657,0.01000,3.654 +18,mlp.down_proj,0.00122959,0.01000,23.619 +19,self_attn.k_proj,0.04987283,0.01000,2.876 +19,self_attn.v_proj,0.01152044,0.01000,2.852 +19,self_attn.q_proj,0.23667189,0.01000,3.034 +19,self_attn.o_proj,0.00088504,0.01000,3.051 +19,mlp.up_proj,0.16981483,0.01000,3.699 +19,mlp.gate_proj,0.17935821,0.01000,3.665 +19,mlp.down_proj,0.00161302,0.01000,23.574 +20,self_attn.k_proj,0.05427954,0.01000,2.861 +20,self_attn.v_proj,0.01272871,0.01000,2.834 +20,self_attn.q_proj,0.24049807,0.01000,3.027 +20,self_attn.o_proj,0.00070158,0.01000,3.040 +20,mlp.up_proj,0.22542092,0.01000,3.684 +20,mlp.gate_proj,0.24176314,0.01000,3.756 +20,mlp.down_proj,0.00204766,0.01000,23.569 +21,self_attn.k_proj,0.02494409,0.01000,2.855 +21,self_attn.v_proj,0.00563191,0.01000,2.835 +21,self_attn.q_proj,0.08746912,0.01000,3.029 +21,self_attn.o_proj,0.00031802,0.01000,3.164 +21,mlp.up_proj,0.27131084,0.01000,3.702 +21,mlp.gate_proj,0.29088163,0.01000,3.666 +21,mlp.down_proj,0.00235122,0.01000,23.866 +22,self_attn.k_proj,0.03556939,0.01000,2.869 +22,self_attn.v_proj,0.00862655,0.01000,2.841 +22,self_attn.q_proj,0.13518177,0.01000,3.028 +22,self_attn.o_proj,0.00061836,0.01000,3.039 +22,mlp.up_proj,0.30826789,0.01000,3.699 +22,mlp.gate_proj,0.32763052,0.01000,3.662 +22,mlp.down_proj,0.00285140,0.01000,23.919 +23,self_attn.k_proj,0.05327557,0.01000,2.874 +23,self_attn.v_proj,0.01095798,0.01000,2.851 +23,self_attn.q_proj,0.18898712,0.01000,3.041 +23,self_attn.o_proj,0.00056104,0.01000,3.044 +23,mlp.up_proj,0.36285511,0.01000,3.712 +23,mlp.gate_proj,0.37368101,0.01000,3.658 +23,mlp.down_proj,0.00361535,0.01000,24.233 +24,self_attn.k_proj,0.03823718,0.01000,2.868 +24,self_attn.v_proj,0.00773104,0.01000,2.845 +24,self_attn.q_proj,0.10366546,0.01000,3.034 +24,self_attn.o_proj,0.00022847,0.01000,3.037 +24,mlp.up_proj,0.42964947,0.01000,3.712 +24,mlp.gate_proj,0.44238710,0.01000,3.663 +24,mlp.down_proj,0.00405051,0.01000,23.933 +25,self_attn.k_proj,0.04427643,0.01000,2.874 +25,self_attn.v_proj,0.00842728,0.01000,2.849 +25,self_attn.q_proj,0.11555678,0.01000,3.037 +25,self_attn.o_proj,0.00054910,0.01000,3.115 +25,mlp.up_proj,0.48171091,0.01000,3.756 +25,mlp.gate_proj,0.49526206,0.01000,3.765 +25,mlp.down_proj,0.00455085,0.01000,23.683 +26,self_attn.k_proj,0.05220067,0.01000,2.873 +26,self_attn.v_proj,0.01245648,0.01000,2.840 +26,self_attn.q_proj,0.15528464,0.01000,3.035 +26,self_attn.o_proj,0.00044229,0.01000,3.041 +26,mlp.up_proj,0.53073263,0.01000,3.710 +26,mlp.gate_proj,0.54655004,0.01000,3.662 +26,mlp.down_proj,0.00514483,0.01000,23.606 +27,self_attn.k_proj,0.07086261,0.01000,2.869 +27,self_attn.v_proj,0.01621779,0.01000,2.841 +27,self_attn.q_proj,0.22274941,0.01000,3.036 +27,self_attn.o_proj,0.00056478,0.01000,3.043 +27,mlp.up_proj,0.58440363,0.01000,3.705 +27,mlp.gate_proj,0.60328418,0.01000,3.654 +27,mlp.down_proj,0.00590638,0.01000,23.601 +28,self_attn.k_proj,0.06540029,0.01000,2.877 +28,self_attn.v_proj,0.01417490,0.01000,2.849 +28,self_attn.q_proj,0.21564806,0.01000,3.042 +28,self_attn.o_proj,0.00086175,0.01000,3.048 +28,mlp.up_proj,0.64220130,0.01000,3.706 +28,mlp.gate_proj,0.66457200,0.01000,3.656 +28,mlp.down_proj,0.00696370,0.01000,23.649 +29,self_attn.k_proj,0.06633534,0.01000,2.874 +29,self_attn.v_proj,0.01460105,0.01000,2.845 +29,self_attn.q_proj,0.20281756,0.01000,3.035 +29,self_attn.o_proj,0.00085880,0.01000,3.044 +29,mlp.up_proj,0.69239128,0.01000,3.722 +29,mlp.gate_proj,0.71667749,0.01000,3.667 +29,mlp.down_proj,0.00798881,0.01000,23.854 +30,self_attn.k_proj,0.06208337,0.01000,2.923 +30,self_attn.v_proj,0.01452993,0.01000,2.848 +30,self_attn.q_proj,0.22849074,0.01000,3.040 +30,self_attn.o_proj,39.02008820,0.01250,3.446 +30,mlp.up_proj,0.74068320,0.01000,3.723 +30,mlp.gate_proj,0.76780957,0.01000,3.665 +30,mlp.down_proj,0.00959350,0.01000,23.704 +31,self_attn.k_proj,0.04250758,0.01000,2.928 +31,self_attn.v_proj,0.01003543,0.01000,2.843 +31,self_attn.q_proj,0.11947796,0.01000,3.034 +31,self_attn.o_proj,0.00048994,0.01000,3.044 +31,mlp.up_proj,0.80648619,0.01000,3.877 +31,mlp.gate_proj,0.83697611,0.01000,3.725 +31,mlp.down_proj,0.01128164,0.01000,23.931 +32,self_attn.k_proj,0.06081090,0.01000,2.870 +32,self_attn.v_proj,0.01490181,0.01000,2.842 +32,self_attn.q_proj,0.19380055,0.01000,3.037 +32,self_attn.o_proj,0.00080986,0.01000,3.220 +32,mlp.up_proj,0.86384434,0.01000,3.709 +32,mlp.gate_proj,0.89658046,0.01000,3.669 +32,mlp.down_proj,0.01298850,0.01000,23.685 +33,self_attn.k_proj,0.11575995,0.01000,2.863 +33,self_attn.v_proj,0.02088767,0.01000,2.852 +33,self_attn.q_proj,0.46211475,0.01000,3.038 +33,self_attn.o_proj,0.00328360,0.01000,3.057 +33,mlp.up_proj,0.89181459,0.01000,3.747 +33,mlp.gate_proj,0.92805189,0.01000,3.668 +33,mlp.down_proj,0.01596361,0.01000,23.972 +34,self_attn.k_proj,0.13273700,0.01000,2.915 +34,self_attn.v_proj,0.02182782,0.01000,2.853 +34,self_attn.q_proj,0.54459333,0.01000,3.071 +34,self_attn.o_proj,0.00468417,0.01000,3.049 +34,mlp.up_proj,0.91019976,0.01000,3.844 +34,mlp.gate_proj,0.95086002,0.01000,3.693 +34,mlp.down_proj,3.75284410,0.01250,37.177 +35,self_attn.k_proj,0.11273406,0.01000,2.870 +35,self_attn.v_proj,0.01845640,0.01000,2.843 +35,self_attn.q_proj,0.50896466,0.01000,3.037 +35,self_attn.o_proj,0.00834611,0.01000,3.092 +35,mlp.up_proj,0.91458881,0.01000,3.736 +35,mlp.gate_proj,0.95795333,0.01000,3.661 +35,mlp.down_proj,0.02598865,0.01000,23.646 +36,self_attn.k_proj,0.21857107,0.01000,2.907 +36,self_attn.v_proj,0.02554663,0.01000,2.859 +36,self_attn.q_proj,1.00961685,0.01000,3.038 +36,self_attn.o_proj,0.01518390,0.01000,3.096 +36,mlp.up_proj,1.07854950,0.01000,3.822 +36,mlp.gate_proj,1.13364911,0.01000,3.805 +36,mlp.down_proj,0.03592963,0.01000,23.598 +37,self_attn.k_proj,0.25464824,0.01000,2.885 +37,self_attn.v_proj,0.02577686,0.01000,2.853 +37,self_attn.q_proj,1.04892945,0.01000,3.038 +37,self_attn.o_proj,0.01800542,0.01000,3.044 +37,mlp.up_proj,1.29188228,0.01000,3.701 +37,mlp.gate_proj,1.35657763,0.01000,3.660 +37,mlp.down_proj,0.04599481,0.01000,23.596 +38,self_attn.k_proj,0.28796399,0.01000,2.892 +38,self_attn.v_proj,0.03155093,0.01000,2.857 +38,self_attn.q_proj,1.17120147,0.01000,3.041 +38,self_attn.o_proj,0.02665423,0.01000,3.050 +38,mlp.up_proj,1.50040531,0.01000,3.714 +38,mlp.gate_proj,1.58298051,0.01000,3.658 +38,mlp.down_proj,0.06129718,0.01000,23.586 +39,self_attn.k_proj,0.31726360,0.01000,2.917 +39,self_attn.v_proj,0.02809570,0.01000,2.857 +39,self_attn.q_proj,1.20108891,0.01000,3.047 +39,self_attn.o_proj,0.03054131,0.01000,3.051 +39,mlp.up_proj,1.62905383,0.01000,3.767 +39,mlp.gate_proj,1.72419214,0.01000,3.735 +39,mlp.down_proj,0.07263771,0.01000,23.599 +40,self_attn.k_proj,0.29717666,0.01000,2.877 +40,self_attn.v_proj,0.03210151,0.01000,2.853 +40,self_attn.q_proj,1.11615050,0.01000,3.044 +40,self_attn.o_proj,0.03615309,0.01000,3.061 +40,mlp.up_proj,1.73242307,0.01000,3.696 +40,mlp.gate_proj,1.83084393,0.01000,3.660 +40,mlp.down_proj,0.08665400,0.01000,23.727 +41,self_attn.k_proj,0.35400414,0.01000,2.892 +41,self_attn.v_proj,0.03717133,0.01000,2.855 +41,self_attn.q_proj,1.41656673,0.01000,3.046 +41,self_attn.o_proj,0.04859113,0.01000,3.048 +41,mlp.up_proj,1.99087596,0.01000,3.726 +41,mlp.gate_proj,2.09421349,0.01000,3.668 +41,mlp.down_proj,0.10088766,0.01000,23.793 +42,self_attn.k_proj,0.37885195,0.01000,2.881 +42,self_attn.v_proj,0.04624011,0.01000,2.850 +42,self_attn.q_proj,1.52454686,0.01000,3.042 +42,self_attn.o_proj,0.04581728,0.01000,3.049 +42,mlp.up_proj,2.23701382,0.01000,3.691 +42,mlp.gate_proj,2.34554672,0.01000,3.658 +42,mlp.down_proj,0.11399209,0.01000,23.871 +43,self_attn.k_proj,0.38304120,0.01000,2.898 +43,self_attn.v_proj,0.04468669,0.01000,2.850 +43,self_attn.q_proj,1.46446180,0.01000,3.039 +43,self_attn.o_proj,0.05251001,0.01000,3.069 +43,mlp.up_proj,2.37427568,0.01000,3.720 +43,mlp.gate_proj,2.48136878,0.01000,3.676 +43,mlp.down_proj,0.12359466,0.01000,23.662 +44,self_attn.k_proj,0.44479889,0.01000,2.897 +44,self_attn.v_proj,0.05272421,0.01000,2.863 +44,self_attn.q_proj,1.61379039,0.01000,3.075 +44,self_attn.o_proj,0.05454624,0.01000,3.050 +44,mlp.up_proj,2.43377733,0.01000,3.771 +44,mlp.gate_proj,2.51743793,0.01000,3.674 +44,mlp.down_proj,0.12756382,0.01000,23.740 +45,self_attn.k_proj,0.48751697,0.01000,2.882 +45,self_attn.v_proj,0.05374137,0.01000,2.856 +45,self_attn.q_proj,1.82801819,0.01000,3.051 +45,self_attn.o_proj,0.05500936,0.01000,3.101 +45,mlp.up_proj,2.52707529,0.01000,3.723 +45,mlp.gate_proj,2.62176514,0.01000,3.669 +45,mlp.down_proj,0.13558790,0.01000,23.694 +46,self_attn.k_proj,0.44735336,0.01000,2.889 +46,self_attn.v_proj,0.03933088,0.01000,2.852 +46,self_attn.q_proj,1.55006874,0.01000,3.040 +46,self_attn.o_proj,0.06336483,0.01000,3.052 +46,mlp.up_proj,2.51128960,0.01000,3.713 +46,mlp.gate_proj,2.61802244,0.01000,3.659 +46,mlp.down_proj,0.14045243,0.01000,23.865 +47,self_attn.k_proj,0.49522117,0.01000,2.881 +47,self_attn.v_proj,0.04373708,0.01000,2.848 +47,self_attn.q_proj,1.55803537,0.01000,3.044 +47,self_attn.o_proj,0.06389661,0.01000,3.161 +47,mlp.up_proj,2.51895142,0.01000,3.724 +47,mlp.gate_proj,2.66818476,0.01000,3.670 +47,mlp.down_proj,0.14550176,0.01000,23.745 +48,self_attn.k_proj,0.36102581,0.01000,2.868 +48,self_attn.v_proj,0.04326666,0.01000,2.839 +48,self_attn.q_proj,1.07343698,0.01000,3.038 +48,self_attn.o_proj,0.03980022,0.01000,3.058 +48,mlp.up_proj,2.58153296,0.01000,3.709 +48,mlp.gate_proj,2.77953815,0.01000,3.662 +48,mlp.down_proj,0.13908562,0.01000,23.791 +49,self_attn.k_proj,0.40504301,0.01000,2.874 +49,self_attn.v_proj,0.04752368,0.01000,2.848 +49,self_attn.q_proj,1.15267301,0.01000,3.050 +49,self_attn.o_proj,0.03977250,0.01000,3.057 +49,mlp.up_proj,2.73269010,0.01000,3.723 +49,mlp.gate_proj,2.99971056,0.01000,3.667 +49,mlp.down_proj,0.13276882,0.01000,23.744 +50,self_attn.k_proj,0.40104854,0.01000,2.869 +50,self_attn.v_proj,0.05431717,0.01000,2.840 +50,self_attn.q_proj,1.07831752,0.01000,3.038 +50,self_attn.o_proj,0.02396395,0.01000,3.051 +50,mlp.up_proj,2.82059050,0.01000,3.717 +50,mlp.gate_proj,3.11768246,0.01000,3.673 +50,mlp.down_proj,0.12617889,0.01000,23.663 +51,self_attn.k_proj,0.34904447,0.01000,2.915 +51,self_attn.v_proj,0.05765019,0.01000,2.869 +51,self_attn.q_proj,0.96013999,0.01000,3.044 +51,self_attn.o_proj,0.02243298,0.01000,3.051 +51,mlp.up_proj,2.94489574,0.01000,3.772 +51,mlp.gate_proj,3.25056434,0.01000,3.672 +51,mlp.down_proj,0.12476718,0.01000,23.725 +52,self_attn.k_proj,0.38028318,0.01000,2.872 +52,self_attn.v_proj,0.06023370,0.01000,2.868 +52,self_attn.q_proj,0.94192517,0.01000,3.046 +52,self_attn.o_proj,0.02262055,0.01000,3.039 +52,mlp.up_proj,3.00342035,0.01000,3.838 +52,mlp.gate_proj,3.27413774,0.01000,3.698 +52,mlp.down_proj,0.12815255,0.01000,23.572 +53,self_attn.k_proj,0.39918196,0.01000,2.990 +53,self_attn.v_proj,0.05320951,0.01000,2.959 +53,self_attn.q_proj,1.01303983,0.01000,3.150 +53,self_attn.o_proj,0.02987859,0.01000,3.171 +53,mlp.up_proj,2.94009781,0.01000,3.809 +53,mlp.gate_proj,3.24977589,0.01000,3.760 +53,mlp.down_proj,0.13589117,0.01000,23.977 +54,self_attn.k_proj,0.29156125,0.01000,3.117 +54,self_attn.v_proj,0.05463770,0.01000,3.039 +54,self_attn.q_proj,0.69465566,0.01000,3.159 +54,self_attn.o_proj,0.01863490,0.01000,3.165 +54,mlp.up_proj,3.04756975,0.01000,3.817 +54,mlp.gate_proj,3.42895651,0.01000,3.757 +54,mlp.down_proj,0.13801783,0.01000,23.974 +55,self_attn.k_proj,0.29436815,0.01000,2.983 +55,self_attn.v_proj,0.06037501,0.01000,2.962 +55,self_attn.q_proj,0.69079077,0.01000,3.146 +55,self_attn.o_proj,0.02140802,0.01000,3.161 +55,mlp.up_proj,3.16579223,0.01000,3.928 +55,mlp.gate_proj,3.56116772,0.01000,3.826 +55,mlp.down_proj,0.14120401,0.01000,23.967 +56,self_attn.k_proj,0.32151800,0.01000,3.002 +56,self_attn.v_proj,0.06054703,0.01000,2.980 +56,self_attn.q_proj,0.72300643,0.01000,3.160 +56,self_attn.o_proj,0.01343636,0.01000,3.162 +56,mlp.up_proj,3.28707027,0.01000,3.802 +56,mlp.gate_proj,3.68873525,0.01000,3.765 +56,mlp.down_proj,0.14419352,0.01000,23.997 +57,self_attn.k_proj,0.32900959,0.01000,2.995 +57,self_attn.v_proj,0.06180292,0.01000,2.969 +57,self_attn.q_proj,0.72549963,0.01000,3.156 +57,self_attn.o_proj,0.01484916,0.01000,3.216 +57,mlp.up_proj,3.41336131,0.01000,3.856 +57,mlp.gate_proj,3.83690763,0.01000,3.816 +57,mlp.down_proj,0.15301809,0.01000,24.022 +58,self_attn.k_proj,0.47999173,0.01000,2.990 +58,self_attn.v_proj,0.06798664,0.01000,2.966 +58,self_attn.q_proj,1.21168077,0.01000,3.163 +58,self_attn.o_proj,0.02610476,0.01000,3.160 +58,mlp.up_proj,3.48228192,0.01000,3.790 +58,mlp.gate_proj,3.94697499,0.01000,3.756 +58,mlp.down_proj,0.15896690,0.01000,23.998 +59,self_attn.k_proj,0.51474702,0.01000,3.098 +59,self_attn.v_proj,0.07658368,0.01000,3.031 +59,self_attn.q_proj,1.34825325,0.01000,3.307 +59,self_attn.o_proj,0.02822880,0.01000,3.203 +59,mlp.up_proj,3.57816696,0.01000,3.813 +59,mlp.gate_proj,4.05744505,0.01000,3.768 +59,mlp.down_proj,0.16594706,0.01000,24.028 +60,self_attn.k_proj,0.42529619,0.01000,3.076 +60,self_attn.v_proj,0.06737301,0.01000,2.991 +60,self_attn.q_proj,1.04251540,0.01000,3.157 +60,self_attn.o_proj,0.01963641,0.01000,3.252 +60,mlp.up_proj,3.64369416,0.01000,3.805 +60,mlp.gate_proj,4.11684561,0.01000,3.776 +60,mlp.down_proj,0.17070547,0.01000,24.203 +61,self_attn.k_proj,0.38535088,0.01000,3.041 +61,self_attn.v_proj,0.05904239,0.01000,3.072 +61,self_attn.q_proj,0.99338090,0.01000,3.281 +61,self_attn.o_proj,0.04793409,0.01000,3.157 +61,mlp.up_proj,3.69138598,0.01000,3.815 +61,mlp.gate_proj,4.19136906,0.01000,3.866 +61,mlp.down_proj,0.18564591,0.01000,23.975 +62,self_attn.k_proj,0.40277070,0.01000,3.085 +62,self_attn.v_proj,0.06261566,0.01000,3.011 +62,self_attn.q_proj,1.05124092,0.01000,3.146 +62,self_attn.o_proj,0.04467377,0.01000,3.154 +62,mlp.up_proj,3.70292091,0.01000,3.801 +62,mlp.gate_proj,4.21846390,0.01000,3.763 +62,mlp.down_proj,0.21794230,0.01000,24.110 +63,self_attn.k_proj,0.39282632,0.01000,3.054 +63,self_attn.v_proj,0.05969462,0.01000,3.039 +63,self_attn.q_proj,1.01897359,0.01000,3.168 +63,self_attn.o_proj,0.03331522,0.01000,3.161 +63,mlp.up_proj,3.71761847,0.01000,3.806 +63,mlp.gate_proj,4.26381588,0.01000,3.763 +63,mlp.down_proj,0.26135504,0.01000,23.999 +64,self_attn.k_proj,0.46097943,0.01000,2.989 +64,self_attn.v_proj,0.07023165,0.01000,2.958 +64,self_attn.q_proj,1.20084310,0.01000,3.151 +64,self_attn.o_proj,0.03521793,0.01000,3.166 +64,mlp.up_proj,3.76330948,0.01000,3.808 +64,mlp.gate_proj,4.30178547,0.01000,3.761 +64,mlp.down_proj,0.24099487,0.01000,24.022 +65,self_attn.k_proj,0.38237023,0.01000,2.965 +65,self_attn.v_proj,0.05751929,0.01000,2.954 +65,self_attn.q_proj,0.86196846,0.01000,3.144 +65,self_attn.o_proj,0.03021962,0.01000,3.161 +65,mlp.up_proj,3.66591787,0.01000,3.795 +65,mlp.gate_proj,4.28166103,0.01000,3.759 +65,mlp.down_proj,0.21255144,0.01000,24.178 +66,self_attn.k_proj,0.37788883,0.01000,3.036 +66,self_attn.v_proj,0.06458911,0.01000,2.957 +66,self_attn.q_proj,0.94858336,0.01000,3.144 +66,self_attn.o_proj,0.03743310,0.01000,3.199 +66,mlp.up_proj,3.74793983,0.01000,3.789 +66,mlp.gate_proj,4.37713146,0.01000,3.758 +66,mlp.down_proj,0.21309337,0.01000,23.970 +67,self_attn.k_proj,0.40760538,0.01000,2.976 +67,self_attn.v_proj,0.07310674,0.01000,2.962 +67,self_attn.q_proj,1.01889122,0.01000,3.151 +67,self_attn.o_proj,0.04029405,0.01000,3.153 +67,mlp.up_proj,3.83434510,0.01000,3.799 +67,mlp.gate_proj,4.50831366,0.01000,3.762 +67,mlp.down_proj,0.22395623,0.01000,23.957 +68,self_attn.k_proj,0.32451558,0.01000,3.018 +68,self_attn.v_proj,0.06332302,0.01000,2.958 +68,self_attn.q_proj,0.70388466,0.01000,3.153 +68,self_attn.o_proj,0.01848056,0.01000,3.199 +68,mlp.up_proj,3.92450428,0.01000,3.797 +68,mlp.gate_proj,4.64310265,0.01000,3.761 +68,mlp.down_proj,0.23298344,0.01000,23.935 +69,self_attn.k_proj,0.28042096,0.01000,2.975 +69,self_attn.v_proj,0.05284975,0.01000,2.948 +69,self_attn.q_proj,0.56274700,0.01000,3.142 +69,self_attn.o_proj,0.02368817,0.01000,3.216 +69,mlp.up_proj,3.98886538,0.01000,3.789 +69,mlp.gate_proj,4.69375992,0.01000,3.752 +69,mlp.down_proj,0.24038924,0.01000,23.938 +70,self_attn.k_proj,0.35089734,0.01000,2.977 +70,self_attn.v_proj,0.07521341,0.01000,2.954 +70,self_attn.q_proj,0.82929134,0.01000,3.145 +70,self_attn.o_proj,0.02283092,0.01000,3.156 +70,mlp.up_proj,4.03632069,0.01000,3.808 +70,mlp.gate_proj,4.70279360,0.01000,3.758 +70,mlp.down_proj,0.23927595,0.01000,24.154 +71,self_attn.k_proj,0.33203787,0.01000,2.992 +71,self_attn.v_proj,0.07004619,0.01000,2.948 +71,self_attn.q_proj,0.71842867,0.01000,3.143 +71,self_attn.o_proj,0.01962153,0.01000,3.155 +71,mlp.up_proj,4.06384087,0.01000,3.797 +71,mlp.gate_proj,4.72325993,0.01000,3.750 +71,mlp.down_proj,0.23815823,0.01000,23.934 +72,self_attn.k_proj,0.23616463,0.01000,2.978 +72,self_attn.v_proj,0.05522804,0.01000,2.949 +72,self_attn.q_proj,0.37876058,0.01000,3.141 +72,self_attn.o_proj,0.00529693,0.01000,3.153 +72,mlp.up_proj,4.12277365,0.01000,3.805 +72,mlp.gate_proj,4.78202343,0.01000,3.755 +72,mlp.down_proj,0.24076617,0.01000,23.961 +73,self_attn.k_proj,0.33402500,0.01000,3.089 +73,self_attn.v_proj,0.07489772,0.01000,3.032 +73,self_attn.q_proj,0.74307632,0.01000,3.200 +73,self_attn.o_proj,0.01953132,0.01000,3.166 +73,mlp.up_proj,4.15922451,0.01000,3.802 +73,mlp.gate_proj,4.81442738,0.01000,3.749 +73,mlp.down_proj,0.24650751,0.01000,24.176 +74,self_attn.k_proj,0.37715226,0.01000,2.973 +74,self_attn.v_proj,0.08905508,0.01000,3.002 +74,self_attn.q_proj,0.91868663,0.01000,3.154 +74,self_attn.o_proj,0.02344134,0.01000,3.177 +74,mlp.up_proj,4.22110844,0.01000,3.811 +74,mlp.gate_proj,4.86965609,0.01000,3.772 +74,mlp.down_proj,0.24981582,0.01000,24.024 +75,self_attn.k_proj,0.30000970,0.01000,3.169 +75,self_attn.v_proj,0.06850379,0.01000,3.127 +75,self_attn.q_proj,0.61720216,0.01000,3.157 +75,self_attn.o_proj,0.01489818,0.01000,3.385 +75,mlp.up_proj,4.25210142,0.01000,3.981 +75,mlp.gate_proj,4.88587141,0.01000,3.767 +75,mlp.down_proj,0.25220692,0.01000,24.003 +76,self_attn.k_proj,0.32943577,0.01000,2.990 +76,self_attn.v_proj,0.07318573,0.01000,2.959 +76,self_attn.q_proj,0.72651207,0.01000,3.153 +76,self_attn.o_proj,0.01490472,0.01000,3.160 +76,mlp.up_proj,4.29693890,0.01000,3.826 +76,mlp.gate_proj,4.92639208,0.01000,3.760 +76,mlp.down_proj,0.25657284,0.01000,23.981 +77,self_attn.k_proj,0.32774276,0.01000,2.990 +77,self_attn.v_proj,0.07041748,0.01000,2.959 +77,self_attn.q_proj,0.67166424,0.01000,3.153 +77,self_attn.o_proj,0.01676882,0.01000,3.157 +77,mlp.up_proj,4.31973410,0.01000,3.791 +77,mlp.gate_proj,4.98892975,0.01000,3.751 +77,mlp.down_proj,0.26171321,0.01000,23.971 +78,self_attn.k_proj,0.33256626,0.01000,2.992 +78,self_attn.v_proj,0.07743990,0.01000,2.965 +78,self_attn.q_proj,0.75554824,0.01000,3.158 +78,self_attn.o_proj,0.02410802,0.01000,3.166 +78,mlp.up_proj,4.33316135,0.01000,3.934 +78,mlp.gate_proj,4.96317434,0.01000,3.884 +78,mlp.down_proj,0.27104366,0.01000,24.696 +79,self_attn.k_proj,0.36951751,0.01000,2.985 +79,self_attn.v_proj,0.07467420,0.01000,2.963 +79,self_attn.q_proj,0.77670276,0.01000,3.212 +79,self_attn.o_proj,0.02118552,0.01000,3.218 +79,mlp.up_proj,4.37143135,0.01000,3.924 +79,mlp.gate_proj,5.07775116,0.01000,3.923 +79,mlp.down_proj,0.31796801,0.01000,24.063 +80,self_attn.k_proj,0.24544287,0.01000,3.003 +80,self_attn.v_proj,0.04570962,0.01000,2.976 +80,self_attn.q_proj,0.31773978,0.01000,3.169 +80,self_attn.o_proj,0.01628411,0.01000,3.171 +80,mlp.up_proj,4.37757206,0.01000,3.815 +80,mlp.gate_proj,5.03681135,0.01000,3.775 +80,mlp.down_proj,0.29214674,0.01000,24.062 +81,self_attn.k_proj,0.22727039,0.01000,2.992 +81,self_attn.v_proj,0.03885635,0.01000,2.967 +81,self_attn.q_proj,0.19104597,0.01000,3.161 +81,self_attn.o_proj,0.00215094,0.01000,3.170 +81,mlp.up_proj,4.38108301,0.01000,3.810 +81,mlp.gate_proj,5.00936747,0.01000,3.772 +81,mlp.down_proj,0.26303947,0.01000,24.147 +82,self_attn.k_proj,0.22519225,0.01000,3.055 +82,self_attn.v_proj,0.03628106,0.01000,3.020 +82,self_attn.q_proj,0.17980447,0.01000,3.224 +82,self_attn.o_proj,0.00217593,0.01000,3.157 +82,mlp.up_proj,4.36705780,0.01000,3.799 +82,mlp.gate_proj,4.96079111,0.01000,3.788 +82,mlp.down_proj,0.24839677,0.01000,23.991 +83,self_attn.k_proj,0.22979859,0.01000,2.979 +83,self_attn.v_proj,0.03165802,0.01000,3.033 +83,self_attn.q_proj,0.14812446,0.01000,3.314 +83,self_attn.o_proj,0.00163801,0.01000,3.165 +83,mlp.up_proj,4.43501663,0.01000,3.800 +83,mlp.gate_proj,5.01992846,0.01000,3.760 +83,mlp.down_proj,0.24432391,0.01000,24.021 +84,self_attn.k_proj,0.23477116,0.01000,2.983 +84,self_attn.v_proj,0.03422052,0.01000,2.952 +84,self_attn.q_proj,0.17082308,0.01000,3.146 +84,self_attn.o_proj,0.00244106,0.01000,3.166 +84,mlp.up_proj,4.52563667,0.01000,3.796 +84,mlp.gate_proj,5.12133408,0.01000,3.758 +84,mlp.down_proj,0.25407451,0.01000,23.972 +85,self_attn.k_proj,0.25626156,0.01000,3.050 +85,self_attn.v_proj,0.04464536,0.01000,3.024 +85,self_attn.q_proj,0.29323554,0.01000,3.210 +85,self_attn.o_proj,0.00449265,0.01000,3.299 +85,mlp.up_proj,4.55304718,0.01000,3.895 +85,mlp.gate_proj,5.14762449,0.01000,3.818 +85,mlp.down_proj,0.24894002,0.01000,24.037 +86,self_attn.k_proj,0.24631341,0.01000,3.094 +86,self_attn.v_proj,0.04632462,0.01000,3.025 +86,self_attn.q_proj,0.29648083,0.01000,3.332 +86,self_attn.o_proj,0.00564192,0.01000,3.183 +86,mlp.up_proj,4.63581562,0.01000,3.803 +86,mlp.gate_proj,5.23101234,0.01000,3.758 +86,mlp.down_proj,0.25253561,0.01000,24.257 +87,self_attn.k_proj,0.25991723,0.01000,2.875 +87,self_attn.v_proj,0.04951914,0.01000,2.837 +87,self_attn.q_proj,0.32328671,0.01000,3.040 +87,self_attn.o_proj,0.00662299,0.01000,3.053 +87,mlp.up_proj,4.71944237,0.01000,3.762 +87,mlp.gate_proj,5.32375669,0.01000,3.669 +87,mlp.down_proj,0.25535131,0.01000,23.879 +88,self_attn.k_proj,0.25833249,0.01000,2.874 +88,self_attn.v_proj,0.04221261,0.01000,2.851 +88,self_attn.q_proj,0.25684297,0.01000,3.048 +88,self_attn.o_proj,0.00373890,0.01000,3.065 +88,mlp.up_proj,4.80329323,0.01000,3.719 +88,mlp.gate_proj,5.41401815,0.01000,3.672 +88,mlp.down_proj,0.26542714,0.01000,23.748 +89,self_attn.k_proj,0.29540002,0.01000,2.876 +89,self_attn.v_proj,0.06226634,0.01000,2.839 +89,self_attn.q_proj,0.51509064,0.01000,3.037 +89,self_attn.o_proj,0.01419941,0.01000,3.154 +89,mlp.up_proj,4.88891220,0.01000,3.696 +89,mlp.gate_proj,5.51603365,0.01000,3.661 +89,mlp.down_proj,0.27464589,0.01000,23.744 +90,self_attn.k_proj,0.30758756,0.01000,2.874 +90,self_attn.v_proj,0.07978672,0.01000,2.847 +90,self_attn.q_proj,0.69067073,0.01000,3.046 +90,self_attn.o_proj,0.01465273,0.01000,3.058 +90,mlp.up_proj,4.95264244,0.01000,3.716 +90,mlp.gate_proj,5.59407043,0.01000,3.664 +90,mlp.down_proj,0.27416804,0.01000,23.639 +91,self_attn.k_proj,0.29949823,0.01000,3.004 +91,self_attn.v_proj,0.07352249,0.01000,2.960 +91,self_attn.q_proj,0.56337321,0.01000,3.151 +91,self_attn.o_proj,0.02519920,0.01000,3.220 +91,mlp.up_proj,5.05726719,0.01000,3.851 +91,mlp.gate_proj,5.72195625,0.01000,3.761 +91,mlp.down_proj,0.28127199,0.01000,23.992 +92,self_attn.k_proj,0.38835877,0.01000,2.883 +92,self_attn.v_proj,0.09978136,0.01000,2.847 +92,self_attn.q_proj,0.96952260,0.01000,3.063 +92,self_attn.o_proj,0.03872257,0.01000,3.094 +92,mlp.up_proj,5.15710735,0.01000,3.772 +92,mlp.gate_proj,5.84712744,0.01000,3.660 +92,mlp.down_proj,0.29408342,0.01000,24.088 +93,self_attn.k_proj,0.31899413,0.01000,3.060 +93,self_attn.v_proj,0.07746388,0.01000,3.027 +93,self_attn.q_proj,0.61558229,0.01000,3.223 +93,self_attn.o_proj,0.01283797,0.01000,3.239 +93,mlp.up_proj,5.25052452,0.01000,3.870 +93,mlp.gate_proj,5.97865391,0.01000,3.823 +93,mlp.down_proj,0.30153495,0.01000,24.014 +94,self_attn.k_proj,0.33564955,0.01000,3.063 +94,self_attn.v_proj,0.07524215,0.01000,3.017 +94,self_attn.q_proj,0.66267246,0.01000,3.215 +94,self_attn.o_proj,0.01723994,0.01000,3.294 +94,mlp.up_proj,5.35795498,0.01000,3.860 +94,mlp.gate_proj,6.13542175,0.01000,3.805 +94,mlp.down_proj,0.31616414,0.01000,23.986 +95,self_attn.k_proj,0.51306427,0.01000,2.886 +95,self_attn.v_proj,0.10710293,0.01000,2.857 +95,self_attn.q_proj,1.44587100,0.01000,3.041 +95,self_attn.o_proj,0.03162030,0.01000,3.066 +95,mlp.up_proj,5.47087049,0.01000,3.708 +95,mlp.gate_proj,6.36106205,0.01000,3.661 +95,mlp.down_proj,0.37017322,0.01000,23.666 +96,self_attn.k_proj,0.25305077,0.01000,2.887 +96,self_attn.v_proj,0.05330018,0.01000,2.854 +96,self_attn.q_proj,0.29774117,0.01000,3.042 +96,self_attn.o_proj,0.00259345,0.01000,3.057 +96,mlp.up_proj,5.38642406,0.01000,3.732 +96,mlp.gate_proj,6.22862244,0.01000,3.671 +96,mlp.down_proj,0.33717293,0.01000,23.631 +97,self_attn.k_proj,0.25711370,0.01000,2.889 +97,self_attn.v_proj,0.04117794,0.01000,2.848 +97,self_attn.q_proj,0.21586779,0.01000,3.046 +97,self_attn.o_proj,0.00133171,0.01000,3.056 +97,mlp.up_proj,5.38330507,0.01000,3.724 +97,mlp.gate_proj,6.17678928,0.01000,3.668 +97,mlp.down_proj,0.30833608,0.01000,23.769 +98,self_attn.k_proj,0.24343455,0.01000,2.890 +98,self_attn.v_proj,0.03673466,0.01000,2.853 +98,self_attn.q_proj,0.19779912,0.01000,3.054 +98,self_attn.o_proj,0.00182172,0.01000,3.074 +98,mlp.up_proj,5.39940548,0.01000,3.725 +98,mlp.gate_proj,6.12171650,0.01000,3.667 +98,mlp.down_proj,0.30475682,0.01000,23.720 +99,self_attn.k_proj,0.24955244,0.01000,2.883 +99,self_attn.v_proj,0.03584138,0.01000,2.849 +99,self_attn.q_proj,0.16865364,0.01000,3.045 +99,self_attn.o_proj,0.00110970,0.01000,3.060 +99,mlp.up_proj,5.36008501,0.01000,3.722 +99,mlp.gate_proj,6.05855083,0.01000,3.688 +99,mlp.down_proj,0.28650719,0.01000,23.686 +100,self_attn.k_proj,0.24380288,0.01000,2.974 +100,self_attn.v_proj,0.03494181,0.01000,2.868 +100,self_attn.q_proj,0.16455841,0.01000,3.153 +100,self_attn.o_proj,0.00118948,0.01000,3.058 +100,mlp.up_proj,5.45997190,0.01000,3.713 +100,mlp.gate_proj,6.15322876,0.01000,3.776 +100,mlp.down_proj,0.29507601,0.01000,24.183 +101,self_attn.k_proj,0.25931284,0.01000,2.887 +101,self_attn.v_proj,0.05811056,0.01000,2.861 +101,self_attn.q_proj,0.33433765,0.01000,3.051 +101,self_attn.o_proj,0.00674569,0.01000,3.064 +101,mlp.up_proj,5.58221626,0.01000,3.911 +101,mlp.gate_proj,6.27407837,0.01000,3.828 +101,mlp.down_proj,0.30314332,0.01000,23.670 +102,self_attn.k_proj,0.24433655,0.01000,2.890 +102,self_attn.v_proj,0.03565801,0.01000,2.867 +102,self_attn.q_proj,0.17172939,0.01000,3.048 +102,self_attn.o_proj,0.00165733,0.01000,3.063 +102,mlp.up_proj,5.65192032,0.01000,3.733 +102,mlp.gate_proj,6.35287809,0.01000,3.673 +102,mlp.down_proj,0.30830440,0.01000,23.678 +103,self_attn.k_proj,0.21869095,0.01000,2.891 +103,self_attn.v_proj,0.03465326,0.01000,2.879 +103,self_attn.q_proj,0.16411979,0.01000,3.101 +103,self_attn.o_proj,0.00228987,0.01000,3.064 +103,mlp.up_proj,5.77248287,0.01000,3.715 +103,mlp.gate_proj,6.47285175,0.01000,3.671 +103,mlp.down_proj,0.31878665,0.01000,23.672 +104,self_attn.k_proj,0.27833816,0.01000,3.098 +104,self_attn.v_proj,0.05653447,0.01000,2.894 +104,self_attn.q_proj,0.33181605,0.01000,3.050 +104,self_attn.o_proj,0.00632260,0.01000,3.065 +104,mlp.up_proj,5.92299604,0.01000,3.898 +104,mlp.gate_proj,6.62810659,0.01000,3.839 +104,mlp.down_proj,0.33463770,0.01000,23.667 +105,self_attn.k_proj,0.23251824,0.01000,2.882 +105,self_attn.v_proj,0.04988948,0.01000,2.853 +105,self_attn.q_proj,0.26777545,0.01000,3.043 +105,self_attn.o_proj,0.00370652,0.01000,3.064 +105,mlp.up_proj,6.04110050,0.01000,3.708 +105,mlp.gate_proj,6.76609421,0.01000,3.669 +105,mlp.down_proj,0.34246147,0.01000,23.766 +106,self_attn.k_proj,0.24649963,0.01000,2.884 +106,self_attn.v_proj,0.05303718,0.01000,2.854 +106,self_attn.q_proj,0.31437218,0.01000,3.051 +106,self_attn.o_proj,0.00818158,0.01000,3.069 +106,mlp.up_proj,6.20051622,0.01000,3.721 +106,mlp.gate_proj,6.91358566,0.01000,3.668 +106,mlp.down_proj,0.35747206,0.01000,23.652 +107,self_attn.k_proj,0.26935774,0.01000,2.873 +107,self_attn.v_proj,0.07399143,0.01000,2.862 +107,self_attn.q_proj,0.52347749,0.01000,3.048 +107,self_attn.o_proj,0.02350343,0.01000,3.063 +107,mlp.up_proj,6.36273432,0.01000,3.738 +107,mlp.gate_proj,7.07827950,0.01000,3.672 +107,mlp.down_proj,0.38489127,0.01000,23.655 +108,self_attn.k_proj,0.31541759,0.01000,2.889 +108,self_attn.v_proj,0.09756853,0.01000,2.856 +108,self_attn.q_proj,0.68302226,0.01000,3.046 +108,self_attn.o_proj,0.02506516,0.01000,3.062 +108,mlp.up_proj,6.56144905,0.01000,3.729 +108,mlp.gate_proj,7.27471352,0.01000,3.669 +108,mlp.down_proj,0.41431051,0.01000,23.899 +109,self_attn.k_proj,0.41287538,0.01000,2.881 +109,self_attn.v_proj,0.11051632,0.01000,2.861 +109,self_attn.q_proj,1.06271303,0.01000,3.062 +109,self_attn.o_proj,0.06111147,0.01000,3.070 +109,mlp.up_proj,6.80175447,0.01000,3.741 +109,mlp.gate_proj,7.53095341,0.01000,3.684 +109,mlp.down_proj,0.43263084,0.01000,23.877 +110,self_attn.k_proj,0.34959465,0.01000,2.888 +110,self_attn.v_proj,0.10869827,0.01000,2.869 +110,self_attn.q_proj,0.78934175,0.01000,3.059 +110,self_attn.o_proj,0.04443160,0.01000,3.062 +110,mlp.up_proj,6.99190807,0.01000,3.776 +110,mlp.gate_proj,7.73158836,0.01000,3.703 +110,mlp.down_proj,0.47574228,0.01000,23.641 +111,self_attn.k_proj,0.48460689,0.01000,3.003 +111,self_attn.v_proj,0.13581827,0.01000,2.976 +111,self_attn.q_proj,1.18839431,0.01000,3.162 +111,self_attn.o_proj,0.05220971,0.01000,3.163 +111,mlp.up_proj,7.29503870,0.01000,3.833 +111,mlp.gate_proj,8.15610886,0.01000,3.762 +111,mlp.down_proj,0.54147720,0.01000,24.167 +112,self_attn.k_proj,0.40064055,0.01000,3.109 +112,self_attn.v_proj,0.13832238,0.01000,3.080 +112,self_attn.q_proj,1.13341451,0.01000,3.240 +112,self_attn.o_proj,0.05223420,0.01000,3.243 +112,mlp.up_proj,7.45650768,0.01000,3.981 +112,mlp.gate_proj,8.27934933,0.01000,3.849 +112,mlp.down_proj,0.57087278,0.01000,24.064 +113,self_attn.k_proj,0.26745471,0.01000,2.999 +113,self_attn.v_proj,0.08530372,0.01000,2.960 +113,self_attn.q_proj,0.57435006,0.01000,3.164 +113,self_attn.o_proj,0.03999488,0.01000,3.160 +113,mlp.up_proj,7.44573402,0.01000,3.903 +113,mlp.gate_proj,8.11815834,0.01000,3.842 +113,mlp.down_proj,0.54559553,0.01000,23.945 +114,self_attn.k_proj,0.27175841,0.01000,2.884 +114,self_attn.v_proj,0.07138506,0.01000,2.866 +114,self_attn.q_proj,0.40150535,0.01000,3.047 +114,self_attn.o_proj,0.04010369,0.01000,3.068 +114,mlp.up_proj,7.53834248,0.01000,3.730 +114,mlp.gate_proj,8.15936279,0.01000,3.669 +114,mlp.down_proj,0.56933641,0.01000,23.684 +115,self_attn.k_proj,0.20326231,0.01000,2.875 +115,self_attn.v_proj,0.05090861,0.01000,2.840 +115,self_attn.q_proj,0.25653815,0.01000,3.038 +115,self_attn.o_proj,0.01262367,0.01000,3.050 +115,mlp.up_proj,7.71780300,0.01000,3.721 +115,mlp.gate_proj,8.29847145,0.01000,3.654 +115,mlp.down_proj,0.61317605,0.01000,23.622 +116,self_attn.k_proj,0.20196617,0.01000,2.986 +116,self_attn.v_proj,0.06781015,0.01000,2.957 +116,self_attn.q_proj,0.40639585,0.01000,3.143 +116,self_attn.o_proj,0.03301121,0.01000,3.163 +116,mlp.up_proj,8.07445526,0.01000,3.790 +116,mlp.gate_proj,8.65225887,0.01000,3.764 +116,mlp.down_proj,0.70528197,0.01000,23.915 +117,self_attn.k_proj,0.26332399,0.01000,2.995 +117,self_attn.v_proj,0.09148645,0.01000,2.953 +117,self_attn.q_proj,0.66883528,0.01000,3.143 +117,self_attn.o_proj,0.07002541,0.01000,3.162 +117,mlp.up_proj,8.46139336,0.01000,3.792 +117,mlp.gate_proj,9.08536243,0.01000,3.743 +117,mlp.down_proj,0.79289079,0.01000,24.390 +118,self_attn.k_proj,0.13599819,0.01000,3.055 +118,self_attn.v_proj,0.05755990,0.01000,3.039 +118,self_attn.q_proj,0.36659241,0.01000,3.212 +118,self_attn.o_proj,0.02626957,0.01000,3.225 +118,mlp.up_proj,8.75907326,0.01000,3.873 +118,mlp.gate_proj,9.43249702,0.01000,3.822 +118,mlp.down_proj,0.91107446,0.01000,24.235 +119,self_attn.k_proj,0.32519650,0.01000,2.987 +119,self_attn.v_proj,0.13961217,0.01000,2.954 +119,self_attn.q_proj,0.89187199,0.01000,3.114 +119,self_attn.o_proj,0.06746665,0.01000,3.127 +119,mlp.up_proj,10.05993176,0.01000,3.720 +119,mlp.gate_proj,11.40488815,0.01000,3.668 +119,mlp.down_proj,1.42022395,0.01000,23.689 +120,self_attn.k_proj,0.39963067,0.01000,2.875 +120,self_attn.v_proj,0.18884216,0.01000,2.840 +120,self_attn.q_proj,1.27754569,0.01000,3.039 +120,self_attn.o_proj,0.17780617,0.01000,3.053 +120,mlp.up_proj,10.65344620,0.01000,3.702 +120,mlp.gate_proj,12.24440670,0.01000,3.661 +120,mlp.down_proj,1.51386595,0.01000,23.693 +121,self_attn.k_proj,0.36563653,0.01000,2.878 +121,self_attn.v_proj,0.16182208,0.01000,2.845 +121,self_attn.q_proj,1.15949345,0.01000,3.046 +121,self_attn.o_proj,0.18820892,0.01000,3.055 +121,mlp.up_proj,11.27324581,0.01000,3.723 +121,mlp.gate_proj,12.97904396,0.01000,3.705 +121,mlp.down_proj,1.76101518,0.01000,23.695 +122,self_attn.k_proj,0.44027224,0.01000,2.893 +122,self_attn.v_proj,0.23480010,0.01000,2.849 +122,self_attn.q_proj,1.58412385,0.01000,3.044 +122,self_attn.o_proj,0.32005700,0.01000,3.068 +122,mlp.up_proj,11.34827137,0.01000,3.706 +122,mlp.gate_proj,13.04171467,0.01000,3.679 +122,mlp.down_proj,2.13088989,0.01000,23.610 +123,self_attn.k_proj,0.53668171,0.01000,2.870 +123,self_attn.v_proj,0.29302818,0.01000,2.841 +123,self_attn.q_proj,2.16953516,0.01000,3.044 +123,self_attn.o_proj,0.67833418,0.01000,3.058 +123,mlp.up_proj,11.48632622,0.01000,3.717 +123,mlp.gate_proj,13.14237595,0.01000,3.678 +123,mlp.down_proj,3.32353497,0.01000,23.707 +124,self_attn.k_proj,0.45076931,0.01000,2.877 +124,self_attn.v_proj,0.20754635,0.01000,2.863 +124,self_attn.q_proj,1.72256780,0.01000,3.061 +124,self_attn.o_proj,0.57295287,0.01000,3.148 +124,mlp.up_proj,10.53095722,0.01000,3.712 +124,mlp.gate_proj,12.15573597,0.01000,3.674 +124,mlp.down_proj,5.73145533,0.01000,23.910 +125,self_attn.k_proj,0.24226147,0.01000,2.879 +125,self_attn.v_proj,0.07440731,0.01000,2.847 +125,self_attn.q_proj,0.75529593,0.01000,3.044 +125,self_attn.o_proj,0.08420333,0.01000,3.052 +125,mlp.up_proj,3.35766482,0.01000,3.773 +125,mlp.gate_proj,3.85698128,0.01000,3.828 +125,mlp.down_proj,0.02208812,0.01250,37.806 diff --git a/quantize_config.json b/quantize_config.json new file mode 100644 index 0000000000000000000000000000000000000000..75ef9703cacc73a52663543f4a1651e69ccf7ace --- /dev/null +++ b/quantize_config.json @@ -0,0 +1,21 @@ +{ + "bits": 4, + "group_size": 128, + "desc_act": true, + "sym": true, + "lm_head": false, + "quant_method": "gptq", + "checkpoint_format": "gptq", + "pack_dtype": "int32", + "meta": { + "quantizer": [ + "gptqmodel:2.2.0" + ], + "uri": "https://github.com/modelcloud/gptqmodel", + "damp_percent": 0.01, + "damp_auto_increment": 0.0025, + "static_groups": false, + "true_sequential": true, + "mse": 0.0 + } +} \ No newline at end of file diff --git a/special_tokens_map.json b/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..344c8261025248cbe380e52f8730a03149d599e1 --- /dev/null +++ b/special_tokens_map.json @@ -0,0 +1,23 @@ +{ + "bos_token": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "eos_token": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/tokenizer.json b/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..6ac83ff885a7c13c549808db25dd230af9d67648 --- /dev/null +++ b/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:384a7e7c676f7be2e5d2e8449c508be9b00e5b18c5b3c39ebc626e96b3f4b988 +size 17210019 diff --git a/tokenizer_config.json b/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..2c7095124f783bd0c1ab751763abd2b544bc5cda --- /dev/null +++ b/tokenizer_config.json @@ -0,0 +1,2067 @@ +{ + "added_tokens_decoder": { + "128000": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128001": { + "content": "<|end_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128002": { + "content": "<|reserved_special_token_0|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128003": { + "content": "<|reserved_special_token_1|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128004": { + "content": "<|finetune_right_pad_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128005": { + "content": "<|reserved_special_token_2|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128006": { + "content": "<|start_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128007": { + "content": "<|end_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128008": { + "content": "<|eom_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128009": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128010": { + "content": "<|python_tag|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128011": { + "content": "<|reserved_special_token_3|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128012": { + "content": "<|reserved_special_token_4|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128013": { + "content": "<|reserved_special_token_5|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128014": { + "content": "<|reserved_special_token_6|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128015": { + "content": "<|reserved_special_token_7|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128016": { + "content": "<|reserved_special_token_8|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128017": { + "content": "<|reserved_special_token_9|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128018": { + "content": "<|reserved_special_token_10|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128019": { + "content": "<|reserved_special_token_11|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128020": { + "content": "<|reserved_special_token_12|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128021": { + "content": "<|reserved_special_token_13|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128022": { + "content": "<|reserved_special_token_14|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128023": { + "content": "<|reserved_special_token_15|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128024": { + "content": "<|reserved_special_token_16|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128025": { + "content": "<|reserved_special_token_17|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128026": { + "content": "<|reserved_special_token_18|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128027": { + "content": "<|reserved_special_token_19|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128028": { + "content": "<|reserved_special_token_20|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128029": { + "content": "<|reserved_special_token_21|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128030": { + "content": "<|reserved_special_token_22|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128031": { + "content": "<|reserved_special_token_23|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128032": { + "content": "<|reserved_special_token_24|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128033": { + "content": "<|reserved_special_token_25|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128034": { + "content": "<|reserved_special_token_26|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128035": { + "content": "<|reserved_special_token_27|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128036": { + "content": "<|reserved_special_token_28|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128037": { + "content": "<|reserved_special_token_29|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128038": { + "content": "<|reserved_special_token_30|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128039": { + "content": "<|reserved_special_token_31|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128040": { + "content": "<|reserved_special_token_32|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128041": { + "content": "<|reserved_special_token_33|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128042": { + "content": "<|reserved_special_token_34|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128043": { + "content": "<|reserved_special_token_35|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128044": { + "content": "<|reserved_special_token_36|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128045": { + "content": "<|reserved_special_token_37|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128046": { + "content": "<|reserved_special_token_38|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128047": { + "content": "<|reserved_special_token_39|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128048": { + "content": "<|reserved_special_token_40|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128049": { + "content": "<|reserved_special_token_41|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128050": { + "content": "<|reserved_special_token_42|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128051": { + "content": "<|reserved_special_token_43|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128052": { + "content": "<|reserved_special_token_44|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128053": { + "content": "<|reserved_special_token_45|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128054": { + "content": "<|reserved_special_token_46|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128055": { + "content": "<|reserved_special_token_47|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128056": { + "content": "<|reserved_special_token_48|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128057": { + "content": "<|reserved_special_token_49|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128058": { + "content": "<|reserved_special_token_50|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128059": { + "content": "<|reserved_special_token_51|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128060": { + "content": "<|reserved_special_token_52|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128061": { + "content": "<|reserved_special_token_53|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128062": { + "content": "<|reserved_special_token_54|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128063": { + "content": "<|reserved_special_token_55|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128064": { + "content": "<|reserved_special_token_56|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128065": { + "content": "<|reserved_special_token_57|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128066": { + "content": "<|reserved_special_token_58|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128067": { + "content": "<|reserved_special_token_59|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128068": { + "content": "<|reserved_special_token_60|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128069": { + "content": "<|reserved_special_token_61|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128070": { + "content": "<|reserved_special_token_62|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128071": { + "content": "<|reserved_special_token_63|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128072": { + "content": "<|reserved_special_token_64|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128073": { + "content": "<|reserved_special_token_65|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128074": { + "content": "<|reserved_special_token_66|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128075": { + "content": "<|reserved_special_token_67|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128076": { + "content": "<|reserved_special_token_68|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128077": { + "content": "<|reserved_special_token_69|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128078": { + "content": "<|reserved_special_token_70|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128079": { + "content": "<|reserved_special_token_71|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128080": { + "content": "<|reserved_special_token_72|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128081": { + "content": "<|reserved_special_token_73|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128082": { + "content": "<|reserved_special_token_74|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128083": { + "content": "<|reserved_special_token_75|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128084": { + "content": "<|reserved_special_token_76|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128085": { + "content": "<|reserved_special_token_77|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128086": { + "content": "<|reserved_special_token_78|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128087": { + "content": "<|reserved_special_token_79|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128088": { + "content": "<|reserved_special_token_80|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128089": { + "content": "<|reserved_special_token_81|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128090": { + "content": "<|reserved_special_token_82|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128091": { + "content": "<|reserved_special_token_83|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128092": { + "content": "<|reserved_special_token_84|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128093": { + "content": "<|reserved_special_token_85|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128094": { + "content": "<|reserved_special_token_86|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128095": { + "content": "<|reserved_special_token_87|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128096": { + "content": "<|reserved_special_token_88|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128097": { + "content": "<|reserved_special_token_89|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128098": { + "content": "<|reserved_special_token_90|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128099": { + "content": "<|reserved_special_token_91|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128100": { + "content": "<|reserved_special_token_92|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128101": { + "content": "<|reserved_special_token_93|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128102": { + "content": "<|reserved_special_token_94|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128103": { + "content": "<|reserved_special_token_95|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128104": { + "content": "<|reserved_special_token_96|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128105": { + "content": "<|reserved_special_token_97|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128106": { + "content": "<|reserved_special_token_98|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128107": { + "content": "<|reserved_special_token_99|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128108": { + "content": "<|reserved_special_token_100|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128109": { + "content": "<|reserved_special_token_101|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128110": { + "content": "<|reserved_special_token_102|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128111": { + "content": "<|reserved_special_token_103|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128112": { + "content": "<|reserved_special_token_104|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128113": { + "content": "<|reserved_special_token_105|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128114": { + "content": "<|reserved_special_token_106|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128115": { + "content": "<|reserved_special_token_107|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128116": { + "content": "<|reserved_special_token_108|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128117": { + "content": "<|reserved_special_token_109|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128118": { + "content": "<|reserved_special_token_110|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128119": { + "content": "<|reserved_special_token_111|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128120": { + "content": "<|reserved_special_token_112|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128121": { + "content": "<|reserved_special_token_113|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128122": { + "content": "<|reserved_special_token_114|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128123": { + "content": "<|reserved_special_token_115|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128124": { + "content": "<|reserved_special_token_116|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128125": { + "content": "<|reserved_special_token_117|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128126": { + "content": "<|reserved_special_token_118|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128127": { + "content": "<|reserved_special_token_119|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128128": { + "content": "<|reserved_special_token_120|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128129": { + "content": "<|reserved_special_token_121|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128130": { + "content": "<|reserved_special_token_122|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128131": { + "content": "<|reserved_special_token_123|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128132": { + "content": "<|reserved_special_token_124|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128133": { + "content": "<|reserved_special_token_125|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128134": { + "content": "<|reserved_special_token_126|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128135": { + "content": "<|reserved_special_token_127|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128136": { + "content": "<|reserved_special_token_128|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128137": { + "content": "<|reserved_special_token_129|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128138": { + "content": "<|reserved_special_token_130|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128139": { + "content": "<|reserved_special_token_131|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128140": { + "content": "<|reserved_special_token_132|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128141": { + "content": "<|reserved_special_token_133|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128142": { + "content": "<|reserved_special_token_134|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128143": { + "content": "<|reserved_special_token_135|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128144": { + "content": "<|reserved_special_token_136|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128145": { + "content": "<|reserved_special_token_137|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128146": { + "content": "<|reserved_special_token_138|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128147": { + "content": "<|reserved_special_token_139|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128148": { + "content": "<|reserved_special_token_140|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128149": { + "content": "<|reserved_special_token_141|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128150": { + "content": "<|reserved_special_token_142|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128151": { + "content": "<|reserved_special_token_143|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128152": { + "content": "<|reserved_special_token_144|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128153": { + "content": "<|reserved_special_token_145|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128154": { + "content": "<|reserved_special_token_146|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128155": { + "content": "<|reserved_special_token_147|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128156": { + "content": "<|reserved_special_token_148|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128157": { + "content": "<|reserved_special_token_149|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128158": { + "content": "<|reserved_special_token_150|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128159": { + "content": "<|reserved_special_token_151|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128160": { + "content": "<|reserved_special_token_152|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128161": { + "content": "<|reserved_special_token_153|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128162": { + "content": "<|reserved_special_token_154|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128163": { + "content": "<|reserved_special_token_155|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128164": { + "content": "<|reserved_special_token_156|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128165": { + "content": "<|reserved_special_token_157|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128166": { + "content": "<|reserved_special_token_158|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128167": { + "content": "<|reserved_special_token_159|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128168": { + "content": "<|reserved_special_token_160|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128169": { + "content": "<|reserved_special_token_161|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128170": { + "content": "<|reserved_special_token_162|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128171": { + "content": "<|reserved_special_token_163|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128172": { + "content": "<|reserved_special_token_164|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128173": { + "content": "<|reserved_special_token_165|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128174": { + "content": "<|reserved_special_token_166|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128175": { + "content": "<|reserved_special_token_167|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128176": { + "content": "<|reserved_special_token_168|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128177": { + "content": "<|reserved_special_token_169|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128178": { + "content": "<|reserved_special_token_170|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128179": { + "content": "<|reserved_special_token_171|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128180": { + "content": "<|reserved_special_token_172|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128181": { + "content": "<|reserved_special_token_173|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128182": { + "content": "<|reserved_special_token_174|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128183": { + "content": "<|reserved_special_token_175|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128184": { + "content": "<|reserved_special_token_176|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128185": { + "content": "<|reserved_special_token_177|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128186": { + "content": "<|reserved_special_token_178|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128187": { + "content": "<|reserved_special_token_179|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128188": { + "content": "<|reserved_special_token_180|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128189": { + "content": "<|reserved_special_token_181|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128190": { + "content": "<|reserved_special_token_182|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128191": { + "content": "<|reserved_special_token_183|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128192": { + "content": "<|reserved_special_token_184|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128193": { + "content": "<|reserved_special_token_185|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128194": { + "content": "<|reserved_special_token_186|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128195": { + "content": "<|reserved_special_token_187|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128196": { + "content": "<|reserved_special_token_188|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128197": { + "content": "<|reserved_special_token_189|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128198": { + "content": "<|reserved_special_token_190|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128199": { + "content": "<|reserved_special_token_191|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128200": { + "content": "<|reserved_special_token_192|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128201": { + "content": "<|reserved_special_token_193|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128202": { + "content": "<|reserved_special_token_194|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128203": { + "content": "<|reserved_special_token_195|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128204": { + "content": "<|reserved_special_token_196|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128205": { + "content": "<|reserved_special_token_197|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128206": { + "content": "<|reserved_special_token_198|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128207": { + "content": "<|reserved_special_token_199|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128208": { + "content": "<|reserved_special_token_200|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128209": { + "content": "<|reserved_special_token_201|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128210": { + "content": "<|reserved_special_token_202|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128211": { + "content": "<|reserved_special_token_203|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128212": { + "content": "<|reserved_special_token_204|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128213": { + "content": "<|reserved_special_token_205|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128214": { + "content": "<|reserved_special_token_206|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128215": { + "content": "<|reserved_special_token_207|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128216": { + "content": "<|reserved_special_token_208|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128217": { + "content": "<|reserved_special_token_209|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128218": { + "content": "<|reserved_special_token_210|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128219": { + "content": "<|reserved_special_token_211|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128220": { + "content": "<|reserved_special_token_212|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128221": { + "content": "<|reserved_special_token_213|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128222": { + "content": "<|reserved_special_token_214|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128223": { + "content": "<|reserved_special_token_215|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128224": { + "content": "<|reserved_special_token_216|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128225": { + "content": "<|reserved_special_token_217|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128226": { + "content": "<|reserved_special_token_218|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128227": { + "content": "<|reserved_special_token_219|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128228": { + "content": "<|reserved_special_token_220|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128229": { + "content": "<|reserved_special_token_221|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128230": { + "content": "<|reserved_special_token_222|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128231": { + "content": "<|reserved_special_token_223|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128232": { + "content": "<|reserved_special_token_224|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128233": { + "content": "<|reserved_special_token_225|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128234": { + "content": "<|reserved_special_token_226|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128235": { + "content": "<|reserved_special_token_227|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128236": { + "content": "<|reserved_special_token_228|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128237": { + "content": "<|reserved_special_token_229|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128238": { + "content": "<|reserved_special_token_230|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128239": { + "content": "<|reserved_special_token_231|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128240": { + "content": "<|reserved_special_token_232|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128241": { + "content": "<|reserved_special_token_233|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128242": { + "content": "<|reserved_special_token_234|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128243": { + "content": "<|reserved_special_token_235|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128244": { + "content": "<|reserved_special_token_236|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128245": { + "content": "<|reserved_special_token_237|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128246": { + "content": "<|reserved_special_token_238|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128247": { + "content": "<|reserved_special_token_239|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128248": { + "content": "<|reserved_special_token_240|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128249": { + "content": "<|reserved_special_token_241|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128250": { + "content": "<|reserved_special_token_242|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128251": { + "content": "<|reserved_special_token_243|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128252": { + "content": "<|reserved_special_token_244|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128253": { + "content": "<|reserved_special_token_245|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128254": { + "content": "<|reserved_special_token_246|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128255": { + "content": "<|reserved_special_token_247|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + } + }, + "bos_token": "<|begin_of_text|>", + "clean_up_tokenization_spaces": true, + "eos_token": "<|eot_id|>", + "extra_special_tokens": {}, + "max_length": 4096, + "model_input_names": [ + "input_ids", + "attention_mask" + ], + "model_max_length": 131072, + "pad_token": "<|eot_id|>", + "stride": 0, + "tokenizer_class": "PreTrainedTokenizer", + "truncation_side": "right", + "truncation_strategy": "longest_first" +}