onebitquantized
commited on
Commit
•
70b86ed
1
Parent(s):
b3c2e3a
Upload of AutoGPTQ quantized model
Browse files- config.json +4 -4
- model.safetensors.index.json +1 -1
- model_000.safetensors +2 -2
- model_001.safetensors +2 -2
- model_002.safetensors +2 -2
- model_003.safetensors +2 -2
- model_004.safetensors +2 -2
- quantize_config.json +4 -4
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "meta-llama/
|
3 |
"architectures": [
|
4 |
"LlamaForCausalLM"
|
5 |
],
|
@@ -25,14 +25,14 @@
|
|
25 |
"quantization_config": {
|
26 |
"bits": 4,
|
27 |
"checkpoint_format": "gptq",
|
28 |
-
"damp_percent": 0.
|
29 |
"desc_act": true,
|
30 |
"exponent_hinv": 4.0,
|
31 |
-
"group_size":
|
32 |
"model_file_base_name": null,
|
33 |
"model_name_or_path": null,
|
34 |
"quant_method": "gptq",
|
35 |
-
"shrink": 0.
|
36 |
"static_groups": false,
|
37 |
"sym": false,
|
38 |
"true_sequential": true
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "meta-llama/Llama-3.1-405B-Instruct",
|
3 |
"architectures": [
|
4 |
"LlamaForCausalLM"
|
5 |
],
|
|
|
25 |
"quantization_config": {
|
26 |
"bits": 4,
|
27 |
"checkpoint_format": "gptq",
|
28 |
+
"damp_percent": 0.025,
|
29 |
"desc_act": true,
|
30 |
"exponent_hinv": 4.0,
|
31 |
+
"group_size": 128,
|
32 |
"model_file_base_name": null,
|
33 |
"model_name_or_path": null,
|
34 |
"quant_method": "gptq",
|
35 |
+
"shrink": 0.0625,
|
36 |
"static_groups": false,
|
37 |
"sym": false,
|
38 |
"true_sequential": true
|
model.safetensors.index.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"weight_map": {
|
3 |
-
"lm_head.weight": "
|
4 |
"model.embed_tokens.weight": "model_000.safetensors",
|
5 |
"model.layers.0.input_layernorm.weight": "model_000.safetensors",
|
6 |
"model.layers.0.mlp.down_proj.g_idx": "model_000.safetensors",
|
|
|
1 |
{
|
2 |
"weight_map": {
|
3 |
+
"lm_head.weight": "model_001.safetensors",
|
4 |
"model.embed_tokens.weight": "model_000.safetensors",
|
5 |
"model.layers.0.input_layernorm.weight": "model_000.safetensors",
|
6 |
"model.layers.0.mlp.down_proj.g_idx": "model_000.safetensors",
|
model_000.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8423bb31b093ec0f90a8c173c40dda3ca0fbb48b62bfa6c5e56633f604bfb4ce
|
3 |
+
size 46075435392
|
model_001.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:61c72c3359917c910414718b9c9af81956e2303a513850fab3f411d657181fed
|
3 |
+
size 46514935960
|
model_002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d4e434399cfeb463e1aa5edcdecc47c3ed81299d854c1102e032ad4da613bee5
|
3 |
+
size 41441761136
|
model_003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c37fb8f1f302056aaabd8f84d6f7bee18ebd4c50c6642355b600a17b8d169f37
|
3 |
+
size 41698236320
|
model_004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f616abec4fcabe795c5d26af181b3d7254fc3a2e174a902f803e7babffdc05c9
|
3 |
+
size 41428063888
|
quantize_config.json
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
{
|
2 |
"bits": 4,
|
3 |
-
"group_size":
|
4 |
-
"damp_percent": 0.
|
5 |
"desc_act": true,
|
6 |
"static_groups": false,
|
7 |
"sym": false,
|
@@ -9,7 +9,7 @@
|
|
9 |
"model_name_or_path": null,
|
10 |
"model_file_base_name": null,
|
11 |
"exponent_hinv": 4.0,
|
12 |
-
"shrink": 0.
|
13 |
-
"quant_method": "
|
14 |
"checkpoint_format": "gptq"
|
15 |
}
|
|
|
1 |
{
|
2 |
"bits": 4,
|
3 |
+
"group_size": 128,
|
4 |
+
"damp_percent": 0.025,
|
5 |
"desc_act": true,
|
6 |
"static_groups": false,
|
7 |
"sym": false,
|
|
|
9 |
"model_name_or_path": null,
|
10 |
"model_file_base_name": null,
|
11 |
"exponent_hinv": 4.0,
|
12 |
+
"shrink": 0.0625,
|
13 |
+
"quant_method": "gptq",
|
14 |
"checkpoint_format": "gptq"
|
15 |
}
|