onebitquantized commited on
Commit
70b86ed
1 Parent(s): b3c2e3a

Upload of AutoGPTQ quantized model

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "meta-llama/Meta-Llama-3.1-405B-Instruct",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -25,14 +25,14 @@
25
  "quantization_config": {
26
  "bits": 4,
27
  "checkpoint_format": "gptq",
28
- "damp_percent": 0.01,
29
  "desc_act": true,
30
  "exponent_hinv": 4.0,
31
- "group_size": -1,
32
  "model_file_base_name": null,
33
  "model_name_or_path": null,
34
  "quant_method": "gptq",
35
- "shrink": 0.1,
36
  "static_groups": false,
37
  "sym": false,
38
  "true_sequential": true
 
1
  {
2
+ "_name_or_path": "meta-llama/Llama-3.1-405B-Instruct",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
25
  "quantization_config": {
26
  "bits": 4,
27
  "checkpoint_format": "gptq",
28
+ "damp_percent": 0.025,
29
  "desc_act": true,
30
  "exponent_hinv": 4.0,
31
+ "group_size": 128,
32
  "model_file_base_name": null,
33
  "model_name_or_path": null,
34
  "quant_method": "gptq",
35
+ "shrink": 0.0625,
36
  "static_groups": false,
37
  "sym": false,
38
  "true_sequential": true
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "weight_map": {
3
- "lm_head.weight": "model_000.safetensors",
4
  "model.embed_tokens.weight": "model_000.safetensors",
5
  "model.layers.0.input_layernorm.weight": "model_000.safetensors",
6
  "model.layers.0.mlp.down_proj.g_idx": "model_000.safetensors",
 
1
  {
2
  "weight_map": {
3
+ "lm_head.weight": "model_001.safetensors",
4
  "model.embed_tokens.weight": "model_000.safetensors",
5
  "model.layers.0.input_layernorm.weight": "model_000.safetensors",
6
  "model.layers.0.mlp.down_proj.g_idx": "model_000.safetensors",
model_000.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4b2c0207cc9b3c5ee37655efe94b64822fecc359a6294b60426a3f6fdd628e4f
3
- size 48714504984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8423bb31b093ec0f90a8c173c40dda3ca0fbb48b62bfa6c5e56633f604bfb4ce
3
+ size 46075435392
model_001.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6a9fc3a3f483c0ba6adcec77c5fc97aaa99fe23a6f6e547edf2cad356d9feb31
3
- size 40745331064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61c72c3359917c910414718b9c9af81956e2303a513850fab3f411d657181fed
3
+ size 46514935960
model_002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7d43728978c9d83ecb5f04bfcc2acc7e5f625389ae38fa459dca4cb1f6cafe41
3
- size 39881286320
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4e434399cfeb463e1aa5edcdecc47c3ed81299d854c1102e032ad4da613bee5
3
+ size 41441761136
model_003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f5d816f7b7760b92be06ea378eaf525202df12cf7a8195a7082986d535458829
3
- size 40141207768
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c37fb8f1f302056aaabd8f84d6f7bee18ebd4c50c6642355b600a17b8d169f37
3
+ size 41698236320
model_004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ced7c1739f523f6ad4e8fb39666cc5014cebd7b1015a57a38430e857d6ed4b7b
3
- size 39881114056
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f616abec4fcabe795c5d26af181b3d7254fc3a2e174a902f803e7babffdc05c9
3
+ size 41428063888
quantize_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "bits": 4,
3
- "group_size": -1,
4
- "damp_percent": 0.01,
5
  "desc_act": true,
6
  "static_groups": false,
7
  "sym": false,
@@ -9,7 +9,7 @@
9
  "model_name_or_path": null,
10
  "model_file_base_name": null,
11
  "exponent_hinv": 4.0,
12
- "shrink": 0.1,
13
- "quant_method": "xmad",
14
  "checkpoint_format": "gptq"
15
  }
 
1
  {
2
  "bits": 4,
3
+ "group_size": 128,
4
+ "damp_percent": 0.025,
5
  "desc_act": true,
6
  "static_groups": false,
7
  "sym": false,
 
9
  "model_name_or_path": null,
10
  "model_file_base_name": null,
11
  "exponent_hinv": 4.0,
12
+ "shrink": 0.0625,
13
+ "quant_method": "gptq",
14
  "checkpoint_format": "gptq"
15
  }