danielhanchen commited on
Commit
07571c6
·
verified ·
1 Parent(s): 32c20b1

Add files using upload-large-folder tool

Browse files
Qwen3-30B-A3B-UD-IQ1_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bf4f5a91caa6f42d3f61ea0378fddce6790cc7971fc4f30f04b65e090f28f6b7
3
- size 9021281120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef75cdd5f6bad30e114cc2f780953f55cde9b02934fe8f8f7c7b3083cfcc44fe
3
+ size 9021280608
Qwen3-30B-A3B-UD-IQ2_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8074036a4a2b95ac775b2218676486a66297c988f0534f89beef84e03227f1d3
3
- size 10865578848
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70e34f3ec8c79af04e35e8acbd2ab15ad1cf7a096d00513a1c1afd40d4863f9c
3
+ size 10865578336
Qwen3-30B-A3B-UD-Q2_K_XL.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c8cd2fc3ce4b41b72174c74351fcbbea02a289fd9efaf8e5c8a2c9c3e18fbc6c
3
- size 11814277984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f6e67b99c35a0da407034fd49e69b64e187d01ef393e303f38187f10401c09d
3
+ size 11814277472
Qwen3-30B-A3B-UD-Q4_K_XL.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:609302337c6e2b0619d6a2aa0c54bc9cdebe2302d8ae5cdb8cfe33310c0124f3
3
- size 17715663712
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac5fe037006087ea052c4dcf4398f8cdd25effed176a67682e75402bb5f7e7aa
3
+ size 17715663200
config.json CHANGED
@@ -4,7 +4,6 @@
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
- "bos_token_id": 151643,
8
  "decoder_sparse_step": 1,
9
  "eos_token_id": 151645,
10
  "head_dim": 128,
@@ -24,6 +23,7 @@
24
  "num_hidden_layers": 48,
25
  "num_key_value_heads": 4,
26
  "output_router_logits": false,
 
27
  "rms_norm_eps": 1e-06,
28
  "rope_scaling": null,
29
  "rope_theta": 1000000.0,
@@ -31,8 +31,9 @@
31
  "sliding_window": null,
32
  "tie_word_embeddings": false,
33
  "torch_dtype": "bfloat16",
34
- "transformers_version": "4.51.0",
 
35
  "use_cache": true,
36
  "use_sliding_window": false,
37
  "vocab_size": 151936
38
- }
 
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
 
7
  "decoder_sparse_step": 1,
8
  "eos_token_id": 151645,
9
  "head_dim": 128,
 
23
  "num_hidden_layers": 48,
24
  "num_key_value_heads": 4,
25
  "output_router_logits": false,
26
+ "pad_token_id": 151654,
27
  "rms_norm_eps": 1e-06,
28
  "rope_scaling": null,
29
  "rope_theta": 1000000.0,
 
31
  "sliding_window": null,
32
  "tie_word_embeddings": false,
33
  "torch_dtype": "bfloat16",
34
+ "transformers_version": "4.52.0.dev0",
35
+ "unsloth_fixed": true,
36
  "use_cache": true,
37
  "use_sliding_window": false,
38
  "vocab_size": 151936
39
+ }