Upload folder using huggingface_hub
Browse files- config.json +38 -0
- generation_config.json +13 -0
- model-00001-of-00013.safetensors +3 -0
- model-00002-of-00013.safetensors +3 -0
- model-00003-of-00013.safetensors +3 -0
- model-00004-of-00013.safetensors +3 -0
- model-00005-of-00013.safetensors +3 -0
- model-00006-of-00013.safetensors +3 -0
- model-00007-of-00013.safetensors +3 -0
- model-00008-of-00013.safetensors +3 -0
- model-00009-of-00013.safetensors +3 -0
- model-00010-of-00013.safetensors +3 -0
- model-00011-of-00013.safetensors +3 -0
- model-00012-of-00013.safetensors +3 -0
- model-00013-of-00013.safetensors +3 -0
- model.safetensors.index.json +0 -0
config.json
ADDED
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"Qwen3MoeForCausalLM"
|
4 |
+
],
|
5 |
+
"attention_bias": false,
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 151643,
|
8 |
+
"decoder_sparse_step": 1,
|
9 |
+
"eos_token_id": 151645,
|
10 |
+
"head_dim": 128,
|
11 |
+
"hidden_act": "silu",
|
12 |
+
"hidden_size": 2048,
|
13 |
+
"initializer_range": 0.02,
|
14 |
+
"intermediate_size": 6144,
|
15 |
+
"max_position_embeddings": 262144,
|
16 |
+
"max_window_layers": 48,
|
17 |
+
"mlp_only_layers": [],
|
18 |
+
"model_type": "qwen3_moe",
|
19 |
+
"moe_intermediate_size": 768,
|
20 |
+
"norm_topk_prob": true,
|
21 |
+
"num_attention_heads": 32,
|
22 |
+
"num_experts": 128,
|
23 |
+
"num_experts_per_tok": 8,
|
24 |
+
"num_hidden_layers": 48,
|
25 |
+
"num_key_value_heads": 4,
|
26 |
+
"output_router_logits": false,
|
27 |
+
"rms_norm_eps": 1e-06,
|
28 |
+
"rope_scaling": null,
|
29 |
+
"rope_theta": 10000000,
|
30 |
+
"router_aux_loss_coef": 0.001,
|
31 |
+
"sliding_window": null,
|
32 |
+
"tie_word_embeddings": false,
|
33 |
+
"torch_dtype": "bfloat16",
|
34 |
+
"transformers_version": "4.51.3",
|
35 |
+
"use_cache": true,
|
36 |
+
"use_sliding_window": false,
|
37 |
+
"vocab_size": 151936
|
38 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token_id": 151643,
|
3 |
+
"do_sample": true,
|
4 |
+
"eos_token_id": [
|
5 |
+
151645,
|
6 |
+
151643
|
7 |
+
],
|
8 |
+
"pad_token_id": 151643,
|
9 |
+
"temperature": 0.7,
|
10 |
+
"top_k": 20,
|
11 |
+
"top_p": 0.8,
|
12 |
+
"transformers_version": "4.51.3"
|
13 |
+
}
|
model-00001-of-00013.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:96d9164864228161953a3bdf8550c5b06a3517e35f60f29aade2a5cf43167f49
|
3 |
+
size 4997184968
|
model-00002-of-00013.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:35c8a68d6df191ee28426d7f74db964f477872dbcd9fa1837494cf423431bacc
|
3 |
+
size 4997741608
|
model-00003-of-00013.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f7277c090bdf132edf7574610d51fb950cf1a5f8d5bd093389c9933446c6b4c6
|
3 |
+
size 4997742208
|
model-00004-of-00013.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a675a0f70573c565e4257747ed45608f4fc983062fead6ef02ce97c8f680d71b
|
3 |
+
size 4997743184
|
model-00005-of-00013.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f3a9ea6fece9760c059f1a374122e03dfcac619305d51f3e33f2cd11dfbf571e
|
3 |
+
size 4997743184
|
model-00006-of-00013.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eb8821b85f771b45293d8a7eb81ce274a80484baba7c3c482fa528fcc1fd0309
|
3 |
+
size 4997743184
|
model-00007-of-00013.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4024e1e545b3bec3212078e69f1dccb0d6c2f6fd0f71b2c2cd0b50881c3393a4
|
3 |
+
size 4997743184
|
model-00008-of-00013.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:602dc38a979aff898fdf03735be53a93bd5c4fecd1850afe0f6ec27bcc9dc76b
|
3 |
+
size 4997743184
|
model-00009-of-00013.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:db5acc43c93133236fe4072a633f0df9adb26793b9645bd8af290c02f11fd513
|
3 |
+
size 4997743184
|
model-00010-of-00013.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ff6e6e4f80f528330dd3d7b13ad8fdf8f08d6a314360bb4e6d306b5dacb981c2
|
3 |
+
size 4997743184
|
model-00011-of-00013.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:917ae77c78ac479104acab51a044f12a24338ac650ceccf27e5dd0d8c5e1ed6c
|
3 |
+
size 4997743184
|
model-00012-of-00013.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:50d5bc67d9d58fee880f11ff94bc4199fd85363947d59a5d59d9b94d507e68d5
|
3 |
+
size 4997743184
|
model-00013-of-00013.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:75f065d707a79540b448f2f5a869e2d803c33ac2a6fe6a62163698ab53628696
|
3 |
+
size 1094220288
|
model.safetensors.index.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|