Upload folder using huggingface_hub
Browse files- README.md +2 -1
- config.json +1 -0
README.md
CHANGED
@@ -17,7 +17,7 @@ This tiny model is for debugging. It is randomly initialized with the config ada
|
|
17 |
- vLLM
|
18 |
|
19 |
```bash
|
20 |
-
vllm serve tiny-random/kimi-k2
|
21 |
```
|
22 |
|
23 |
- Transformers
|
@@ -94,6 +94,7 @@ config_json.update({
|
|
94 |
'v_head_dim': 64,
|
95 |
'tie_word_embeddings': False,
|
96 |
})
|
|
|
97 |
del config_json['quantization_config']
|
98 |
with open(f"{save_folder}/config.json", "w", encoding='utf-8') as f:
|
99 |
json.dump(config_json, f, indent=2)
|
|
|
17 |
- vLLM
|
18 |
|
19 |
```bash
|
20 |
+
vllm serve tiny-random/kimi-k2 --trust-remote-code
|
21 |
```
|
22 |
|
23 |
- Transformers
|
|
|
94 |
'v_head_dim': 64,
|
95 |
'tie_word_embeddings': False,
|
96 |
})
|
97 |
+
config_json['rope_scaling']['rope_type'] = 'yarn'
|
98 |
del config_json['quantization_config']
|
99 |
with open(f"{save_folder}/config.json", "w", encoding='utf-8') as f:
|
100 |
json.dump(config_json, f, indent=2)
|
config.json
CHANGED
@@ -44,6 +44,7 @@
|
|
44 |
"mscale": 1.0,
|
45 |
"mscale_all_dim": 1.0,
|
46 |
"original_max_position_embeddings": 4096,
|
|
|
47 |
"type": "yarn"
|
48 |
},
|
49 |
"rope_theta": 50000.0,
|
|
|
44 |
"mscale": 1.0,
|
45 |
"mscale_all_dim": 1.0,
|
46 |
"original_max_position_embeddings": 4096,
|
47 |
+
"rope_type": "yarn",
|
48 |
"type": "yarn"
|
49 |
},
|
50 |
"rope_theta": 50000.0,
|