yujiepan commited on
Commit
1bec50b
·
verified ·
1 Parent(s): 93c8028

Upload folder using huggingface_hub

Browse files
Files changed (2) hide show
  1. README.md +2 -1
  2. config.json +1 -0
README.md CHANGED
@@ -17,7 +17,7 @@ This tiny model is for debugging. It is randomly initialized with the config ada
17
  - vLLM
18
 
19
  ```bash
20
- vllm serve tiny-random/kimi-k2
21
  ```
22
 
23
  - Transformers
@@ -94,6 +94,7 @@ config_json.update({
94
  'v_head_dim': 64,
95
  'tie_word_embeddings': False,
96
  })
 
97
  del config_json['quantization_config']
98
  with open(f"{save_folder}/config.json", "w", encoding='utf-8') as f:
99
  json.dump(config_json, f, indent=2)
 
17
  - vLLM
18
 
19
  ```bash
20
+ vllm serve tiny-random/kimi-k2 --trust-remote-code
21
  ```
22
 
23
  - Transformers
 
94
  'v_head_dim': 64,
95
  'tie_word_embeddings': False,
96
  })
97
+ config_json['rope_scaling']['rope_type'] = 'yarn'
98
  del config_json['quantization_config']
99
  with open(f"{save_folder}/config.json", "w", encoding='utf-8') as f:
100
  json.dump(config_json, f, indent=2)
config.json CHANGED
@@ -44,6 +44,7 @@
44
  "mscale": 1.0,
45
  "mscale_all_dim": 1.0,
46
  "original_max_position_embeddings": 4096,
 
47
  "type": "yarn"
48
  },
49
  "rope_theta": 50000.0,
 
44
  "mscale": 1.0,
45
  "mscale_all_dim": 1.0,
46
  "original_max_position_embeddings": 4096,
47
+ "rope_type": "yarn",
48
  "type": "yarn"
49
  },
50
  "rope_theta": 50000.0,