Xenova HF Staff commited on
Commit
3572710
·
verified ·
1 Parent(s): 67c08d6

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +4 -4
config.json CHANGED
@@ -4,10 +4,6 @@
4
  "Phi3ForCausalLM"
5
  ],
6
  "attention_dropout": 0.0,
7
- "auto_map": {
8
- "AutoConfig": "configuration_phi3.Phi3Config",
9
- "AutoModelForCausalLM": "modeling_phi3.Phi3ForCausalLM"
10
- },
11
  "bos_token_id": 1,
12
  "embd_pdrop": 0.0,
13
  "eos_token_id": 32000,
@@ -30,6 +26,10 @@
30
  "tie_word_embeddings": false,
31
  "torch_dtype": "bfloat16",
32
  "transformers_version": "4.39.3",
 
 
 
 
33
  "use_cache": true,
34
  "vocab_size": 32064
35
  }
 
4
  "Phi3ForCausalLM"
5
  ],
6
  "attention_dropout": 0.0,
 
 
 
 
7
  "bos_token_id": 1,
8
  "embd_pdrop": 0.0,
9
  "eos_token_id": 32000,
 
26
  "tie_word_embeddings": false,
27
  "torch_dtype": "bfloat16",
28
  "transformers_version": "4.39.3",
29
+ "transformers.js_config": {
30
+ "dtype": "q4",
31
+ "use_external_data_format": true
32
+ },
33
  "use_cache": true,
34
  "vocab_size": 32064
35
  }