alvarobartt HF Staff commited on
Commit
8133ef0
·
verified ·
1 Parent(s): e753abb

Remove `auto_map` from `config.json`

Browse files

Jamba native support in Transformers was introduced upon https://github.com/huggingface/transformers/releases/tag/v4.40.0, so `auto_map` not required as it's pointing to remote code at https://huggingface.co/ai21labs/Jamba-v0.1 which can be a security risk as it requires the Transformers pipeline to add `trust_remote_code=True`

Files changed (1) hide show
  1. config.json +0 -6
config.json CHANGED
@@ -6,12 +6,6 @@
6
  "attention_dropout": 0.0,
7
  "attn_layer_offset": 4,
8
  "attn_layer_period": 8,
9
- "auto_map": {
10
- "AutoConfig": "ai21labs/Jamba-v0.1--configuration_jamba.JambaConfig",
11
- "AutoModel": "ai21labs/Jamba-v0.1--modeling_jamba.JambaModel",
12
- "AutoModelForCausalLM": "ai21labs/Jamba-v0.1--modeling_jamba.JambaForCausalLM",
13
- "AutoModelForSequenceClassification": "ai21labs/Jamba-v0.1--model.JambaForSequenceClassification"
14
- },
15
  "bos_token_id": 29,
16
  "eos_token_id": 27,
17
  "expert_layer_offset": 1,
 
6
  "attention_dropout": 0.0,
7
  "attn_layer_offset": 4,
8
  "attn_layer_period": 8,
 
 
 
 
 
 
9
  "bos_token_id": 29,
10
  "eos_token_id": 27,
11
  "expert_layer_offset": 1,