Transformers
PyTorch
esm
airkingbd commited on
Commit
7a7e651
·
1 Parent(s): a743f1a

add "dplm_type" to config

Browse files
Files changed (1) hide show
  1. config.json +1 -1
config.json CHANGED
@@ -5,6 +5,7 @@
5
  ],
6
  "attention_probs_dropout_prob": 0.0,
7
  "classifier_dropout": null,
 
8
  "emb_layer_norm_before": false,
9
  "esmfold_config": null,
10
  "hidden_act": "gelu",
@@ -17,7 +18,6 @@
17
  "mask_token_id": 32,
18
  "max_position_embeddings": 1026,
19
  "model_type": "esm",
20
- "dplm_type": "dplm_esm",
21
  "num_attention_heads": 20,
22
  "num_hidden_layers": 33,
23
  "pad_token_id": 1,
 
5
  ],
6
  "attention_probs_dropout_prob": 0.0,
7
  "classifier_dropout": null,
8
+ "dplm_type": "dplm_esm",
9
  "emb_layer_norm_before": false,
10
  "esmfold_config": null,
11
  "hidden_act": "gelu",
 
18
  "mask_token_id": 32,
19
  "max_position_embeddings": 1026,
20
  "model_type": "esm",
 
21
  "num_attention_heads": 20,
22
  "num_hidden_layers": 33,
23
  "pad_token_id": 1,