yangheng commited on
Commit
3b26205
·
verified ·
1 Parent(s): 7018149

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +12 -14
config.json CHANGED
@@ -1,28 +1,27 @@
1
  {
2
  "OmniGenomefold_config": null,
3
  "_name_or_path": "./",
4
- "architectures": [
5
- "OmniGenomeModel",
6
- "OmniGenomeForTokenClassification",
7
- "OmniGenomeForMaskedLM",
8
- "OmniGenomeModelForSeq2SeqLM",
9
- "OmniGenomeForTSequenceClassification",
10
- "OmniGenomeForTokenClassification",
11
- "OmniGenomeForSeq2SeqLM"
12
- ],
13
  "attention_probs_dropout_prob": 0.0,
 
 
 
 
 
 
 
 
14
  "classifier_dropout": null,
15
  "emb_layer_norm_before": false,
16
  "hidden_act": "gelu",
17
  "hidden_dropout_prob": 0,
18
- "hidden_size": 480,
19
  "id2label": {
20
  "0": "(",
21
  "1": ")",
22
  "2": "."
23
  },
24
  "initializer_range": 0.02,
25
- "intermediate_size": 2400,
26
  "is_folding_model": false,
27
  "label2id": {
28
  "(": 0,
@@ -33,9 +32,9 @@
33
  "mask_token_id": 23,
34
  "max_position_embeddings": 1026,
35
  "model_type": "omnigenome",
36
- "num_attention_heads": 24,
 
37
  "num_generation": 50,
38
- "num_hidden_layers": 16,
39
  "num_population": 100,
40
  "pad_token_id": 1,
41
  "position_embedding_type": "rotary",
@@ -43,7 +42,6 @@
43
  "torch_dtype": "float32",
44
  "transformers_version": "4.41.0.dev0",
45
  "use_cache": true,
46
- "verify_ss": true,
47
  "vocab_list": null,
48
  "vocab_size": 24
49
  }
 
1
  {
2
  "OmniGenomefold_config": null,
3
  "_name_or_path": "./",
 
 
 
 
 
 
 
 
 
4
  "attention_probs_dropout_prob": 0.0,
5
+ "auto_map": {
6
+ "AutoConfig": "configuration_omnigenome.OmniGenomeConfig",
7
+ "AutoModel": "modeling_omnigenome.OmniGenomeModel",
8
+ "AutoModelForMaskedLM": "modeling_omnigenome.OmniGenomeForMaskedLM",
9
+ "AutoModelForSeq2SeqLM": "modeling_omnigenome.OmniGenomeForSeq2SeqLM",
10
+ "AutoModelForSequenceClassification": "modeling_omnigenome.OmniGenomeForSequenceClassification",
11
+ "AutoModelForTokenClassification": "modeling_omnigenome.OmniGenomeForTokenClassification"
12
+ },
13
  "classifier_dropout": null,
14
  "emb_layer_norm_before": false,
15
  "hidden_act": "gelu",
16
  "hidden_dropout_prob": 0,
17
+ "hidden_size": 720,
18
  "id2label": {
19
  "0": "(",
20
  "1": ")",
21
  "2": "."
22
  },
23
  "initializer_range": 0.02,
24
+ "intermediate_size": 2560,
25
  "is_folding_model": false,
26
  "label2id": {
27
  "(": 0,
 
32
  "mask_token_id": 23,
33
  "max_position_embeddings": 1026,
34
  "model_type": "omnigenome",
35
+ "num_attention_heads": 30,
36
+ "num_hidden_layers": 32,
37
  "num_generation": 50,
 
38
  "num_population": 100,
39
  "pad_token_id": 1,
40
  "position_embedding_type": "rotary",
 
42
  "torch_dtype": "float32",
43
  "transformers_version": "4.41.0.dev0",
44
  "use_cache": true,
 
45
  "vocab_list": null,
46
  "vocab_size": 24
47
  }