lhallee commited on
Commit
26032b8
·
verified ·
1 Parent(s): fee7bb3

Upload ProteinVec

Browse files
config.json CHANGED
@@ -1,20 +1,120 @@
1
  {
2
- "activation": "relu",
3
  "architectures": [
4
  "ProteinVec"
5
  ],
6
- "d_model": 512,
7
- "dim_feedforward": 2048,
8
- "dropout": 0.1,
9
- "embedded_path": "models/protein_vec/src_run/protein_vec_models",
10
- "lr0": 0.0001,
11
- "nhead": 4,
12
- "num_layers": 2,
13
- "num_variables": 10,
14
- "out_dim": 512,
15
- "p_bernoulli": 0.5,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  "torch_dtype": "float32",
17
- "transformers_version": "4.38.2",
18
- "vocab": 20,
19
- "warmup_steps": 500
 
 
 
 
 
 
 
 
 
 
 
 
20
  }
 
1
  {
 
2
  "architectures": [
3
  "ProteinVec"
4
  ],
5
+ "bp_activation": "relu",
6
+ "bp_d_model": 1024,
7
+ "bp_dim_feedforward": 2048,
8
+ "bp_dropout": 0.1,
9
+ "bp_lr0": 0.0001,
10
+ "bp_nhead": 4,
11
+ "bp_num_layers": 4,
12
+ "bp_num_variables": 10,
13
+ "bp_out_dim": 512,
14
+ "bp_p_bernoulli": 0.5,
15
+ "bp_vocab": 20,
16
+ "bp_warmup_steps": 500,
17
+ "cc_activation": "relu",
18
+ "cc_d_model": 1024,
19
+ "cc_dim_feedforward": 2048,
20
+ "cc_dropout": 0.1,
21
+ "cc_lr0": 0.0001,
22
+ "cc_nhead": 4,
23
+ "cc_num_layers": 4,
24
+ "cc_num_variables": 10,
25
+ "cc_out_dim": 512,
26
+ "cc_p_bernoulli": 0.5,
27
+ "cc_vocab": 20,
28
+ "cc_warmup_steps": 500,
29
+ "classifier_dropout": 0.0,
30
+ "d_ff": 16384,
31
+ "d_kv": 128,
32
+ "d_model": 1024,
33
+ "dense_act_fn": "relu",
34
+ "dropout_rate": 0.1,
35
+ "ec_activation": "relu",
36
+ "ec_d_model": 1024,
37
+ "ec_dim_feedforward": 2048,
38
+ "ec_dropout": 0.1,
39
+ "ec_lr0": 0.0001,
40
+ "ec_nhead": 4,
41
+ "ec_num_layers": 2,
42
+ "ec_num_variables": 10,
43
+ "ec_out_dim": 512,
44
+ "ec_p_bernoulli": 0.5,
45
+ "ec_vocab": 20,
46
+ "ec_warmup_steps": 500,
47
+ "eos_token_id": 1,
48
+ "feed_forward_proj": "relu",
49
+ "gene3d_activation": "relu",
50
+ "gene3d_d_model": 1024,
51
+ "gene3d_dim_feedforward": 2048,
52
+ "gene3d_dropout": 0.1,
53
+ "gene3d_lr0": 0.0001,
54
+ "gene3d_nhead": 4,
55
+ "gene3d_num_layers": 2,
56
+ "gene3d_num_variables": 10,
57
+ "gene3d_out_dim": 512,
58
+ "gene3d_p_bernoulli": 0.5,
59
+ "gene3d_vocab": 20,
60
+ "gene3d_warmup_steps": 500,
61
+ "initializer_factor": 1.0,
62
+ "is_gated_act": false,
63
+ "layer_norm_epsilon": 1e-06,
64
+ "mf_activation": "relu",
65
+ "mf_d_model": 1024,
66
+ "mf_dim_feedforward": 2048,
67
+ "mf_dropout": 0.1,
68
+ "mf_lr0": 0.0001,
69
+ "mf_nhead": 4,
70
+ "mf_num_layers": 4,
71
+ "mf_num_variables": 10,
72
+ "mf_out_dim": 512,
73
+ "mf_p_bernoulli": 0.5,
74
+ "mf_vocab": 20,
75
+ "mf_warmup_steps": 500,
76
+ "model_type": "t5",
77
+ "num_decoder_layers": 24,
78
+ "num_heads": 32,
79
+ "num_layers": 24,
80
+ "pad_token_id": 0,
81
+ "pfam_activation": "relu",
82
+ "pfam_d_model": 1024,
83
+ "pfam_dim_feedforward": 2048,
84
+ "pfam_dropout": 0.1,
85
+ "pfam_lr0": 0.0001,
86
+ "pfam_nhead": 4,
87
+ "pfam_num_layers": 2,
88
+ "pfam_num_variables": 10,
89
+ "pfam_out_dim": 512,
90
+ "pfam_p_bernoulli": 0.5,
91
+ "pfam_vocab": 20,
92
+ "pfam_warmup_steps": 500,
93
+ "relative_attention_max_distance": null,
94
+ "relative_attention_num_buckets": 32,
95
+ "tm_activation": "relu",
96
+ "tm_d_model": 1024,
97
+ "tm_dim_feedforward": 2048,
98
+ "tm_dropout": 0.1,
99
+ "tm_lr0": 0.0001,
100
+ "tm_nhead": 4,
101
+ "tm_num_layers": 4,
102
+ "tm_out_dim": 512,
103
+ "tm_warmup_steps": 300,
104
  "torch_dtype": "float32",
105
+ "transformers_version": "4.38.1",
106
+ "use_cache": true,
107
+ "vec_activation": "relu",
108
+ "vec_d_model": 1024,
109
+ "vec_dim_feedforward": 2048,
110
+ "vec_dropout": 0.1,
111
+ "vec_lr0": 0.0001,
112
+ "vec_nhead": 4,
113
+ "vec_num_layers": 2,
114
+ "vec_num_variables": 10,
115
+ "vec_out_dim": 512,
116
+ "vec_p_bernoulli": 0.5,
117
+ "vec_vocab": 20,
118
+ "vec_warmup_steps": 500,
119
+ "vocab_size": 128
120
  }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6a2e2a081a78c4d5ce1d72a5e545e211aa6a83140cd6954b097ceff1a2e52d49
3
- size 4996418440
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac4ed24c74dce0afa06faf9efdd2e32fa179002491ddc2e8eeac83cb9e93403f
3
+ size 4995336712
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:365bbe7727ae454be8bb8fdb173a49f0ca762a3afb3b4b6a7055a02f45eb92b3
3
- size 623719592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8753bd949ac67f613ff84915e80694d153a9dcae910c3b587522a20360baadc7
3
+ size 667829856
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 5620074496
4
  },
5
  "weight_map": {
6
  "moe.encoder.layers.0.linear1.bias": "model-00001-of-00002.safetensors",
@@ -297,30 +297,30 @@
297
  "moe.model_aspect_tmvec.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
298
  "moe.model_aspect_tmvec.encoder.layers.2.linear1.bias": "model-00001-of-00002.safetensors",
299
  "moe.model_aspect_tmvec.encoder.layers.2.linear1.weight": "model-00001-of-00002.safetensors",
300
- "moe.model_aspect_tmvec.encoder.layers.2.linear2.bias": "model-00001-of-00002.safetensors",
301
- "moe.model_aspect_tmvec.encoder.layers.2.linear2.weight": "model-00001-of-00002.safetensors",
302
- "moe.model_aspect_tmvec.encoder.layers.2.norm1.bias": "model-00001-of-00002.safetensors",
303
- "moe.model_aspect_tmvec.encoder.layers.2.norm1.weight": "model-00001-of-00002.safetensors",
304
- "moe.model_aspect_tmvec.encoder.layers.2.norm2.bias": "model-00001-of-00002.safetensors",
305
- "moe.model_aspect_tmvec.encoder.layers.2.norm2.weight": "model-00001-of-00002.safetensors",
306
  "moe.model_aspect_tmvec.encoder.layers.2.self_attn.in_proj_bias": "model-00001-of-00002.safetensors",
307
  "moe.model_aspect_tmvec.encoder.layers.2.self_attn.in_proj_weight": "model-00001-of-00002.safetensors",
308
  "moe.model_aspect_tmvec.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
309
  "moe.model_aspect_tmvec.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
310
- "moe.model_aspect_tmvec.encoder.layers.3.linear1.bias": "model-00001-of-00002.safetensors",
311
- "moe.model_aspect_tmvec.encoder.layers.3.linear1.weight": "model-00001-of-00002.safetensors",
312
- "moe.model_aspect_tmvec.encoder.layers.3.linear2.bias": "model-00001-of-00002.safetensors",
313
- "moe.model_aspect_tmvec.encoder.layers.3.linear2.weight": "model-00001-of-00002.safetensors",
314
- "moe.model_aspect_tmvec.encoder.layers.3.norm1.bias": "model-00001-of-00002.safetensors",
315
- "moe.model_aspect_tmvec.encoder.layers.3.norm1.weight": "model-00001-of-00002.safetensors",
316
- "moe.model_aspect_tmvec.encoder.layers.3.norm2.bias": "model-00001-of-00002.safetensors",
317
- "moe.model_aspect_tmvec.encoder.layers.3.norm2.weight": "model-00001-of-00002.safetensors",
318
- "moe.model_aspect_tmvec.encoder.layers.3.self_attn.in_proj_bias": "model-00001-of-00002.safetensors",
319
- "moe.model_aspect_tmvec.encoder.layers.3.self_attn.in_proj_weight": "model-00001-of-00002.safetensors",
320
- "moe.model_aspect_tmvec.encoder.layers.3.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
321
- "moe.model_aspect_tmvec.encoder.layers.3.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
322
- "moe.model_aspect_tmvec.mlp.bias": "model-00001-of-00002.safetensors",
323
- "moe.model_aspect_tmvec.mlp.weight": "model-00001-of-00002.safetensors",
324
  "t5.encoder.block.0.layer.0.SelfAttention.k.weight": "model-00001-of-00002.safetensors",
325
  "t5.encoder.block.0.layer.0.SelfAttention.o.weight": "model-00001-of-00002.safetensors",
326
  "t5.encoder.block.0.layer.0.SelfAttention.q.weight": "model-00001-of-00002.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 5663102976
4
  },
5
  "weight_map": {
6
  "moe.encoder.layers.0.linear1.bias": "model-00001-of-00002.safetensors",
 
297
  "moe.model_aspect_tmvec.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
298
  "moe.model_aspect_tmvec.encoder.layers.2.linear1.bias": "model-00001-of-00002.safetensors",
299
  "moe.model_aspect_tmvec.encoder.layers.2.linear1.weight": "model-00001-of-00002.safetensors",
300
+ "moe.model_aspect_tmvec.encoder.layers.2.linear2.bias": "model-00002-of-00002.safetensors",
301
+ "moe.model_aspect_tmvec.encoder.layers.2.linear2.weight": "model-00002-of-00002.safetensors",
302
+ "moe.model_aspect_tmvec.encoder.layers.2.norm1.bias": "model-00002-of-00002.safetensors",
303
+ "moe.model_aspect_tmvec.encoder.layers.2.norm1.weight": "model-00002-of-00002.safetensors",
304
+ "moe.model_aspect_tmvec.encoder.layers.2.norm2.bias": "model-00002-of-00002.safetensors",
305
+ "moe.model_aspect_tmvec.encoder.layers.2.norm2.weight": "model-00002-of-00002.safetensors",
306
  "moe.model_aspect_tmvec.encoder.layers.2.self_attn.in_proj_bias": "model-00001-of-00002.safetensors",
307
  "moe.model_aspect_tmvec.encoder.layers.2.self_attn.in_proj_weight": "model-00001-of-00002.safetensors",
308
  "moe.model_aspect_tmvec.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
309
  "moe.model_aspect_tmvec.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
310
+ "moe.model_aspect_tmvec.encoder.layers.3.linear1.bias": "model-00002-of-00002.safetensors",
311
+ "moe.model_aspect_tmvec.encoder.layers.3.linear1.weight": "model-00002-of-00002.safetensors",
312
+ "moe.model_aspect_tmvec.encoder.layers.3.linear2.bias": "model-00002-of-00002.safetensors",
313
+ "moe.model_aspect_tmvec.encoder.layers.3.linear2.weight": "model-00002-of-00002.safetensors",
314
+ "moe.model_aspect_tmvec.encoder.layers.3.norm1.bias": "model-00002-of-00002.safetensors",
315
+ "moe.model_aspect_tmvec.encoder.layers.3.norm1.weight": "model-00002-of-00002.safetensors",
316
+ "moe.model_aspect_tmvec.encoder.layers.3.norm2.bias": "model-00002-of-00002.safetensors",
317
+ "moe.model_aspect_tmvec.encoder.layers.3.norm2.weight": "model-00002-of-00002.safetensors",
318
+ "moe.model_aspect_tmvec.encoder.layers.3.self_attn.in_proj_bias": "model-00002-of-00002.safetensors",
319
+ "moe.model_aspect_tmvec.encoder.layers.3.self_attn.in_proj_weight": "model-00002-of-00002.safetensors",
320
+ "moe.model_aspect_tmvec.encoder.layers.3.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
321
+ "moe.model_aspect_tmvec.encoder.layers.3.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
322
+ "moe.model_aspect_tmvec.mlp.bias": "model-00002-of-00002.safetensors",
323
+ "moe.model_aspect_tmvec.mlp.weight": "model-00002-of-00002.safetensors",
324
  "t5.encoder.block.0.layer.0.SelfAttention.k.weight": "model-00001-of-00002.safetensors",
325
  "t5.encoder.block.0.layer.0.SelfAttention.o.weight": "model-00001-of-00002.safetensors",
326
  "t5.encoder.block.0.layer.0.SelfAttention.q.weight": "model-00001-of-00002.safetensors",