Upload ProteinVec
Browse files- config.json +114 -14
- model-00001-of-00002.safetensors +2 -2
- model-00002-of-00002.safetensors +2 -2
- model.safetensors.index.json +21 -21
config.json
CHANGED
@@ -1,20 +1,120 @@
|
|
1 |
{
|
2 |
-
"activation": "relu",
|
3 |
"architectures": [
|
4 |
"ProteinVec"
|
5 |
],
|
6 |
-
"
|
7 |
-
"
|
8 |
-
"
|
9 |
-
"
|
10 |
-
"
|
11 |
-
"
|
12 |
-
"
|
13 |
-
"
|
14 |
-
"
|
15 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
"torch_dtype": "float32",
|
17 |
-
"transformers_version": "4.38.
|
18 |
-
"
|
19 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
20 |
}
|
|
|
1 |
{
|
|
|
2 |
"architectures": [
|
3 |
"ProteinVec"
|
4 |
],
|
5 |
+
"bp_activation": "relu",
|
6 |
+
"bp_d_model": 1024,
|
7 |
+
"bp_dim_feedforward": 2048,
|
8 |
+
"bp_dropout": 0.1,
|
9 |
+
"bp_lr0": 0.0001,
|
10 |
+
"bp_nhead": 4,
|
11 |
+
"bp_num_layers": 4,
|
12 |
+
"bp_num_variables": 10,
|
13 |
+
"bp_out_dim": 512,
|
14 |
+
"bp_p_bernoulli": 0.5,
|
15 |
+
"bp_vocab": 20,
|
16 |
+
"bp_warmup_steps": 500,
|
17 |
+
"cc_activation": "relu",
|
18 |
+
"cc_d_model": 1024,
|
19 |
+
"cc_dim_feedforward": 2048,
|
20 |
+
"cc_dropout": 0.1,
|
21 |
+
"cc_lr0": 0.0001,
|
22 |
+
"cc_nhead": 4,
|
23 |
+
"cc_num_layers": 4,
|
24 |
+
"cc_num_variables": 10,
|
25 |
+
"cc_out_dim": 512,
|
26 |
+
"cc_p_bernoulli": 0.5,
|
27 |
+
"cc_vocab": 20,
|
28 |
+
"cc_warmup_steps": 500,
|
29 |
+
"classifier_dropout": 0.0,
|
30 |
+
"d_ff": 16384,
|
31 |
+
"d_kv": 128,
|
32 |
+
"d_model": 1024,
|
33 |
+
"dense_act_fn": "relu",
|
34 |
+
"dropout_rate": 0.1,
|
35 |
+
"ec_activation": "relu",
|
36 |
+
"ec_d_model": 1024,
|
37 |
+
"ec_dim_feedforward": 2048,
|
38 |
+
"ec_dropout": 0.1,
|
39 |
+
"ec_lr0": 0.0001,
|
40 |
+
"ec_nhead": 4,
|
41 |
+
"ec_num_layers": 2,
|
42 |
+
"ec_num_variables": 10,
|
43 |
+
"ec_out_dim": 512,
|
44 |
+
"ec_p_bernoulli": 0.5,
|
45 |
+
"ec_vocab": 20,
|
46 |
+
"ec_warmup_steps": 500,
|
47 |
+
"eos_token_id": 1,
|
48 |
+
"feed_forward_proj": "relu",
|
49 |
+
"gene3d_activation": "relu",
|
50 |
+
"gene3d_d_model": 1024,
|
51 |
+
"gene3d_dim_feedforward": 2048,
|
52 |
+
"gene3d_dropout": 0.1,
|
53 |
+
"gene3d_lr0": 0.0001,
|
54 |
+
"gene3d_nhead": 4,
|
55 |
+
"gene3d_num_layers": 2,
|
56 |
+
"gene3d_num_variables": 10,
|
57 |
+
"gene3d_out_dim": 512,
|
58 |
+
"gene3d_p_bernoulli": 0.5,
|
59 |
+
"gene3d_vocab": 20,
|
60 |
+
"gene3d_warmup_steps": 500,
|
61 |
+
"initializer_factor": 1.0,
|
62 |
+
"is_gated_act": false,
|
63 |
+
"layer_norm_epsilon": 1e-06,
|
64 |
+
"mf_activation": "relu",
|
65 |
+
"mf_d_model": 1024,
|
66 |
+
"mf_dim_feedforward": 2048,
|
67 |
+
"mf_dropout": 0.1,
|
68 |
+
"mf_lr0": 0.0001,
|
69 |
+
"mf_nhead": 4,
|
70 |
+
"mf_num_layers": 4,
|
71 |
+
"mf_num_variables": 10,
|
72 |
+
"mf_out_dim": 512,
|
73 |
+
"mf_p_bernoulli": 0.5,
|
74 |
+
"mf_vocab": 20,
|
75 |
+
"mf_warmup_steps": 500,
|
76 |
+
"model_type": "t5",
|
77 |
+
"num_decoder_layers": 24,
|
78 |
+
"num_heads": 32,
|
79 |
+
"num_layers": 24,
|
80 |
+
"pad_token_id": 0,
|
81 |
+
"pfam_activation": "relu",
|
82 |
+
"pfam_d_model": 1024,
|
83 |
+
"pfam_dim_feedforward": 2048,
|
84 |
+
"pfam_dropout": 0.1,
|
85 |
+
"pfam_lr0": 0.0001,
|
86 |
+
"pfam_nhead": 4,
|
87 |
+
"pfam_num_layers": 2,
|
88 |
+
"pfam_num_variables": 10,
|
89 |
+
"pfam_out_dim": 512,
|
90 |
+
"pfam_p_bernoulli": 0.5,
|
91 |
+
"pfam_vocab": 20,
|
92 |
+
"pfam_warmup_steps": 500,
|
93 |
+
"relative_attention_max_distance": null,
|
94 |
+
"relative_attention_num_buckets": 32,
|
95 |
+
"tm_activation": "relu",
|
96 |
+
"tm_d_model": 1024,
|
97 |
+
"tm_dim_feedforward": 2048,
|
98 |
+
"tm_dropout": 0.1,
|
99 |
+
"tm_lr0": 0.0001,
|
100 |
+
"tm_nhead": 4,
|
101 |
+
"tm_num_layers": 4,
|
102 |
+
"tm_out_dim": 512,
|
103 |
+
"tm_warmup_steps": 300,
|
104 |
"torch_dtype": "float32",
|
105 |
+
"transformers_version": "4.38.1",
|
106 |
+
"use_cache": true,
|
107 |
+
"vec_activation": "relu",
|
108 |
+
"vec_d_model": 1024,
|
109 |
+
"vec_dim_feedforward": 2048,
|
110 |
+
"vec_dropout": 0.1,
|
111 |
+
"vec_lr0": 0.0001,
|
112 |
+
"vec_nhead": 4,
|
113 |
+
"vec_num_layers": 2,
|
114 |
+
"vec_num_variables": 10,
|
115 |
+
"vec_out_dim": 512,
|
116 |
+
"vec_p_bernoulli": 0.5,
|
117 |
+
"vec_vocab": 20,
|
118 |
+
"vec_warmup_steps": 500,
|
119 |
+
"vocab_size": 128
|
120 |
}
|
model-00001-of-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ac4ed24c74dce0afa06faf9efdd2e32fa179002491ddc2e8eeac83cb9e93403f
|
3 |
+
size 4995336712
|
model-00002-of-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8753bd949ac67f613ff84915e80694d153a9dcae910c3b587522a20360baadc7
|
3 |
+
size 667829856
|
model.safetensors.index.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
-
"total_size":
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"moe.encoder.layers.0.linear1.bias": "model-00001-of-00002.safetensors",
|
@@ -297,30 +297,30 @@
|
|
297 |
"moe.model_aspect_tmvec.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
298 |
"moe.model_aspect_tmvec.encoder.layers.2.linear1.bias": "model-00001-of-00002.safetensors",
|
299 |
"moe.model_aspect_tmvec.encoder.layers.2.linear1.weight": "model-00001-of-00002.safetensors",
|
300 |
-
"moe.model_aspect_tmvec.encoder.layers.2.linear2.bias": "model-
|
301 |
-
"moe.model_aspect_tmvec.encoder.layers.2.linear2.weight": "model-
|
302 |
-
"moe.model_aspect_tmvec.encoder.layers.2.norm1.bias": "model-
|
303 |
-
"moe.model_aspect_tmvec.encoder.layers.2.norm1.weight": "model-
|
304 |
-
"moe.model_aspect_tmvec.encoder.layers.2.norm2.bias": "model-
|
305 |
-
"moe.model_aspect_tmvec.encoder.layers.2.norm2.weight": "model-
|
306 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.in_proj_bias": "model-00001-of-00002.safetensors",
|
307 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.in_proj_weight": "model-00001-of-00002.safetensors",
|
308 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
309 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
310 |
-
"moe.model_aspect_tmvec.encoder.layers.3.linear1.bias": "model-
|
311 |
-
"moe.model_aspect_tmvec.encoder.layers.3.linear1.weight": "model-
|
312 |
-
"moe.model_aspect_tmvec.encoder.layers.3.linear2.bias": "model-
|
313 |
-
"moe.model_aspect_tmvec.encoder.layers.3.linear2.weight": "model-
|
314 |
-
"moe.model_aspect_tmvec.encoder.layers.3.norm1.bias": "model-
|
315 |
-
"moe.model_aspect_tmvec.encoder.layers.3.norm1.weight": "model-
|
316 |
-
"moe.model_aspect_tmvec.encoder.layers.3.norm2.bias": "model-
|
317 |
-
"moe.model_aspect_tmvec.encoder.layers.3.norm2.weight": "model-
|
318 |
-
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.in_proj_bias": "model-
|
319 |
-
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.in_proj_weight": "model-
|
320 |
-
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.out_proj.bias": "model-
|
321 |
-
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.out_proj.weight": "model-
|
322 |
-
"moe.model_aspect_tmvec.mlp.bias": "model-
|
323 |
-
"moe.model_aspect_tmvec.mlp.weight": "model-
|
324 |
"t5.encoder.block.0.layer.0.SelfAttention.k.weight": "model-00001-of-00002.safetensors",
|
325 |
"t5.encoder.block.0.layer.0.SelfAttention.o.weight": "model-00001-of-00002.safetensors",
|
326 |
"t5.encoder.block.0.layer.0.SelfAttention.q.weight": "model-00001-of-00002.safetensors",
|
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
+
"total_size": 5663102976
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"moe.encoder.layers.0.linear1.bias": "model-00001-of-00002.safetensors",
|
|
|
297 |
"moe.model_aspect_tmvec.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
298 |
"moe.model_aspect_tmvec.encoder.layers.2.linear1.bias": "model-00001-of-00002.safetensors",
|
299 |
"moe.model_aspect_tmvec.encoder.layers.2.linear1.weight": "model-00001-of-00002.safetensors",
|
300 |
+
"moe.model_aspect_tmvec.encoder.layers.2.linear2.bias": "model-00002-of-00002.safetensors",
|
301 |
+
"moe.model_aspect_tmvec.encoder.layers.2.linear2.weight": "model-00002-of-00002.safetensors",
|
302 |
+
"moe.model_aspect_tmvec.encoder.layers.2.norm1.bias": "model-00002-of-00002.safetensors",
|
303 |
+
"moe.model_aspect_tmvec.encoder.layers.2.norm1.weight": "model-00002-of-00002.safetensors",
|
304 |
+
"moe.model_aspect_tmvec.encoder.layers.2.norm2.bias": "model-00002-of-00002.safetensors",
|
305 |
+
"moe.model_aspect_tmvec.encoder.layers.2.norm2.weight": "model-00002-of-00002.safetensors",
|
306 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.in_proj_bias": "model-00001-of-00002.safetensors",
|
307 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.in_proj_weight": "model-00001-of-00002.safetensors",
|
308 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
|
309 |
"moe.model_aspect_tmvec.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
|
310 |
+
"moe.model_aspect_tmvec.encoder.layers.3.linear1.bias": "model-00002-of-00002.safetensors",
|
311 |
+
"moe.model_aspect_tmvec.encoder.layers.3.linear1.weight": "model-00002-of-00002.safetensors",
|
312 |
+
"moe.model_aspect_tmvec.encoder.layers.3.linear2.bias": "model-00002-of-00002.safetensors",
|
313 |
+
"moe.model_aspect_tmvec.encoder.layers.3.linear2.weight": "model-00002-of-00002.safetensors",
|
314 |
+
"moe.model_aspect_tmvec.encoder.layers.3.norm1.bias": "model-00002-of-00002.safetensors",
|
315 |
+
"moe.model_aspect_tmvec.encoder.layers.3.norm1.weight": "model-00002-of-00002.safetensors",
|
316 |
+
"moe.model_aspect_tmvec.encoder.layers.3.norm2.bias": "model-00002-of-00002.safetensors",
|
317 |
+
"moe.model_aspect_tmvec.encoder.layers.3.norm2.weight": "model-00002-of-00002.safetensors",
|
318 |
+
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.in_proj_bias": "model-00002-of-00002.safetensors",
|
319 |
+
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.in_proj_weight": "model-00002-of-00002.safetensors",
|
320 |
+
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.out_proj.bias": "model-00002-of-00002.safetensors",
|
321 |
+
"moe.model_aspect_tmvec.encoder.layers.3.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
322 |
+
"moe.model_aspect_tmvec.mlp.bias": "model-00002-of-00002.safetensors",
|
323 |
+
"moe.model_aspect_tmvec.mlp.weight": "model-00002-of-00002.safetensors",
|
324 |
"t5.encoder.block.0.layer.0.SelfAttention.k.weight": "model-00001-of-00002.safetensors",
|
325 |
"t5.encoder.block.0.layer.0.SelfAttention.o.weight": "model-00001-of-00002.safetensors",
|
326 |
"t5.encoder.block.0.layer.0.SelfAttention.q.weight": "model-00001-of-00002.safetensors",
|