loodvanniekerkginkgo commited on
Commit
5b5ee28
·
1 Parent(s): a99fd50

Removed pycache (hopefully), and fixed assay names

Browse files
.gitignore ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ __pycache__/
2
+ *.pyc
__pycache__/about.cpython-311.pyc DELETED
Binary file (767 Bytes)
 
__pycache__/evaluation.cpython-311.pyc DELETED
Binary file (481 Bytes)
 
__pycache__/submit.cpython-311.pyc DELETED
Binary file (4.06 kB)
 
__pycache__/utils.cpython-311.pyc DELETED
Binary file (3.2 kB)
 
__pycache__/visualize.cpython-311.pyc DELETED
Binary file (679 Bytes)
 
about.py CHANGED
@@ -1,7 +1,21 @@
1
  import os
2
  from huggingface_hub import HfApi
3
 
4
- ASSAY_LIST = ["AC-SINS_pH7.4", "PSP_CHO", "HIC", "Tm2", "Titer"]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  TOKEN = os.environ.get("HF_TOKEN")
6
  CACHE_PATH=os.getenv("HF_HOME", ".")
7
  API = HfApi(token=TOKEN)
 
1
  import os
2
  from huggingface_hub import HfApi
3
 
4
+ ASSAY_LIST = ["AC-SINS_pH7.4", "PR_CHO", "HIC", "Tm2", "Titer"]
5
+ ASSAY_RENAME = {
6
+ "AC-SINS_pH7.4": "Self-association",
7
+ "PR_CHO": "Polyreactivity",
8
+ "HIC": "Hydrophobicity",
9
+ "Tm2": "Thermostability",
10
+ "Titer": "Titer",
11
+ }
12
+ ASSAY_EMOJIS = {
13
+ "AC-SINS_pH7.4": "🧲",
14
+ "PR_CHO": "🎯",
15
+ "HIC": "💧",
16
+ "Tm2": "🌡️",
17
+ "Titer": "🧪",
18
+ }
19
  TOKEN = os.environ.get("HF_TOKEN")
20
  CACHE_PATH=os.getenv("HF_HOME", ".")
21
  API = HfApi(token=TOKEN)
app.py CHANGED
@@ -7,7 +7,7 @@ from gradio_leaderboard import Leaderboard
7
  from evaluation import evaluate_problem
8
 
9
  from utils import read_submission_from_hub, write_results
10
- from about import ASSAY_LIST
11
 
12
  def evaluate_boundary(filename):
13
  print(filename)
@@ -36,8 +36,9 @@ def get_leaderboard_table(assay: str | None = None):
36
  # to_show['user'] = to_show['user'].apply(lambda x: make_user_clickable(x)).astype(str)
37
 
38
  # Previously hosted on HF hub, local for now
39
- column_order = ["model", "feature", "assay", "spearman"]
40
  df = pd.read_csv("data/metrics_all.csv")
 
41
  if assay is not None:
42
  df = df[df['assay'] == assay]
43
  df = df[column_order]
@@ -48,11 +49,11 @@ def get_leaderboard_object(assay: str | None = None):
48
  df = get_leaderboard_table(assay=assay)
49
  filter_columns = ["model"]
50
  if assay is None:
51
- filter_columns.append("assay")
52
  Leaderboard(
53
  value=df,
54
  datatype=["str", "str", "str", "number"],
55
- select_columns=["model", "assay", "feature", "spearman"],
56
  search_columns=["model"],
57
  # hide_columns=["spearman_abs"],
58
  filter_columns=filter_columns,
@@ -77,8 +78,8 @@ with gr.Blocks() as demo:
77
 
78
  # Procedurally make these 5 tabs
79
  for assay in ASSAY_LIST:
80
- with gr.TabItem(assay, elem_id=f"abdev-benchmark-tab-table"):
81
- gr.Markdown(f"# {assay}")
82
  get_leaderboard_object(assay=assay)
83
 
84
  with gr.TabItem("❔About", elem_id="abdev-benchmark-tab-table"):
 
7
  from evaluation import evaluate_problem
8
 
9
  from utils import read_submission_from_hub, write_results
10
+ from about import ASSAY_LIST, ASSAY_RENAME, ASSAY_EMOJIS
11
 
12
  def evaluate_boundary(filename):
13
  print(filename)
 
36
  # to_show['user'] = to_show['user'].apply(lambda x: make_user_clickable(x)).astype(str)
37
 
38
  # Previously hosted on HF hub, local for now
39
+ column_order = ["model", "feature", "property", "spearman", "spearman_abs"] # "assay",
40
  df = pd.read_csv("data/metrics_all.csv")
41
+ df["property"] = df["assay"].map(ASSAY_RENAME)
42
  if assay is not None:
43
  df = df[df['assay'] == assay]
44
  df = df[column_order]
 
49
  df = get_leaderboard_table(assay=assay)
50
  filter_columns = ["model"]
51
  if assay is None:
52
+ filter_columns.append("property")
53
  Leaderboard(
54
  value=df,
55
  datatype=["str", "str", "str", "number"],
56
+ select_columns=["model", "feature", "property", "spearman"],
57
  search_columns=["model"],
58
  # hide_columns=["spearman_abs"],
59
  filter_columns=filter_columns,
 
78
 
79
  # Procedurally make these 5 tabs
80
  for assay in ASSAY_LIST:
81
+ with gr.TabItem(f"{ASSAY_EMOJIS[assay]} {ASSAY_RENAME[assay]}", elem_id=f"abdev-benchmark-tab-table"):
82
+ gr.Markdown(f"# {ASSAY_RENAME[assay]} (measured by {assay})")
83
  get_leaderboard_object(assay=assay)
84
 
85
  with gr.TabItem("❔About", elem_id="abdev-benchmark-tab-table"):