Spaces:
Runtime error
Runtime error
ycy
commited on
Commit
·
ae0984e
1
Parent(s):
4024a0b
- app.py +2 -0
- src/about.py +4 -0
- src/display/formatting.py +2 -2
- src/envs.py +1 -0
- src/submission/submit.py +2 -2
app.py
CHANGED
|
@@ -13,6 +13,7 @@ from src.about import (
|
|
| 13 |
get_INTRODUCTION_TEXT,
|
| 14 |
LLM_BENCHMARKS_TEXT,
|
| 15 |
TITLE,
|
|
|
|
| 16 |
)
|
| 17 |
from src.display.css_html_js import custom_css
|
| 18 |
from src.display.utils import (
|
|
@@ -93,6 +94,7 @@ with demo:
|
|
| 93 |
gr.HTML(get_INTRODUCTION_TEXT(LEADERBOARD_DF.shape[0] , datetime.now(pytz.timezone('US/Pacific')).strftime("%Y-%m-%d %H:%M:%S")), elem_classes="markdown-text")
|
| 94 |
|
| 95 |
with gr.Tabs(elem_classes="tab-buttons") as tabs:
|
|
|
|
| 96 |
with gr.TabItem("🏅 LLM Benchmark", elem_id="llm-benchmark-tab-table", id=0):
|
| 97 |
|
| 98 |
leaderboard = init_leaderboard(LEADERBOARD_DF)
|
|
|
|
| 13 |
get_INTRODUCTION_TEXT,
|
| 14 |
LLM_BENCHMARKS_TEXT,
|
| 15 |
TITLE,
|
| 16 |
+
INTRODUCE_BENCHMARK
|
| 17 |
)
|
| 18 |
from src.display.css_html_js import custom_css
|
| 19 |
from src.display.utils import (
|
|
|
|
| 94 |
gr.HTML(get_INTRODUCTION_TEXT(LEADERBOARD_DF.shape[0] , datetime.now(pytz.timezone('US/Pacific')).strftime("%Y-%m-%d %H:%M:%S")), elem_classes="markdown-text")
|
| 95 |
|
| 96 |
with gr.Tabs(elem_classes="tab-buttons") as tabs:
|
| 97 |
+
gr.HTML(INTRODUCE_BENCHMARK) #TODO
|
| 98 |
with gr.TabItem("🏅 LLM Benchmark", elem_id="llm-benchmark-tab-table", id=0):
|
| 99 |
|
| 100 |
leaderboard = init_leaderboard(LEADERBOARD_DF)
|
src/about.py
CHANGED
|
@@ -44,6 +44,10 @@ def get_INTRODUCTION_TEXT(model_num: int, LAST_UPDATED: str, paper_link="TODO"):
|
|
| 44 |
</div>
|
| 45 |
"""
|
| 46 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 47 |
|
| 48 |
#TODO About
|
| 49 |
LLM_BENCHMARKS_TEXT = f"""
|
|
|
|
| 44 |
</div>
|
| 45 |
"""
|
| 46 |
|
| 47 |
+
#TODO
|
| 48 |
+
INTRODUCE_BENCHMARK = f"""
|
| 49 |
+
<h2 style="color: #2c3e50;"> Introduction </h2>
|
| 50 |
+
"""
|
| 51 |
|
| 52 |
#TODO About
|
| 53 |
LLM_BENCHMARKS_TEXT = f"""
|
src/display/formatting.py
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
-
def model_hyperlink(link,
|
| 2 |
-
return f'<a target="_blank" href="{link}" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">{
|
| 3 |
|
| 4 |
|
| 5 |
def make_clickable_model(model_show , model_name , is_open):
|
|
|
|
| 1 |
+
def model_hyperlink(link, model_show):
|
| 2 |
+
return f'<a target="_blank" href="{link}" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">{model_show}</a>'
|
| 3 |
|
| 4 |
|
| 5 |
def make_clickable_model(model_show , model_name , is_open):
|
src/envs.py
CHANGED
|
@@ -13,6 +13,7 @@ REPO_ID = f"{OWNER}/CapArena_Auto_1"
|
|
| 13 |
QUEUE_REPO = f"{OWNER}/requests"
|
| 14 |
RESULTS_REPO = f"{OWNER}/results"
|
| 15 |
|
|
|
|
| 16 |
# If you setup a cache later, just change HF_HOME
|
| 17 |
CACHE_PATH=os.getenv("HF_HOME", ".")
|
| 18 |
|
|
|
|
| 13 |
QUEUE_REPO = f"{OWNER}/requests"
|
| 14 |
RESULTS_REPO = f"{OWNER}/results"
|
| 15 |
|
| 16 |
+
QUEUE_REPO_TEST = f"{OWNER}/requests-test"
|
| 17 |
# If you setup a cache later, just change HF_HOME
|
| 18 |
CACHE_PATH=os.getenv("HF_HOME", ".")
|
| 19 |
|
src/submission/submit.py
CHANGED
|
@@ -3,7 +3,7 @@ import os
|
|
| 3 |
from datetime import datetime, timezone
|
| 4 |
|
| 5 |
from src.display.formatting import styled_error, styled_message, styled_warning
|
| 6 |
-
from src.envs import API, EVAL_REQUESTS_PATH, TOKEN, QUEUE_REPO
|
| 7 |
from src.submission.check_validity import (
|
| 8 |
already_submitted_models,
|
| 9 |
check_model_card,
|
|
@@ -109,7 +109,7 @@ def add_new_eval(
|
|
| 109 |
API.upload_file(
|
| 110 |
path_or_fileobj=out_path,
|
| 111 |
path_in_repo=out_path.split("eval-queue/")[1],
|
| 112 |
-
repo_id=
|
| 113 |
repo_type="dataset",
|
| 114 |
commit_message=f"Add {model} to eval queue",
|
| 115 |
)
|
|
|
|
| 3 |
from datetime import datetime, timezone
|
| 4 |
|
| 5 |
from src.display.formatting import styled_error, styled_message, styled_warning
|
| 6 |
+
from src.envs import API, EVAL_REQUESTS_PATH, TOKEN, QUEUE_REPO, QUEUE_REPO_TEST
|
| 7 |
from src.submission.check_validity import (
|
| 8 |
already_submitted_models,
|
| 9 |
check_model_card,
|
|
|
|
| 109 |
API.upload_file(
|
| 110 |
path_or_fileobj=out_path,
|
| 111 |
path_in_repo=out_path.split("eval-queue/")[1],
|
| 112 |
+
repo_id=QUEUE_REPO_TEST,
|
| 113 |
repo_type="dataset",
|
| 114 |
commit_message=f"Add {model} to eval queue",
|
| 115 |
)
|