Commit
·
11793f2
1
Parent(s):
de75bee
Added logo and fixed spearman_abs
Browse files- .gitattributes +1 -0
- app.py +7 -6
- assets/competition_logo.jpg +3 -0
.gitattributes
CHANGED
|
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
*.jpg filter=lfs diff=lfs merge=lfs -text
|
app.py
CHANGED
|
@@ -69,12 +69,12 @@ def get_leaderboard_table(df_results: pd.DataFrame, assay: str | None = None):
|
|
| 69 |
# to_show['user'] = to_show['user'].apply(lambda x: make_user_clickable(x)).astype(str)
|
| 70 |
|
| 71 |
# Previously hosted on HF hub, local for now (Can also pull directly from github backend)
|
| 72 |
-
column_order = ["model", "property", "spearman", "
|
| 73 |
df = df_results.query("assay.isin(@ASSAY_RENAME.keys())").copy()
|
| 74 |
if assay is not None:
|
| 75 |
df = df[df['assay'] == assay]
|
| 76 |
df = df[column_order]
|
| 77 |
-
return df.sort_values(by="
|
| 78 |
|
| 79 |
def get_leaderboard_object(df_results: pd.DataFrame, assay: str | None = None):
|
| 80 |
df = get_leaderboard_table(df_results=df_results, assay=assay)
|
|
@@ -85,7 +85,7 @@ def get_leaderboard_object(df_results: pd.DataFrame, assay: str | None = None):
|
|
| 85 |
Leaderboard(
|
| 86 |
value=df,
|
| 87 |
datatype=["str", "str", "str", "number"],
|
| 88 |
-
select_columns=["model", "property", "spearman"],
|
| 89 |
search_columns=["model"],
|
| 90 |
filter_columns=filter_columns,
|
| 91 |
every=60,
|
|
@@ -99,13 +99,14 @@ def fetch_hf_results():
|
|
| 99 |
ds = load_dataset(results_repo, split='no_low_spearman', download_mode="force_redownload")
|
| 100 |
df = pd.DataFrame(ds).drop_duplicates(subset=["model", "assay"])
|
| 101 |
df["property"] = df["assay"].map(ASSAY_RENAME)
|
|
|
|
| 102 |
return df
|
| 103 |
|
| 104 |
with gr.Blocks() as demo:
|
| 105 |
gr.Markdown("""
|
| 106 |
-
## Welcome to the Ginkgo Antibody Developability Benchmark
|
| 107 |
|
| 108 |
-
Participants can submit their model to the leaderboard by
|
| 109 |
""")
|
| 110 |
df = fetch_hf_results()
|
| 111 |
with gr.Tabs(elem_classes="tab-buttons"):
|
|
@@ -117,11 +118,11 @@ with gr.Blocks() as demo:
|
|
| 117 |
|
| 118 |
with gr.TabItem("🚀 Overall", elem_id="abdev-benchmark-tab-table"):
|
| 119 |
gr.Markdown("# Antibody Developability Benchmark Leaderboard over all properties")
|
| 120 |
-
|
| 121 |
get_leaderboard_object(df_results=df)
|
| 122 |
# TODO: this is not going to update well, need to fix
|
| 123 |
|
| 124 |
with gr.TabItem("❔About", elem_id="abdev-benchmark-tab-table"):
|
|
|
|
| 125 |
gr.Markdown(
|
| 126 |
"""
|
| 127 |
## About this challenge
|
|
|
|
| 69 |
# to_show['user'] = to_show['user'].apply(lambda x: make_user_clickable(x)).astype(str)
|
| 70 |
|
| 71 |
# Previously hosted on HF hub, local for now (Can also pull directly from github backend)
|
| 72 |
+
column_order = ["model", "property", "spearman", "spearman_cross_val"]
|
| 73 |
df = df_results.query("assay.isin(@ASSAY_RENAME.keys())").copy()
|
| 74 |
if assay is not None:
|
| 75 |
df = df[df['assay'] == assay]
|
| 76 |
df = df[column_order]
|
| 77 |
+
return df.sort_values(by="spearman", ascending=False)
|
| 78 |
|
| 79 |
def get_leaderboard_object(df_results: pd.DataFrame, assay: str | None = None):
|
| 80 |
df = get_leaderboard_table(df_results=df_results, assay=assay)
|
|
|
|
| 85 |
Leaderboard(
|
| 86 |
value=df,
|
| 87 |
datatype=["str", "str", "str", "number"],
|
| 88 |
+
select_columns=["model", "property", "spearman", "spearman_cross_val"],
|
| 89 |
search_columns=["model"],
|
| 90 |
filter_columns=filter_columns,
|
| 91 |
every=60,
|
|
|
|
| 99 |
ds = load_dataset(results_repo, split='no_low_spearman', download_mode="force_redownload")
|
| 100 |
df = pd.DataFrame(ds).drop_duplicates(subset=["model", "assay"])
|
| 101 |
df["property"] = df["assay"].map(ASSAY_RENAME)
|
| 102 |
+
print(df.head())
|
| 103 |
return df
|
| 104 |
|
| 105 |
with gr.Blocks() as demo:
|
| 106 |
gr.Markdown("""
|
| 107 |
+
## Welcome to the Ginkgo Antibody Developability Benchmark!
|
| 108 |
|
| 109 |
+
Participants can submit their model to the leaderboard by uploading a CSV file (see the "✉️ Submit" tab).
|
| 110 |
""")
|
| 111 |
df = fetch_hf_results()
|
| 112 |
with gr.Tabs(elem_classes="tab-buttons"):
|
|
|
|
| 118 |
|
| 119 |
with gr.TabItem("🚀 Overall", elem_id="abdev-benchmark-tab-table"):
|
| 120 |
gr.Markdown("# Antibody Developability Benchmark Leaderboard over all properties")
|
|
|
|
| 121 |
get_leaderboard_object(df_results=df)
|
| 122 |
# TODO: this is not going to update well, need to fix
|
| 123 |
|
| 124 |
with gr.TabItem("❔About", elem_id="abdev-benchmark-tab-table"):
|
| 125 |
+
gr.Image(value="./assets/competition_logo.jpg")
|
| 126 |
gr.Markdown(
|
| 127 |
"""
|
| 128 |
## About this challenge
|
assets/competition_logo.jpg
ADDED
|
Git LFS Details
|