Spaces:
Sleeping
Sleeping
Jimin Huang
commited on
Commit
·
3adbd07
1
Parent(s):
2e3dc13
Add dataset
Browse files- backend/app/services/leaderboard.py +4 -29
- backend/poetry.lock +0 -0
- backend/pyproject.toml +1 -0
- contents.py +4 -0
- frontend/src/pages/LeaderboardPage/components/Leaderboard/constants/defaults.js +2 -32
- frontend/src/pages/LeaderboardPage/components/Leaderboard/constants/tooltips.js +5 -93
- frontend/src/pages/LeaderboardPage/components/Leaderboard/utils/columnUtils.js +3 -48
- frontend/src/pages/QuotePage/QuotePage.js +1 -1
backend/app/services/leaderboard.py
CHANGED
|
@@ -107,35 +107,10 @@ class LeaderboardService:
|
|
| 107 |
|
| 108 |
evaluations = {
|
| 109 |
"ifeval": {
|
| 110 |
-
"name": "
|
| 111 |
-
"value": data.get("
|
| 112 |
-
"normalized_score": data.get("
|
| 113 |
},
|
| 114 |
-
"bbh": {
|
| 115 |
-
"name": "BBH",
|
| 116 |
-
"value": data.get("BBH Raw", 0),
|
| 117 |
-
"normalized_score": data.get("BBH", 0)
|
| 118 |
-
},
|
| 119 |
-
"math": {
|
| 120 |
-
"name": "MATH Level 5",
|
| 121 |
-
"value": data.get("MATH Lvl 5 Raw", 0),
|
| 122 |
-
"normalized_score": data.get("MATH Lvl 5", 0)
|
| 123 |
-
},
|
| 124 |
-
"gpqa": {
|
| 125 |
-
"name": "GPQA",
|
| 126 |
-
"value": data.get("GPQA Raw", 0),
|
| 127 |
-
"normalized_score": data.get("GPQA", 0)
|
| 128 |
-
},
|
| 129 |
-
"musr": {
|
| 130 |
-
"name": "MUSR",
|
| 131 |
-
"value": data.get("MUSR Raw", 0),
|
| 132 |
-
"normalized_score": data.get("MUSR", 0)
|
| 133 |
-
},
|
| 134 |
-
"mmlu_pro": {
|
| 135 |
-
"name": "MMLU-PRO",
|
| 136 |
-
"value": data.get("MMLU-PRO Raw", 0),
|
| 137 |
-
"normalized_score": data.get("MMLU-PRO", 0)
|
| 138 |
-
}
|
| 139 |
}
|
| 140 |
|
| 141 |
features = {
|
|
@@ -205,4 +180,4 @@ class LeaderboardService:
|
|
| 205 |
|
| 206 |
except Exception as e:
|
| 207 |
logger.error(LogFormatter.error(f"Failed to transform data for {data.get('fullname', 'Unknown')}", e))
|
| 208 |
-
raise
|
|
|
|
| 107 |
|
| 108 |
evaluations = {
|
| 109 |
"ifeval": {
|
| 110 |
+
"name": "MultiFin",
|
| 111 |
+
"value": data.get("MultiFin Raw", 0),
|
| 112 |
+
"normalized_score": data.get("MultiFin", 0)
|
| 113 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 114 |
}
|
| 115 |
|
| 116 |
features = {
|
|
|
|
| 180 |
|
| 181 |
except Exception as e:
|
| 182 |
logger.error(LogFormatter.error(f"Failed to transform data for {data.get('fullname', 'Unknown')}", e))
|
| 183 |
+
raise
|
backend/poetry.lock
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
backend/pyproject.toml
CHANGED
|
@@ -19,6 +19,7 @@ safetensors = "^0.4.5"
|
|
| 19 |
aiofiles = "^24.1.0"
|
| 20 |
fastapi-cache2 = "^0.2.1"
|
| 21 |
python-dotenv = "^1.0.1"
|
|
|
|
| 22 |
|
| 23 |
[tool.poetry.group.dev.dependencies]
|
| 24 |
pytest = "^8.3.4"
|
|
|
|
| 19 |
aiofiles = "^24.1.0"
|
| 20 |
fastapi-cache2 = "^0.2.1"
|
| 21 |
python-dotenv = "^1.0.1"
|
| 22 |
+
pydantic = "^2.10.4"
|
| 23 |
|
| 24 |
[tool.poetry.group.dev.dependencies]
|
| 25 |
pytest = "^8.3.4"
|
contents.py
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from datasets import load_dataset
|
| 2 |
+
|
| 3 |
+
ds = load_dataset("open-llm-leaderboard/contents")
|
| 4 |
+
ds.push_to_hub("TheFinAI/greek-contents")
|
frontend/src/pages/LeaderboardPage/components/Leaderboard/constants/defaults.js
CHANGED
|
@@ -125,41 +125,11 @@ const COLUMNS = {
|
|
| 125 |
},
|
| 126 |
},
|
| 127 |
EVALUATION: {
|
| 128 |
-
"evaluations.
|
| 129 |
group: "evaluation",
|
| 130 |
size: COLUMN_SIZES.BENCHMARK,
|
| 131 |
defaultVisible: true,
|
| 132 |
-
label: "
|
| 133 |
-
},
|
| 134 |
-
"evaluations.bbh.normalized_score": {
|
| 135 |
-
group: "evaluation",
|
| 136 |
-
size: COLUMN_SIZES.BENCHMARK,
|
| 137 |
-
defaultVisible: true,
|
| 138 |
-
label: "BBH",
|
| 139 |
-
},
|
| 140 |
-
"evaluations.math.normalized_score": {
|
| 141 |
-
group: "evaluation",
|
| 142 |
-
size: COLUMN_SIZES.BENCHMARK,
|
| 143 |
-
defaultVisible: true,
|
| 144 |
-
label: "MATH",
|
| 145 |
-
},
|
| 146 |
-
"evaluations.gpqa.normalized_score": {
|
| 147 |
-
group: "evaluation",
|
| 148 |
-
size: COLUMN_SIZES.BENCHMARK,
|
| 149 |
-
defaultVisible: true,
|
| 150 |
-
label: "GPQA",
|
| 151 |
-
},
|
| 152 |
-
"evaluations.musr.normalized_score": {
|
| 153 |
-
group: "evaluation",
|
| 154 |
-
size: COLUMN_SIZES.BENCHMARK,
|
| 155 |
-
defaultVisible: true,
|
| 156 |
-
label: "MUSR",
|
| 157 |
-
},
|
| 158 |
-
"evaluations.mmlu_pro.normalized_score": {
|
| 159 |
-
group: "evaluation",
|
| 160 |
-
size: COLUMN_SIZES.BENCHMARK,
|
| 161 |
-
defaultVisible: true,
|
| 162 |
-
label: "MMLU-PRO",
|
| 163 |
},
|
| 164 |
},
|
| 165 |
MODEL_INFO: {
|
|
|
|
| 125 |
},
|
| 126 |
},
|
| 127 |
EVALUATION: {
|
| 128 |
+
"evaluations.multifin.normalized_score": {
|
| 129 |
group: "evaluation",
|
| 130 |
size: COLUMN_SIZES.BENCHMARK,
|
| 131 |
defaultVisible: true,
|
| 132 |
+
label: "MultiFin",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 133 |
},
|
| 134 |
},
|
| 135 |
MODEL_INFO: {
|
frontend/src/pages/LeaderboardPage/components/Leaderboard/constants/tooltips.js
CHANGED
|
@@ -40,107 +40,19 @@ export const COLUMN_TOOLTIPS = {
|
|
| 40 |
},
|
| 41 |
]),
|
| 42 |
|
| 43 |
-
|
| 44 |
{
|
| 45 |
label: "Purpose",
|
| 46 |
description:
|
| 47 |
-
"Tests model's ability to
|
| 48 |
-
subItems: ["
|
| 49 |
},
|
| 50 |
{
|
| 51 |
-
label: "Scoring:
|
| 52 |
-
description: "Was the
|
| 53 |
},
|
| 54 |
]),
|
| 55 |
|
| 56 |
-
BBH: createTooltipContent("Big Bench Hard (BBH):", [
|
| 57 |
-
{
|
| 58 |
-
label: "Overview",
|
| 59 |
-
description: "Collection of challenging for LLM tasks across domains, for example",
|
| 60 |
-
subItems: [
|
| 61 |
-
"Language understanding",
|
| 62 |
-
"Mathematical reasoning",
|
| 63 |
-
"Common sense and world knowledge",
|
| 64 |
-
],
|
| 65 |
-
},
|
| 66 |
-
{
|
| 67 |
-
label: "Scoring: Accuracy",
|
| 68 |
-
description:
|
| 69 |
-
"Was the correct choice selected among the options.",
|
| 70 |
-
},
|
| 71 |
-
]),
|
| 72 |
-
|
| 73 |
-
MATH: createTooltipContent(
|
| 74 |
-
"Mathematics Aptitude Test of Heuristics (MATH), level 5:",
|
| 75 |
-
[
|
| 76 |
-
{
|
| 77 |
-
label: "Content",
|
| 78 |
-
description: "High school level competitions mathematical problems",
|
| 79 |
-
subItems: ["Complex algebra", "Geometry problems", "Advanced calculus"],
|
| 80 |
-
},
|
| 81 |
-
{
|
| 82 |
-
label: "Scoring: Exact match",
|
| 83 |
-
description:
|
| 84 |
-
"Was the solution generated correct and in the expected format",
|
| 85 |
-
},
|
| 86 |
-
]
|
| 87 |
-
),
|
| 88 |
-
|
| 89 |
-
GPQA: createTooltipContent("Graduate-Level Google-Proof Q&A (GPQA):", [
|
| 90 |
-
{
|
| 91 |
-
label: "Focus",
|
| 92 |
-
description: "PhD-level knowledge multiple choice questions in science",
|
| 93 |
-
subItems: [
|
| 94 |
-
"Chemistry",
|
| 95 |
-
"Biology",
|
| 96 |
-
"Physics",
|
| 97 |
-
],
|
| 98 |
-
},
|
| 99 |
-
{
|
| 100 |
-
label: "Scoring: Accuracy",
|
| 101 |
-
description:
|
| 102 |
-
"Was the correct choice selected among the options.",
|
| 103 |
-
},
|
| 104 |
-
]),
|
| 105 |
-
|
| 106 |
-
MUSR: createTooltipContent("Multistep Soft Reasoning (MuSR):", [
|
| 107 |
-
{
|
| 108 |
-
label: "Scope",
|
| 109 |
-
description: "Reasoning and understanding on/of long texts",
|
| 110 |
-
subItems: [
|
| 111 |
-
"Language understanding",
|
| 112 |
-
"Reasoning capabilities",
|
| 113 |
-
"Long context reasoning",
|
| 114 |
-
],
|
| 115 |
-
},
|
| 116 |
-
{
|
| 117 |
-
label: "Scoring: Accuracy",
|
| 118 |
-
description:
|
| 119 |
-
"Was the correct choice selected among the options.",
|
| 120 |
-
},
|
| 121 |
-
]),
|
| 122 |
-
|
| 123 |
-
MMLU_PRO: createTooltipContent(
|
| 124 |
-
"Massive Multitask Language Understanding - Professional (MMLU-Pro):",
|
| 125 |
-
[
|
| 126 |
-
{
|
| 127 |
-
label: "Coverage",
|
| 128 |
-
description: "Expertly reviewed multichoice questions across domains, for example:",
|
| 129 |
-
subItems: [
|
| 130 |
-
"Medicine and healthcare",
|
| 131 |
-
"Law and ethics",
|
| 132 |
-
"Engineering",
|
| 133 |
-
"Mathematics",
|
| 134 |
-
],
|
| 135 |
-
},
|
| 136 |
-
{
|
| 137 |
-
label: "Scoring: Accuracy",
|
| 138 |
-
description:
|
| 139 |
-
"Was the correct choice selected among the options.",
|
| 140 |
-
},
|
| 141 |
-
]
|
| 142 |
-
),
|
| 143 |
-
|
| 144 |
ARCHITECTURE: createTooltipContent("Model Architecture Information:", [
|
| 145 |
{
|
| 146 |
label: "Definition",
|
|
|
|
| 40 |
},
|
| 41 |
]),
|
| 42 |
|
| 43 |
+
MultiFin: createTooltipContent("Multilingual Financial NLP (MultiFin):", [
|
| 44 |
{
|
| 45 |
label: "Purpose",
|
| 46 |
description:
|
| 47 |
+
"Tests model's ability to understand real-world financial article headlines",
|
| 48 |
+
subItems: ["Language Understanding", "Classification"],
|
| 49 |
},
|
| 50 |
{
|
| 51 |
+
label: "Scoring: Micro F1",
|
| 52 |
+
description: "Was the correct choice selected among the options.",
|
| 53 |
},
|
| 54 |
]),
|
| 55 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 56 |
ARCHITECTURE: createTooltipContent("Model Architecture Information:", [
|
| 57 |
{
|
| 58 |
label: "Definition",
|
frontend/src/pages/LeaderboardPage/components/Leaderboard/utils/columnUtils.js
CHANGED
|
@@ -752,56 +752,11 @@ export const createColumns = (
|
|
| 752 |
const evaluationColumns = [
|
| 753 |
{
|
| 754 |
accessorKey: "evaluations.ifeval.normalized_score",
|
| 755 |
-
header: createHeaderCell("
|
| 756 |
cell: ({ row, getValue }) =>
|
| 757 |
-
createScoreCell(getValue, row, "evaluations.
|
| 758 |
size: TABLE_DEFAULTS.COLUMNS.COLUMN_SIZES[
|
| 759 |
-
"evaluations.
|
| 760 |
-
],
|
| 761 |
-
},
|
| 762 |
-
{
|
| 763 |
-
accessorKey: "evaluations.bbh.normalized_score",
|
| 764 |
-
header: createHeaderCell("BBH", COLUMN_TOOLTIPS.BBH),
|
| 765 |
-
cell: ({ row, getValue }) =>
|
| 766 |
-
createScoreCell(getValue, row, "evaluations.bbh.normalized_score"),
|
| 767 |
-
size: TABLE_DEFAULTS.COLUMNS.COLUMN_SIZES[
|
| 768 |
-
"evaluations.bbh.normalized_score"
|
| 769 |
-
],
|
| 770 |
-
},
|
| 771 |
-
{
|
| 772 |
-
accessorKey: "evaluations.math.normalized_score",
|
| 773 |
-
header: createHeaderCell("MATH", COLUMN_TOOLTIPS.MATH),
|
| 774 |
-
cell: ({ row, getValue }) =>
|
| 775 |
-
createScoreCell(getValue, row, "evaluations.math.normalized_score"),
|
| 776 |
-
size: TABLE_DEFAULTS.COLUMNS.COLUMN_SIZES[
|
| 777 |
-
"evaluations.math.normalized_score"
|
| 778 |
-
],
|
| 779 |
-
},
|
| 780 |
-
{
|
| 781 |
-
accessorKey: "evaluations.gpqa.normalized_score",
|
| 782 |
-
header: createHeaderCell("GPQA", COLUMN_TOOLTIPS.GPQA),
|
| 783 |
-
cell: ({ row, getValue }) =>
|
| 784 |
-
createScoreCell(getValue, row, "evaluations.gpqa.normalized_score"),
|
| 785 |
-
size: TABLE_DEFAULTS.COLUMNS.COLUMN_SIZES[
|
| 786 |
-
"evaluations.gpqa.normalized_score"
|
| 787 |
-
],
|
| 788 |
-
},
|
| 789 |
-
{
|
| 790 |
-
accessorKey: "evaluations.musr.normalized_score",
|
| 791 |
-
header: createHeaderCell("MUSR", COLUMN_TOOLTIPS.MUSR),
|
| 792 |
-
cell: ({ row, getValue }) =>
|
| 793 |
-
createScoreCell(getValue, row, "evaluations.musr.normalized_score"),
|
| 794 |
-
size: TABLE_DEFAULTS.COLUMNS.COLUMN_SIZES[
|
| 795 |
-
"evaluations.musr.normalized_score"
|
| 796 |
-
],
|
| 797 |
-
},
|
| 798 |
-
{
|
| 799 |
-
accessorKey: "evaluations.mmlu_pro.normalized_score",
|
| 800 |
-
header: createHeaderCell("MMLU-PRO", COLUMN_TOOLTIPS.MMLU_PRO),
|
| 801 |
-
cell: ({ row, getValue }) =>
|
| 802 |
-
createScoreCell(getValue, row, "evaluations.mmlu_pro.normalized_score"),
|
| 803 |
-
size: TABLE_DEFAULTS.COLUMNS.COLUMN_SIZES[
|
| 804 |
-
"evaluations.mmlu_pro.normalized_score"
|
| 805 |
],
|
| 806 |
},
|
| 807 |
];
|
|
|
|
| 752 |
const evaluationColumns = [
|
| 753 |
{
|
| 754 |
accessorKey: "evaluations.ifeval.normalized_score",
|
| 755 |
+
header: createHeaderCell("MultiFin", COLUMN_TOOLTIPS.IFEVAL),
|
| 756 |
cell: ({ row, getValue }) =>
|
| 757 |
+
createScoreCell(getValue, row, "evaluations.multifin.normalized_score"),
|
| 758 |
size: TABLE_DEFAULTS.COLUMNS.COLUMN_SIZES[
|
| 759 |
+
"evaluations.multifin.normalized_score"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 760 |
],
|
| 761 |
},
|
| 762 |
];
|
frontend/src/pages/QuotePage/QuotePage.js
CHANGED
|
@@ -60,7 +60,7 @@ const priorWork = [
|
|
| 60 |
|
| 61 |
const benchmarks = [
|
| 62 |
{
|
| 63 |
-
title: "
|
| 64 |
authors: "Zhou et al.",
|
| 65 |
citation: `@misc{zhou2023instructionfollowingevaluationlargelanguage,
|
| 66 |
title={Instruction-Following Evaluation for Large Language Models},
|
|
|
|
| 60 |
|
| 61 |
const benchmarks = [
|
| 62 |
{
|
| 63 |
+
title: "MultiFin: Instruction-Following Evaluation",
|
| 64 |
authors: "Zhou et al.",
|
| 65 |
citation: `@misc{zhou2023instructionfollowingevaluationlargelanguage,
|
| 66 |
title={Instruction-Following Evaluation for Large Language Models},
|