wenhu commited on
Commit
bd5120e
β€’
1 Parent(s): 900d902

update interface

Browse files
Files changed (1) hide show
  1. utils.py +48 -13
utils.py CHANGED
@@ -24,37 +24,71 @@ CSV_DIR = "./science_leaderboard_submission/results.csv"
24
 
25
  COLUMN_NAMES = MODEL_INFO
26
 
27
- LEADERBORAD_INTRODUCTION = """# TheoremQA Leaderboard
28
 
29
- *"Which Model is better on STEM QA?"*
30
- πŸ† Welcome to the leaderboard of the **TheoremQA**! 🎦 *A Theorem-driven Question Answering dataset* (**EMNLP 2023**)
31
  <div style="display: flex; flex-wrap: wrap; align-items: center; gap: 10px;">
32
- <a href='https://arxiv.org/abs/2305.12524'><img src='https://img.shields.io/badge/cs.CV-Paper-b31b1b?logo=arxiv&logoColor=red'></a>
33
- <a href='https://github.com/TIGER-AI-Lab/TheoremQA'><img src='https://img.shields.io/badge/Github-Repo-grey?logo=github&logoColor=white'></a>
34
- <a href='https://hits.seeyoufarm.com'><img src='https://hits.seeyoufarm.com/api/count/incr/badge.svg?url=https%3A%2F%2Fhuggingface.co%2Fspaces%2FTIGER-Lab%2FTheoremQA-Leaderboard&count_bg=%23C7C83D&title_bg=%23555555&icon=&icon_color=%23E7E7E7&title=hits&edge_flat=false'></a>
35
  </div>
36
-
37
- We propose the first question-answering dataset driven by STEM theorems. We annotated 800 QA pairs covering 350+ theorems spanning across Math, EE&CS, Physics and Finance. The dataset is collected by human experts with very high quality. We provide the dataset as a new benchmark to test the limit of large language models to apply theorems to solve challenging university-level questions.
38
-
39
- Please follow the instructions in [TheoremQA](https://github.com/TIGER-AI-Lab/TheoremQA) to use.
 
 
 
 
40
  """
41
 
42
  TABLE_INTRODUCTION = """
43
  """
44
 
45
  LEADERBORAD_INFO = """
46
- TheoremQA, a comprehensive benchmark suite for video generative models. We design a comprehensive and hierarchical Evaluation Dimension Suite to decompose "video generation quality" into multiple well-defined dimensions to facilitate fine-grained and objective evaluation. For each dimension and each content category, we carefully design a Prompt Suite as test cases, and sample Generated Videos from a set of video generation models. For each evaluation dimension, we specifically design an Evaluation Method Suite, which uses carefully crafted method or designated pipeline for automatic objective evaluation. We also conduct Human Preference Annotation for the generated videos for each dimension, and show that TheoremQA evaluation results are well aligned with human perceptions. TheoremQA can provide valuable insights from multiple perspectives.
 
 
 
 
 
 
 
 
 
 
 
 
 
47
  """
48
 
49
  CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results"
50
- CITATION_BUTTON_TEXT = r"""@inproceedings{chen2023theoremqa,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
51
  title={Theoremqa: A theorem-driven question answering dataset},
52
  author={Chen, Wenhu and Yin, Ming and Ku, Max and Lu, Pan and Wan, Yixin and Ma, Xueguang and Xu, Jianyu and Wang, Xinyi and Xia, Tony},
53
  booktitle={The 2023 Conference on Empirical Methods in Natural Language Processing},
54
  year={2023}
 
 
 
 
 
 
55
  }"""
56
 
57
- SUBMIT_INTRODUCTION = """# Submit on TheoremQA Leaderboard Introduction
58
 
59
  ## ⚠ Please note that you need to submit the json file with following format:
60
 
@@ -64,6 +98,7 @@ SUBMIT_INTRODUCTION = """# Submit on TheoremQA Leaderboard Introduction
64
  "TheoremQA": 0.5,
65
  "MATH": 0.5,
66
  "GSM": 0.5
 
67
  }
68
  ```
69
  After submitting, you can click the "Refresh" button to see the updated leaderboard(it may takes few seconds).
 
24
 
25
  COLUMN_NAMES = MODEL_INFO
26
 
27
+ LEADERBORAD_INTRODUCTION = """# Science Leaderboard
28
 
29
+ *"Which large language model is the BEST on scinece and engineering?"*
30
+ πŸ† Welcome to the **Science** leaderboard! The leaderboard covers the most popular evaluation for different science subjects including math, phyiscs, biology, chemistry, computer science, finance.
31
  <div style="display: flex; flex-wrap: wrap; align-items: center; gap: 10px;">
 
 
 
32
  </div>
33
+ The evaluation set from the following datasets are being included in the leaderboard.
34
+ <ul>
35
+ <li> MATH: this contains the test set of 5000 questions from American Math contest covering different fields like algebra, calculus, statistics, geometry, linear algebra, number theory.
36
+ <li> GSM8K: this contains the test set of 1320 questions from grade school math word problems. This dataset is mainly covering algebra problems.
37
+ <li> TheoremQA: this contains the test set of 800 questions collected from college-level exams. This covers math, physics, engineering and finance.
38
+ <li> GPQA: this contains the test of 198 questions from college-level dataset GPQA-diamond. This covers many fields like chemistry, genetics, biology, etc.
39
+ </ul>
40
+ <a href='https://hits.seeyoufarm.com'><img src='https://hits.seeyoufarm.com/api/count/incr/badge.svg?url=https%3A%2F%2Fhuggingface.co%2Fspaces%2FTIGER-Lab%2FTheoremQA-Leaderboard&count_bg=%23C7C83D&title_bg=%23555555&icon=&icon_color=%23E7E7E7&title=hits&edge_flat=false'></a>
41
  """
42
 
43
  TABLE_INTRODUCTION = """
44
  """
45
 
46
  LEADERBORAD_INFO = """
47
+ We list the information of the used datasets as follows:<br>
48
+
49
+ MATH: Measuring Mathematical Problem Solving With the MATH Dataset
50
+ <a href='https://arxiv.org/pdf/2103.03874.pdf'><img src='https://img.shields.io/badge/cs.CV-Paper-b31b1b?logo=arxiv&logoColor=red'></a>
51
+ <a href='https://github.com/hendrycks/math'><img src='https://img.shields.io/badge/Github-Repo-grey?logo=github&logoColor=white'></a>
52
+ GSM8K: Training Verifiers to Solve Math Word Problems
53
+ <a href='https://arxiv.org/pdf/2110.14168.pdf'><img src='https://img.shields.io/badge/cs.CV-Paper-b31b1b?logo=arxiv&logoColor=red'></a>
54
+ <a href='https://github.com/openai/grade-school-math'><img src='https://img.shields.io/badge/Github-Repo-grey?logo=github&logoColor=white'></a>
55
+ TheoremQA: TheoremQA: A Theorem-driven Question Answering dataset
56
+ <a href='https://arxiv.org/pdf/2305.12524.pdf'><img src='https://img.shields.io/badge/cs.CV-Paper-b31b1b?logo=arxiv&logoColor=red'></a>
57
+ <a href='https://github.com/TIGER-AI-Lab/TheoremQA'><img src='https://img.shields.io/badge/Github-Repo-grey?logo=github&logoColor=white'></a>
58
+ GPQA: A Graduate-Level Google-Proof Q&A Benchmark
59
+ <a href='https://arxiv.org/pdf/2311.12022.pdf'><img src='https://img.shields.io/badge/cs.CV-Paper-b31b1b?logo=arxiv&logoColor=red'></a>
60
+ <a href='https://github.com/idavidrein/gpqa'><img src='https://img.shields.io/badge/Github-Repo-grey?logo=github&logoColor=white'></a>
61
  """
62
 
63
  CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results"
64
+ CITATION_BUTTON_TEXT = r"""@article{hendrycks2measuring,
65
+ title={Measuring Mathematical Problem Solving With the MATH Dataset},
66
+ author={Hendrycks, Dan and Burns, Collin and Kadavath, Saurav and Arora, Akul and Basart, Steven and Tang, Eric and Song, Dawn and Steinhardt, Jacob},
67
+ journal={Sort},
68
+ volume={2},
69
+ number={4},
70
+ pages={0--6}
71
+ }
72
+ @article{cobbe2021training,
73
+ title={Training verifiers to solve math word problems},
74
+ author={Cobbe, Karl and Kosaraju, Vineet and Bavarian, Mohammad and Chen, Mark and Jun, Heewoo and Kaiser, Lukasz and Plappert, Matthias and Tworek, Jerry and Hilton, Jacob and Nakano, Reiichiro and others},
75
+ journal={arXiv preprint arXiv:2110.14168},
76
+ year={2021}
77
+ }
78
+ @inproceedings{chen2023theoremqa,
79
  title={Theoremqa: A theorem-driven question answering dataset},
80
  author={Chen, Wenhu and Yin, Ming and Ku, Max and Lu, Pan and Wan, Yixin and Ma, Xueguang and Xu, Jianyu and Wang, Xinyi and Xia, Tony},
81
  booktitle={The 2023 Conference on Empirical Methods in Natural Language Processing},
82
  year={2023}
83
+ }
84
+ @article{rein2023gpqa,
85
+ title={Gpqa: A graduate-level google-proof q\&a benchmark},
86
+ author={Rein, David and Hou, Betty Li and Stickland, Asa Cooper and Petty, Jackson and Pang, Richard Yuanzhe and Dirani, Julien and Michael, Julian and Bowman, Samuel R},
87
+ journal={arXiv preprint arXiv:2311.12022},
88
+ year={2023}
89
  }"""
90
 
91
+ SUBMIT_INTRODUCTION = """# Submit on Science Leaderboard Introduction
92
 
93
  ## ⚠ Please note that you need to submit the json file with following format:
94
 
 
98
  "TheoremQA": 0.5,
99
  "MATH": 0.5,
100
  "GSM": 0.5
101
+ "GPQA": 0.5
102
  }
103
  ```
104
  After submitting, you can click the "Refresh" button to see the updated leaderboard(it may takes few seconds).