TITLE = '
Open Multilingual LLM Evaluation Leaderboard
'
INTRO_TEXT = f"""
## About
This leaderboard shows the performance of pretrained models in 29 languages on four benchmarks:
- AI2 Reasoning Challenge (25-shot)
- HellaSwag (10-shot)
- MMLU (5-shot)
- TruthfulQA (0-shot)
The evaluation data was translated into 29 languages using ChatGPT.
"""
HOW_TO = f"""
## How to list your model performance on this leaderboard:
Send an email with title [Open mLLM Loaderboard] to vietl@uoregon.edu with the huggingface's model name.
We will run your model on the four benchmarks and add it to the leaderboard.
"""
CITATION = f"""
## Citation
```
@misc{{lai2023openllmbenchmark,
author = {{Viet Lai and Nghia Trung Ngo and Amir Pouran Ben Veyseh and Franck Dernoncourt and Thien Huu Nguyen}},
title={{Open Multilingual LLM Evaluation Leaderboard}},
year={{2023}}
}}
```
"""