# Your leaderboard name TITLE = """

AIR-Bench

""" # What does your leaderboard evaluate? INTRODUCTION_TEXT = """ AIR-Bench: Automated Heterogeneous Information Retrieval Benchmark """ # Which evaluations are you running? how can people reproduce what you have? BENCHMARKS_TEXT = f""" ## How it works Check more information at [our GitHub repo](https://github.com/AIR-Bench/AIR-Bench) """ EVALUATION_QUEUE_TEXT = """ ## Steps for submit to AIR-Bench 1. Install AIR-Bench ```bash # Clone the repo git clone https://github.com/AIR-Bench/AIR-Bench.git # Install the package cd AIR-Bench pip install . ``` 2. Run the evaluation script ```bash cd AIR-Bench/scripts # Run all tasks python run_AIR-Bench.py \ --output_dir ./search_results \ --encoder BAAI/bge-m3 \ --encoder_link https://huggingface.co/BAAI/bge-m3 \ --reranker BAAI/bge-reranker-v2-m3 \ --reranker_link https://huggingface.co/BAAI/bge-reranker-v2-m3 \ --search_top_k 1000 \ --rerank_top_k 100 \ --max_query_length 512 \ --max_passage_length 512 \ --batch_size 512 \ --pooling_method cls \ --normalize_embeddings True \ --use_fp16 True \ --add_instruction False \ --overwrite False # Run the tasks in the specified task type python run_AIR-Bench.py \ --task_type long-doc \ --output_dir ./search_results \ --encoder BAAI/bge-m3 \ --encoder_link https://huggingface.co/BAAI/bge-m3 \ --reranker BAAI/bge-reranker-v2-m3 \ --reranker_link https://huggingface.co/BAAI/bge-reranker-v2-m3 \ --search_top_k 1000 \ --rerank_top_k 100 \ --max_query_length 512 \ --max_passage_length 512 \ --batch_size 512 \ --pooling_method cls \ --normalize_embeddings True \ --use_fp16 True \ --add_instruction False \ --overwrite False # Run the tasks in the specified task type and domains python run_AIR-Bench.py \ --task_type long-doc \ --domain arxiv book \ --output_dir ./search_results \ --encoder BAAI/bge-m3 \ --encoder_link https://huggingface.co/BAAI/bge-m3 \ --reranker BAAI/bge-reranker-v2-m3 \ --reranker_link https://huggingface.co/BAAI/bge-reranker-v2-m3 \ --search_top_k 1000 \ --rerank_top_k 100 \ --max_query_length 512 \ --max_passage_length 512 \ --batch_size 512 \ --pooling_method cls \ --normalize_embeddings True \ --use_fp16 True \ --add_instruction False \ --overwrite False # Run the tasks in the specified languages python run_AIR-Bench.py \ --language en \ --output_dir ./search_results \ --encoder BAAI/bge-m3 \ --encoder_link https://huggingface.co/BAAI/bge-m3 \ --reranker BAAI/bge-reranker-v2-m3 \ --reranker_link https://huggingface.co/BAAI/bge-reranker-v2-m3 \ --search_top_k 1000 \ --rerank_top_k 100 \ --max_query_length 512 \ --max_passage_length 512 \ --batch_size 512 \ --pooling_method cls \ --normalize_embeddings True \ --use_fp16 True \ --add_instruction False \ --overwrite False # Run the tasks in the specified task type, domains, and languages python run_AIR-Bench.py \ --task_type qa \ --domains wiki web \ --languages en \ --output_dir ./search_results \ --encoder BAAI/bge-m3 \ --encoder_link https://huggingface.co/BAAI/bge-m3 \ --reranker BAAI/bge-reranker-v2-m3 \ --reranker_link https://huggingface.co/BAAI/bge-reranker-v2-m3 \ --search_top_k 1000 \ --rerank_top_k 100 \ --max_query_length 512 \ --max_passage_length 512 \ --batch_size 512 \ --pooling_method cls \ --normalize_embeddings True \ --use_fp16 True \ --add_instruction False \ --overwrite False ``` 3. Package the search results. ```bash python zip_results.py \ --results_path search_results/bge-m3 \ --save_path search_results/zipped_results ``` 4. Upload the `.zip` file on this page and fill in the model information. 5. Congratulation! Your results will be shown on the leaderboard in up to one hour. """ CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results" CITATION_BUTTON_TEXT = r""" """