# This YAML dictionary will expand into 20 (models) x 4 (tasks) = 80 job commands, # where {{ model }} and {{ task }} are filled in with all possible combinations. # {{ gpu }} is defined in `hosts.yaml`, and will be filled in when Pegasus # determines the specific node and gpu the generated job command will run on. - command: - docker exec leaderboard{{ gpu }} python scripts/benchmark.py --input-file sharegpt/sg_90k_part1_html_cleaned_lang_first_sampled_sorted.json --model-path {{ model }} --task {{ task }} --batch-size {{ batch_size }} model: - /data/leaderboard/weights/metaai/llama-7B - /data/leaderboard/weights/metaai/llama-13B - /data/leaderboard/weights/lmsys/vicuna-7B - /data/leaderboard/weights/lmsys/vicuna-13B - /data/leaderboard/weights/tatsu-lab/alpaca-7B - /data/leaderboard/weights/BAIR/koala-7b - /data/leaderboard/weights/BAIR/koala-13b - /data/leaderboard/weights/BlinkDL/RWKV-4-Raven-7B-v12-Eng98%-Other2%-20230521-ctx8192.pth - camel-ai/CAMEL-13B-Combined-Data - databricks/dolly-v2-12b - FreedomIntelligence/phoenix-inst-chat-7b - h2oai/h2ogpt-gm-oasst1-en-2048-open-llama-7b-preview-300bt-v2 - lmsys/fastchat-t5-3b-v1.0 - Neutralzz/BiLLa-7B-SFT - nomic-ai/gpt4all-13b-snoozy - openaccess-ai-collective/manticore-13b-chat-pyg - OpenAssistant/oasst-sft-1-pythia-12b - project-baize/baize-v2-7B - StabilityAI/stablelm-tuned-alpha-7b - togethercomputer/RedPajama-INCITE-7B-Chat - Salesforce/xgen-7b-8k-inst task: - chat - chat-concise - instruct - instruct-concise batch_size: - 1 - 2 - 4 - 8 - 16