File size: 1,653 Bytes
4e9ddf9
 
 
 
 
01bc423
4e9ddf9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
b19efcb
4e9ddf9
 
 
 
 
01bc423
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
# This YAML dictionary will expand into 20 (models) x 4 (tasks) = 80 job commands,
# where {{ model }} and {{ task }} are filled in with all possible combinations.
# {{ gpu }} is defined in `hosts.yaml`, and will be filled in when Pegasus
# determines the specific node and gpu the generated job command will run on.
- command:
    - docker exec leaderboard{{ gpu }} python scripts/benchmark.py --input-file sharegpt/sg_90k_part1_html_cleaned_lang_first_sampled_sorted.json --model-path {{ model }} --task {{ task }} --batch-size {{ batch_size }}
  model:
    - /data/leaderboard/weights/metaai/llama-7B
    - /data/leaderboard/weights/metaai/llama-13B
    - /data/leaderboard/weights/lmsys/vicuna-7B
    - /data/leaderboard/weights/lmsys/vicuna-13B
    - /data/leaderboard/weights/tatsu-lab/alpaca-7B
    - /data/leaderboard/weights/BAIR/koala-7b
    - /data/leaderboard/weights/BAIR/koala-13b
    - /data/leaderboard/weights/BlinkDL/RWKV-4-Raven-7B-v12-Eng98%-Other2%-20230521-ctx8192.pth
    - camel-ai/CAMEL-13B-Combined-Data
    - databricks/dolly-v2-12b
    - FreedomIntelligence/phoenix-inst-chat-7b
    - h2oai/h2ogpt-gm-oasst1-en-2048-open-llama-7b-preview-300bt-v2
    - lmsys/fastchat-t5-3b-v1.0
    - Neutralzz/BiLLa-7B-SFT
    - nomic-ai/gpt4all-13b-snoozy
    - openaccess-ai-collective/manticore-13b-chat-pyg
    - OpenAssistant/oasst-sft-1-pythia-12b
    - project-baize/baize-v2-7B
    - StabilityAI/stablelm-tuned-alpha-7b
    - togethercomputer/RedPajama-INCITE-7B-Chat
    - Salesforce/xgen-7b-8k-inst
  task:
    - chat
    - chat-concise
    - instruct
    - instruct-concise
  batch_size:
    - 1
    - 2
    - 4
    - 8
    - 16