tttoaster commited on
Commit
29fa0ea
1 Parent(s): e5f747b

Update constants.py

Browse files
Files changed (1) hide show
  1. constants.py +13 -2
constants.py CHANGED
@@ -39,11 +39,22 @@ SUBMIT_INTRODUCTION = """# Submit Precautions
39
  1. Attain JSON file from our [github repository](https://github.com/AILab-CVC/SEED-Bench) after evaluation. For example, you can obtain InstructBLIP's JSON file as results/results.json after running
40
  ```shell
41
  python eval.py --model instruct_blip --anno_path SEED-Bench.json --output-dir results
42
- ```
43
- 2. If you want to revise a model, please ensure 'Revision Model Name' align with what's in the leaderboard. For example, if you want to modify InstructBLIP's evaluation result, you need to fill in 'InstructBLIP' in 'Revision Model Name'.
44
  3. Please ensure the right link for each submission. Everyone could go to the model's repository through the model name in the leaderboard.
45
  4. If you don't want to evaluate all dimensions, not evaluated dimension performance, and its corresponding average performance will be set to 0.
46
  5. After clicking 'Submit Eval', you can click 'Refresh' to obtain the latest leaderboard.
 
 
 
 
 
 
 
 
 
 
 
47
  """
48
 
49
  TABLE_INTRODUCTION = """In the table below, we summarize each task performance of all the models.
 
39
  1. Attain JSON file from our [github repository](https://github.com/AILab-CVC/SEED-Bench) after evaluation. For example, you can obtain InstructBLIP's JSON file as results/results.json after running
40
  ```shell
41
  python eval.py --model instruct_blip --anno_path SEED-Bench.json --output-dir results
42
+ ```
43
+ 2. If you want to revise a model, please ensure 'Model Name Revision' align with what's in the leaderboard. For example, if you want to modify InstructBLIP's evaluation result, you need to fill in 'InstructBLIP' in 'Revision Model Name'.
44
  3. Please ensure the right link for each submission. Everyone could go to the model's repository through the model name in the leaderboard.
45
  4. If you don't want to evaluate all dimensions, not evaluated dimension performance, and its corresponding average performance will be set to 0.
46
  5. After clicking 'Submit Eval', you can click 'Refresh' to obtain the latest leaderboard.
47
+
48
+ ## Submit Example
49
+ For example, if you want to revise InstructBLIP's performance in the leaderboard, you need to:
50
+ (1). Fill in 'InstructBLIP' in 'Revision Model Name'.
51
+ (2). Select 'ImageLLM' in 'Model Type'.
52
+ (3). Fill in 'https://github.com/salesforce/LAVIS' in 'Model Link'.
53
+ (4). Select 'Flan-T5-XL' in 'LLM Type'.
54
+ (5). Select 'All' in 'Evaluation Dimension'.
55
+ (6). Upload results.json.
56
+ (7). Click the 'Submit Eval' button.
57
+ (8). Click 'Refresh' to obtain the uploaded leaderboard.
58
  """
59
 
60
  TABLE_INTRODUCTION = """In the table below, we summarize each task performance of all the models.