|
Evaluating Pre-trained Models on Task Datasets
|
|
###############################################
|
|
LAVIS provides pre-trained and finetuned model for off-the-shelf evaluation on task dataset.
|
|
Let's now see an example to evaluate BLIP model on the captioning task, using MSCOCO dataset.
|
|
|
|
.. _prep coco:
|
|
|
|
Preparing Datasets
|
|
******************
|
|
First, let's download the dataset. LAVIS provides `automatic downloading scripts` to help prepare
|
|
most of the public dataset, to download MSCOCO dataset, simply run
|
|
|
|
.. code-block:: bash
|
|
|
|
cd lavis/datasets/download_scripts && python download_coco.py
|
|
|
|
This will put the downloaded dataset at a default cache location ``cache`` used by LAVIS.
|
|
|
|
If you want to use a different cache location, you can specify it by updating ``cache_root`` in ``lavis/configs/default.yaml``.
|
|
|
|
If you have a local copy of the dataset, it is recommended to create a symlink from the cache location to the local copy, e.g.
|
|
|
|
.. code-block:: bash
|
|
|
|
ln -s /path/to/local/coco cache/coco
|
|
|
|
Evaluating pre-trained models
|
|
******************************
|
|
|
|
To evaluate pre-trained model, simply run
|
|
|
|
.. code-block:: bash
|
|
|
|
bash run_scripts/blip/eval/eval_coco_cap.sh
|
|
|
|
Or to evaluate a large model:
|
|
|
|
.. code-block:: bash
|
|
|
|
bash run_scripts/blip/eval/eval_coco_cap_large.sh
|
|
|