OFA-OCR / README_EncouragingLoss.md
JustinLin610's picture
first commit
ee21b96
|
raw
history blame
2.1 kB

Finetuning with Encouraging Loss (EL)

Below we provide methods for finetuning with label smoothed encouraging loss proposed in Well-classified Examples are Underestimated in Classification with Deep Neural Networks on different downstream tasks. The implementation is in label_smoothed_encouraging_loss.py. You can set the --criterion to adjust_label_smoothed_encouraging_loss to use it. This criterion has a hyper-parameter --log-end. --log-end < 1 results in a approximated and conservative version of the full encouraging loss. A high log_end will more strongly weaken the gradient vanishing, enhance the modeling of the data, and increase the growth rate of the margin, but it will also bring a larger gradient norm, which will bring challenges to the existing optimization system. We recommend higher log_end for cases with higher performance, and 0.75 or 0.5 as your first try.

Image Captioning

We provide procedures for image captioning with EL below. The preprocessing is identical to default setting.

Finetuning

We propose two scripts for stage1.

cd run_scripts/caption
nohup sh train_caption_stage1_el.sh > train_stage1_el.out &  # stage 1, train with encouraging loss, expected cider 1.403
nohup sh train_caption_stage1_el_db.sh > train_stage1_el.out &  # stage 1, train with encouraging loss, and drop best examples, expected cider 1.404

Referring Expression Comprehension

We provide procedures for image captioning with EL below. The preprocessing is identical to default setting.

Finetuning
cd run_scripts/refcoco
nohup sh train_refcoco_el.sh > train_refcoco_el.out &  # finetune for refcoco
nohup sh train_refcocoplus_el.sh > train_refcocoplus_el.out &  # finetune for refcoco+
nohup sh train_refcocog_el.sh > train_refcocog_el.out &  # finetune for refcocog
Evaluation is also the same as the default setting.