python trainer.py \
  --dataset_dir=/home/wuwanjie/projects/nlp_algo/app/data/extract \
  --max_seq_length=512 \
  --output_dir=/data/sdv1/wuwanjie/outputs/extract \
  --model_name_or_path=/data/sdv1/wuwanjie/transformers/bert-base-chinese \
  --epoch=30 \
  --batch_size=128 \
  --learning_rate=3e-5 \
  --max_grad_norm=1.0 \
  --dropout=0.1 \
  --eval_count_per_epoch=3 \
  --log_iter_num=10 \
  --warmup_proportion=0.2 \
  --gradient_accumulation_steps=1 \
  --weight_decay=0.01 \
  --do_train \
  --do_predict \
  --model_type=bert_crf \
#  --add_bilstm \
#  --fp16 \
#  --fp16_opt_level=O1
