File size: 907 Bytes
5953ef9 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 |
# task
task_type: MrcQaTask
task_name: Mirror_RobertaBaseWwm_Cons_MsraMrc
comment: 'GlobalPointer with RoPE'
# data preprocessing
max_seq_len: 512
debug_mode: false
mode: cons
# filepaths
plm_dir: hfl/chinese-roberta-wwm-ext
data_dir: resources/NER/msra/mrc
output_dir: outputs
task_dir: ${output_dir}/${task_name}
train_filepath: ${data_dir}/train.jsonl
dev_filepath: ${data_dir}/test.jsonl
test_filepath: ${data_dir}/test.jsonl
dump_cache_dir: ${task_dir}/cache
regenerate_cache: true
# training
random_seed: 1227
eval_on_data: [dev]
select_best_on_data: dev
select_best_by_key: metric
best_metric_field: micro.f1
final_eval_on_test: true
warmup_proportion: 0.1
step_eval_interval: 20000
step_patience: -1
num_epochs: 5
epoch_patience: 5
train_batch_size: 32
eval_batch_size: 64
learning_rate: !!float 5e-5
other_learning_rate: !!float 1e-4
max_grad_norm: 1.0
# model
dropout: 0.3
biaffine_size: 512
|