|
accumulate_grad_batches: 1 |
|
adafactor: false |
|
adam_epsilon: 1.0e-08 |
|
attention_dropout: null |
|
cache_dir: null |
|
config_name: '' |
|
deterministic: false |
|
do_eval: false |
|
do_test: false |
|
do_train: true |
|
dropout_rate: null |
|
eval_batch_size: 16 |
|
eval_dataset_path: '' |
|
force_reinit: false |
|
fp16: true |
|
gpus: -1 |
|
gradient_clip_val: 1.0 |
|
learning_rate: 1.0e-05 |
|
lr_scheduler: linear |
|
max_epochs: 10 |
|
max_history: 1 |
|
max_seq_length: 0 |
|
min_delta: 0.0 |
|
model_name_or_path: roberta-large |
|
num_workers: 8 |
|
output_dir: ./checkpoints/roberta-large-faithcritic |
|
overwrite_output_dir: false |
|
pad_to_multiple_of: 8 |
|
patience: 5 |
|
predict_dataset_path: null |
|
save_last: false |
|
seed: 42 |
|
test_dataset_path: null |
|
test_task: FaithCritic |
|
tokenizer_name: null |
|
train_batch_size: 16 |
|
train_dataset_path: '' |
|
train_task: FaithCritic |
|
val_check_interval: 0.5 |
|
warmup_ratio: 0.08 |
|
warmup_steps: 0 |
|
weight_decay: 0.1 |
|
|