| _config_info: | |
| create_time: '2023-03-14 23:35:21' | |
| use_default_base_config: true | |
| config_filepath: | |
| - /data/tzhu/MRC4IE/conf/mrc.yaml | |
| task_type: MrcQaTask | |
| task_name: RobertaBase_data20230314v2 | |
| comment: GlobalPointer with RoPE | |
| output_dir: outputs | |
| task_dir: outputs/RobertaBase_data20230314v2 | |
| dump_cache_dir: outputs/RobertaBase_data20230314v2/cache | |
| regenerate_cache: true | |
| data_dir: resources/Merged/20230314-v2 | |
| train_filepath: resources/Merged/20230314-v2/train.jsonl | |
| dev_filepath: resources/Merged/20230314-v2/dev.jsonl | |
| test_filepath: resources/Merged/20230314-v2/test.jsonl | |
| random_seed: 1227 | |
| num_epochs: 20 | |
| num_steps: -1 | |
| warmup_proportion: 0.1 | |
| epoch_patience: 10 | |
| step_patience: 5000 | |
| batch_size: 64 | |
| learning_rate: 2.0e-05 | |
| max_grad_norm: -1.0 | |
| skip_train: false | |
| debug_mode: false | |
| grad_accum_steps: 1 | |
| resumed_training_path: null | |
| step_eval_interval: 20000 | |
| epoch_eval_interval: 1 | |
| eval_on_data: | |
| - dev | |
| select_best_on_data: dev | |
| select_best_by_key: metric | |
| best_metric_field: micro.f1 | |
| save_every_ckpt: false | |
| save_best_ckpt: true | |
| final_eval_on_test: true | |
| main_process_logging: true | |
| max_seq_len: 512 | |
| plm_dir: hfl/chinese-roberta-wwm-ext | |
| train_batch_size: 16 | |
| eval_batch_size: 32 | |
| other_learning_rate: 2.0e-05 | |
| dropout: 0.3 | |
| biaffine_size: 512 | |
| include_package: | |
| - src.task | |
| use_default_base_config: true | |