Edit model card

photocatalysisbert-squad2

This model is a fine-tuned version of CambridgeMolecularEngineering/photocatalysisbert on the squad_v2 dataset.

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 3e-05
  • train_batch_size: 1
  • eval_batch_size: 8
  • seed: 0
  • distributed_type: multi-GPU
  • num_devices: 20
  • total_train_batch_size: 20
  • total_eval_batch_size: 160
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 2.0

Framework versions

  • Transformers 4.25.1
  • Pytorch 1.12.0a0+git664058f
  • Datasets 2.7.1
  • Tokenizers 0.12.1

Acknowledgements

This model was trained for the paper "How beneficial is pre-training on a narrow domain-specific corpus for information extraction about photocatalytic water splitting?" by Taketomo Isazawa and Jacqueline M. Cole. J.M.C. is grateful for the BASF/Royal Academy of Engineering Research Chair in Data-Driven Molecular Engineering of Functional Materials, which includes PhD studentship support (for T.I.). This Chair is also partly supported by the Science and Technology Facilities Council. They are also indebted to the Argonne Leadership Computing Facility, which is a DOE Office of Science Facility, for use of its research resources, under contract No. DE-AC02-06CH11357.

Downloads last month
1

Dataset used to train CambridgeMolecularEngineering/photocatalysisbert-squad2