--- license: apache-2.0 datasets: - qiaojin/PubMedQA language: - en metrics: - accuracy base_model: - google-bert/bert-base-uncased --- # GEM_PubMedQA Model Card This model card provides an overview of the GEM_PubMedQA model, a finetuned implementation of the GEM architecture designed for the PubMedQA dataset. ## Purpose The GEM_PubMedQA model was developed to assess the performance of the GEM architecture on domain-specific datasets, with a focus on healthcare. The PubMedQA dataset, a key benchmark in this field, was selected to evaluate its effectiveness. ## Key Details - **License**: Apache-2.0 - **Dataset**: qiaojin/PubMedQA - **Language**: English - **Metrics**: Accuracy: 92.5% - **Base Model**: google-bert/bert-base-uncased ## Model Details The GEM_PubMedQA model is built on the GEM architecture and finetuned from the `google-bert/bert-base-uncased` model using the PubMedQA dataset. The training was performed with the following parameters: - **Number of epochs**: 5 - **Batch size**: 128 - **Learning rate**: 2e-5 - **Maximum sequence length**: 128 - **Gradient accumulation steps**: 2 - **Cluster size**: 256 - **Threshold**: 0.65