pranay-j's picture
Update README.md
23c65b2 verified
---
language:
- en
widget:
- text: "did abraham lincoln write the letter in saving private ryan <sep> In the 1998 war film Saving Private Ryan, General George Marshall (played by Harve Presnell) reads the Bixby letter to his officers before giving the order to find and send home Private James Francis Ryan after Ryan's three brothers died in battle."
example_title: "Bool QA"
license: apache-2.0
base_model: bert-base-uncased
tags:
- generated_from_trainer
datasets:
- google/boolq
metrics:
- accuracy
model-index:
- name: Bert Base Uncased Boolean Question Answer model
results:
- task:
name: Text Classification
type: text-classification
dataset:
name: boolq
type: google/boolq
metrics:
- name: Accuracy
type: accuracy
value: 0.7149847094801223
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# Bert Base Uncased Boolean Question Answer model
This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the boolq dataset.
It achieves the following results on the evaluation set:
- Loss: 0.1993
- Accuracy: 0.7150
## Model description
- **Model type:** Text Classification model
- **Language(s) (NLP):** English
- **License:** Apache 2.0
## Intended uses & limitations
More information needed
## Training and evaluation data
- [Dataset](https://huggingface.co/datasets/google/boolq)
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 64
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 4
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:------:|:----:|:---------------:|:--------:|
| 0.2317 | 0.9966 | 147 | 0.2198 | 0.6569 |
| 0.2 | 2.0 | 295 | 0.2002 | 0.6960 |
| 0.1741 | 2.9966 | 442 | 0.1968 | 0.7122 |
| 0.1469 | 3.9864 | 588 | 0.1993 | 0.7150 |
### Framework versions
- Transformers 4.40.0
- Pytorch 2.2.2+cu121
- Datasets 2.19.0
- Tokenizers 0.19.1