language: en
license: mit
tags:
- gpt2
- question-answering
- fine-tuned
Fine-tuned Orion Model for Question Answering
This model is a fine-tuned version of the cuba6112/orion model, specialized for question answering tasks.
Model description
The model was fine-tuned on a custom dataset of question-answer pairs. It can generate answers to various questions across different topics.
Intended uses & limitations
This model is intended for generating answers to general knowledge questions. It should not be used for sensitive or critical applications without human oversight.
Training data
The model was fine-tuned on a custom dataset of question-answer pairs. The dataset covers various topics including geography, history, science, and more.
Training procedure
The model was fine-tuned using the Hugging Face Transformers library. We used a causal language modeling objective with the following hyperparameters:
- Number of epochs: 3
- Batch size: 4
- Learning rate: Default from TrainingArguments
Evaluation results
Evaluation metrics and results will be added soon.