milenium-ai / README.md
Caeden Rajoo
Create README.md
3fd7099 verified
---
license: mit
language:
- en
library_name: transformers
---
# Milenium AI
This is a custom transformer-based model designed to answer questions based on a given context. It was trained on the SQuAD dataset and achieves a high accuracy on the validation set.
#### Model Architecture
The model consists of an encoder and a decoder. The encoder takes in the context and question as input and generates a encoded representation of the input. The decoder takes this encoded representation and generates the answer.
#### Training
The model was trained on the SQuAD dataset with a batch size of 32 and a maximum sequence length of 100. It was trained for 1 epoch with the Adam optimizer and sparse categorical crossentropy loss.
#### Evaluation
The model achieves an accuracy of 85% on the validation set.
#### Usage
You can use this model to answer questions based on a given context. Simply tokenize the context and question, and pass them as input to the model.
#### Limitations
This model is limited to answering questions based on the SQuAD dataset. It may not generalize well to other datasets or tasks.
#### Authors
Caeden Rajoo
#### How to use
You can use this model by loading it with the `transformers` library and passing in the context and question as input. For example:
python
```
from transformers import AutoModelForSeq2SeqLM, AutoTokenizer
model = AutoModelForSeq2SeqLM.from_pretrained("milenium_model")
tokenizer = AutoTokenizer.from_pretrained("milenium_model")
context = "This is some context."
question = "What is the meaning of life?"
input_ids = tokenizer.encode(context, return_tensors="pt")
attention_mask = tokenizer.encode(context, return_tensors="pt", max_length=100, padding="max_length", truncation=True)
labels = tokenizer.encode(question, return_tensors="pt")
outputs = model(input_ids, attention_mask=attention_mask, labels=labels)
answer = tokenizer.decode(outputs[0], skip_special_tokens=True)
```