evolveon's picture
Update README.md
ca4f70c
---
license: apache-2.0
datasets:
- tatsu-lab/alpaca
language:
- en
---
# Model Card for Flan-Alpaca-GPT4-base-3k
This model was obtained by fine-tuning the `google/flan-t5-base` model on the tatsu-lab/alpaca dataset with the max_source_length option set to 3048. The instructions at the following repository were used for fine-tuning: https://github.com/declare-lab/flan-alpaca
The goal of this model was a learning exercise to determine if setting a higher max_source_length resulted in the model interpreting larger prompts during inference.
### Model Description
- **Language(s) (NLP)**: English
- **Finetuned from model:** google/flan-t5-base
## How to use
```python
from transformers import pipeline
prompt = "Write an email about an alpaca that likes flan"
model = pipeline(model="evolveon/flan-alpaca-gpt4-base-3k")
model(prompt, max_length=3048, do_sample=True)
# Dear AlpacaFriend,
# My name is Alpaca and I'm 10 years old.
# I'm excited to announce that I'm a big fan of flan!
# We like to eat it as a snack and I believe that it can help with our overall growth.
# I'd love to hear your feedback on this idea.
# Have a great day!
# Best, AL Paca
```