|
--- |
|
license: apache-2.0 |
|
datasets: |
|
- tatsu-lab/alpaca |
|
language: |
|
- en |
|
--- |
|
# Model Card for Flan-Alpaca-GPT4-base-3k |
|
|
|
This model was obtained by fine-tuning the `google/flan-t5-base` model on the tatsu-lab/alpaca dataset with the max_source_length option set to 3048. The instructions at the following repository were used for fine-tuning: https://github.com/declare-lab/flan-alpaca |
|
The goal of this model was a learning exercise to determine if setting a higher max_source_length resulted in the model interpreting larger prompts during inference. |
|
|
|
### Model Description |
|
|
|
- **Language(s) (NLP)**: English |
|
- **Finetuned from model:** google/flan-t5-base |
|
|
|
## How to use |
|
|
|
```python |
|
from transformers import pipeline |
|
|
|
prompt = "Write an email about an alpaca that likes flan" |
|
model = pipeline(model="evolveon/flan-alpaca-gpt4-base-3k") |
|
model(prompt, max_length=3048, do_sample=True) |
|
|
|
# Dear AlpacaFriend, |
|
# My name is Alpaca and I'm 10 years old. |
|
# I'm excited to announce that I'm a big fan of flan! |
|
# We like to eat it as a snack and I believe that it can help with our overall growth. |
|
# I'd love to hear your feedback on this idea. |
|
# Have a great day! |
|
# Best, AL Paca |
|
``` |