File size: 1,171 Bytes
4bc6a88
 
ca4f70c
 
 
 
4bc6a88
ca4f70c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
---
license: apache-2.0
datasets:
- tatsu-lab/alpaca
language:
- en
---
# Model Card for Flan-Alpaca-GPT4-base-3k

This model was obtained by fine-tuning the `google/flan-t5-base` model on the tatsu-lab/alpaca dataset with the max_source_length option set to 3048. The instructions at the following repository were used for fine-tuning: https://github.com/declare-lab/flan-alpaca
The goal of this model was a learning exercise to determine if setting a higher max_source_length resulted in the model interpreting larger prompts during inference. 

### Model Description

- **Language(s) (NLP)**: English
- **Finetuned from model:** google/flan-t5-base

## How to use

```python
from transformers import pipeline

prompt = "Write an email about an alpaca that likes flan"
model = pipeline(model="evolveon/flan-alpaca-gpt4-base-3k")
model(prompt, max_length=3048, do_sample=True)

# Dear AlpacaFriend,
# My name is Alpaca and I'm 10 years old.
# I'm excited to announce that I'm a big fan of flan!
# We like to eat it as a snack and I believe that it can help with our overall growth.
# I'd love to hear your feedback on this idea. 
# Have a great day! 
# Best, AL Paca
```