File size: 669 Bytes
75807f1
637332d
75807f1
 
637332d
 
 
 
 
 
 
 
 
2a4f976
637332d
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
---
language: en
license: apache-2.0
---

### Description
Adaptation of the [flan-t5-small](https://huggingface.co/google/flan-t5-small) weights to make it compatible with the [FAT5](https://github.com/catie-aq/flashT5) framework (Flash Attention T5).  
This adaptation should enable the user to efficiently continue the pre-training of the flan-t5 to adapt it to more recent data, or to specialize it in a specific domain, for example.  

### Usage
```
from transformers import AutoModel, AutoTokenizer
model = AutoModel.from_pretrained("CATIE-AQ/FAT5-small-flan-en", trust_remote_code=True)
tokenizer = AutoTokenizer.from_pretrained("CATIE-AQ/FAT5-small-flan-en")
```