eng2french-t5-small / README.md
dmedhi's picture
Update README.md
8441bfc
---
library_name: peft
base_model: t5-small
license: apache-2.0
datasets:
- opus100
tags:
- translation
- safetensors
- transformers
language:
- en
- fr
---
# Model Card for Model ID
A language translation model fine-tuned on **opus100** dataset for *English to French* translation.
## Model Description
- **Model type:** Language Model
- **Language(s) (NLP):** English, French
- **License:** Apache 2.0
- **Finetuned from model:** [T5-small](https://huggingface.co/t5-small)
## Uses
The model is intended to use for English to French translation related tasks.
## How to Get Started with the Model
Install necessary libraries
```
pip install transformers peft accelerate
```
Use the code below to get started with the model.
```python
from peft import PeftModel, PeftConfig
from transformers import AutoModelForSeq2SeqLM, AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained("dmedhi/eng2french-t5-small")
model = AutoModelForSeq2SeqLM.from_pretrained("t5-small")
model = PeftModel.from_pretrained(model, "dmedhi/eng2french-t5-small")
context = tokenizer(["Do you want coffee?"], return_tensors='pt')
output = model.generate(**context)
result = tokenizer.decode(output[0], skip_special_tokens=True)
print(result)
# Output
# Tu veux du café?
```
## Training Details
### Training Data
- Dataset used: [Opus100](https://huggingface.co/datasets/opus100)
- Subset: "en-fr"
## Evaluation
- global_step=5000
- training_loss=1.295289501953125
#### Metrics
- train_runtime = 1672.4371
- train_samples_per_second = 23.917
- train_steps_per_second = 2.99
- total_flos = 685071170273280.0
- train_loss = 1.295289501953125
- epoch = 20.0
## Compute Instance
- Google Colab - T4 GPU (Free)
### Framework versions
- PEFT 0.7.1