|
--- |
|
language: |
|
- en |
|
- de |
|
- fr |
|
- zh |
|
- pt |
|
- nl |
|
- ru |
|
- ko |
|
- it |
|
- es |
|
license: cc-by-nc-4.0 |
|
metrics: |
|
- comet |
|
pipeline_tag: translation |
|
--- |
|
# Model Card for TowerInstruct-WMT24-Chat-7B |
|
|
|
## Model Details |
|
|
|
### Model Description |
|
|
|
TowerInstruct-WMT24-Chat-7B is a language model that results from fine-tuning TowerBase on TowerBlocks and the WMT24 Chat MT Shared task training set. |
|
|
|
TowerInstruct-WMT24-Chat-7B was the best submission of the shared task, winning on all 10 language pairs according to human evaluation (see the task's findings paper [here](https://aclanthology.org/2024.wmt-1.59.pdf)). |
|
|
|
It is specifically tailoured for context-aware translation of customer support chats. |
|
|
|
Check out our [paper](https://aclanthology.org/2024.wmt-1.100/) for more details and information on training and data. |
|
|
|
Information on model usage, out-of-scope usages, risks, etc... are the same as the model cards of the TowerInstruct models. |
|
|
|
## Citation |
|
|
|
```bibtex |
|
@inproceedings{pombal2024improving, |
|
title={Improving Context Usage for Translating Bilingual Customer Support Chat with Large Language Models}, |
|
author={Pombal, Jos{\'e} and Agrawal, Sweta and Martins, Andr{\'e} FT}, |
|
booktitle={Proceedings of the Ninth Conference on Machine Translation}, |
|
pages={993--1003}, |
|
year={2024} |
|
} |
|
``` |
|
|
|
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl) |
|
|