swetaagrawal's picture
Update README.md
74312aa verified
---
language:
- en
- de
- fr
- zh
- pt
- nl
- ru
- ko
- it
- es
license: cc-by-nc-4.0
metrics:
- comet
pipeline_tag: translation
---
# Model Card for TowerInstruct-WMT24-Chat-7B
## Model Details
### Model Description
TowerInstruct-WMT24-Chat-7B is a language model that results from fine-tuning TowerBase on TowerBlocks and the WMT24 Chat MT Shared task training set.
TowerInstruct-WMT24-Chat-7B was the best submission of the shared task, winning on all 10 language pairs according to human evaluation (see the task's findings paper [here](https://aclanthology.org/2024.wmt-1.59.pdf)).
It is specifically tailoured for context-aware translation of customer support chats.
Check out our [paper](https://aclanthology.org/2024.wmt-1.100/) for more details and information on training and data.
Information on model usage, out-of-scope usages, risks, etc... are the same as the model cards of the TowerInstruct models.
## Citation
```bibtex
@inproceedings{pombal2024improving,
title={Improving Context Usage for Translating Bilingual Customer Support Chat with Large Language Models},
author={Pombal, Jos{\'e} and Agrawal, Sweta and Martins, Andr{\'e} FT},
booktitle={Proceedings of the Ninth Conference on Machine Translation},
pages={993--1003},
year={2024}
}
```
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)