futuretod-base-v1.0 / README.md
AndrewZeng's picture
update readme
293eb5f
---
license: apache-2.0
---
# FutureTOD: Teaching Future Knowledge to Pre-trained Language Model for Task-Oriented Dialogue
We present our dialogue-pertaining model, FutureTOD, which distills future knowledge into the representation of the previous dialogue context using a self-training framework. Extensive experiments on diverse downstream dialogue tasks demonstrate the effectiveness of our model, especially its generalization, robustness, and ability to learn discriminative dialogue representations.
[This paper](https://arxiv.org/abs/2306.10315) has been accepted at the ACL 2023 Main Conference.
## Usage
We release our futuretod-base-v1.0 model here. You can use this model for downstream TOD tasks follow instructions in [FutureTOD](https://github.com/Zeng-WH/FutureTOD).
## Quotation
If you find our work helpful, please consider quoting the following papers.
```
@article{zeng2023futuretod,
title={FutureTOD: Teaching Future Knowledge to Pre-trained Language Model for Task-Oriented Dialogue},
author={Zeng, Weihao and He, Keqing and Wang, Yejie and Zeng, Chen and Wang, Jingang and Xian, Yunsen and Xu, Weiran},
journal={arXiv preprint arXiv:2306.10315},
year={2023}
}
```