File size: 1,672 Bytes
b1b7cea
6de8c85
256752c
6c45e86
 
2b4633e
 
6c45e86
9e8c8a1
6c45e86
9e8c8a1
 
 
 
 
 
 
 
 
 
 
 
 
 
ff9bc0d
6de8c85
 
 
 
 
 
 
 
 
09e350f
ff9bc0d
09e350f
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
## Natural Don't Know Response Model

Fine-tuned on [Google's T5](https://ai.googleblog.com/2020/02/exploring-transfer-learning-with-t5.html) using a combination of a dependency-rule based data and [Quora Question Pairs(QQP)](https://huggingface.co/nlp/viewer/?dataset=quora) dataset for **Don't Know Response Generation** task.

Additional information about this model:
- Paper : [Saying No is An Art: Contextualized Fallback Responses for
Unanswerable Dialogue Queries](https://arxiv.org/pdf/2012.01873.pdf)
- Github Repo: https://github.com/kaustubhdhole/natural-dont-know

#### How to use
```python
from transformers import T5ForConditionalGeneration, T5Tokenizer
model_name = "ashish-shrivastava/dont-know-response"
model = T5ForConditionalGeneration.from_pretrained(model_name)
tokenizer = T5Tokenizer.from_pretrained(model_name)

input = "Where can I find good Italian food ?"
input_ids = tokenizer.encode(input, return_tensors="pt")
outputs = model.generate(input_ids)
decoded_output = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(decoded_output) # I'm not sure where you can get good quality Italian food.

```

#### Hyperparameters

```
n_epochs = 2
base_LM_model = "T5-base"
max_seq_len = 256
learning_rate = 3e-4
adam_epsilon = 1e-8
train_batch_size = 6
``` 

#### BibTeX entry and citation info

```bibtex
@misc{shrivastava2020saying,
      title={Saying No is An Art: Contextualized Fallback Responses for Unanswerable Dialogue Queries}, 
      author={Ashish Shrivastava and Kaustubh Dhole and Abhinav Bhatt and Sharvani Raghunath},
      year={2020},
      eprint={2012.01873},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}
```