|
--- |
|
license: cc-by-4.0 |
|
--- |
|
|
|
# GenRead (MergeDPR): FiD model trained on NQ |
|
|
|
-- This is the model checkpoint of GenRead [2], based on the T5-3B and trained on the NQ dataset [1]. |
|
|
|
-- Hyperparameters: 8 x 80GB A100 GPUs; batch size 16; AdamW; LR 3e-5; best dev at 19000 steps. |
|
|
|
References: |
|
|
|
[1] Natural Questions: A Benchmark for Question Answering Research. TACL 2019. |
|
|
|
[2] Generate rather than Retrieve: Large Language Models are Strong Context Generators. arXiv 2022 |
|
|
|
## Model performance |
|
|
|
We evaluate it on the TriviaQA dataset, the EM score is 54.38. |
|
|
|
|
|
<a href="https://huggingface.co/exbert/?model=bert-base-uncased"> |
|
<img width="300px" src="https://cdn-media.huggingface.co/exbert/button.png"> |
|
</a> |
|
--- |
|
license: cc-by-4.0 |
|
--- |
|
|
|
|
|
--- |
|
license: cc-by-4.0 |
|
--- |
|
|