Challenges
Browse files
app.py
CHANGED
@@ -66,6 +66,14 @@ article = """
|
|
66 |
<td align="right">23.8261</td>
|
67 |
</tr>
|
68 |
</tbody></table>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
69 |
<h3>Conclusion and Future Work</h3>
|
70 |
If F1 Score is considered, the results show that there may be no advantage in using domain-specific masked language models to generate Biomedical QA models.
|
71 |
However, the F1 Scores reported for the Biomedical roberta-based models are not far below from those of the general roberta-based model.
|
|
|
66 |
<td align="right">23.8261</td>
|
67 |
</tr>
|
68 |
</tbody></table>
|
69 |
+
|
70 |
+
<h3>Conclusion and Future Work</h3>
|
71 |
+
|
72 |
+
<li>Question Answering is a complex task to understand, as it requires not only pre-processing the inputs, but also post-processing the outputs. Moreover, the metrics used are quite specific.
|
73 |
+
<li>There is not as much documentation and tutorials available for QA as for other more popular NLP tasks. In particular, the examples provided are often focused on the SQUAD v1 format and not on SQUAD v2, the format selected for this project.
|
74 |
+
<li>Before the Hackathon, there was no Biomedical QA dataset in Spanish publicly available (particularly with the SQUAD V2 format). It was necessary to create a validation Biomedical Dataset using the SQUAD_ES Dataset.
|
75 |
+
</ul>
|
76 |
+
|
77 |
<h3>Conclusion and Future Work</h3>
|
78 |
If F1 Score is considered, the results show that there may be no advantage in using domain-specific masked language models to generate Biomedical QA models.
|
79 |
However, the F1 Scores reported for the Biomedical roberta-based models are not far below from those of the general roberta-based model.
|