Recent successes in pre-trained language models, such as BERT, RoBERTa, and XLNet, have yielded state-of-the-art results in the natural language processing field. BioASQ is a question answering (QA) benchmark with a public and competitive leaderboard that spurs advancement in large-scale pre-trained language models for biomedical QA. In this paper, we introduce our system for the BioASQ Task 8b Phase B. We employed a pre-trained biomedical language model, BioBERT, to generate “exact” answers for the questions, and a logistic regression model with our sentence embedding to construct the top-n sentences/snippets as a prediction for “ideal” answers. On the final test batch, our best configuration achieved the highest ROUGE-2 and ROUGE-SU4 F1 scores among all participants in the 8th BioASQ QA task (Task 8b, Phase B).
|Journal||CEUR Workshop Proceedings|
|State||Published - 2020|
|Event||11th Conference and Labs of the Evaluation Forum, CLEF 2020 - Thessaloniki, Greece|
Duration: 22 Sep 2020 → 25 Sep 2020
- Biomedical Question Answering
- Logistic Regression
- Pre-trained Language Model