NCU-IISR: Using a Pre-trained Language Model and Logistic Regression Model for BioASQ Task 8b Phase B

Jen Chieh Han, Richard Tzong Han Tsai

研究成果: 雜誌貢獻會議論文同行評審

摘要

Recent successes in pre-trained language models, such as BERT, RoBERTa, and XLNet, have yielded state-of-the-art results in the natural language processing field. BioASQ is a question answering (QA) benchmark with a public and competitive leaderboard that spurs advancement in large-scale pre-trained language models for biomedical QA. In this paper, we introduce our system for the BioASQ Task 8b Phase B. We employed a pre-trained biomedical language model, BioBERT, to generate “exact” answers for the questions, and a logistic regression model with our sentence embedding to construct the top-n sentences/snippets as a prediction for “ideal” answers. On the final test batch, our best configuration achieved the highest ROUGE-2 and ROUGE-SU4 F1 scores among all participants in the 8th BioASQ QA task (Task 8b, Phase B).

原文???core.languages.en_GB???
期刊CEUR Workshop Proceedings
2696
出版狀態已出版 - 2020
事件11th Conference and Labs of the Evaluation Forum, CLEF 2020 - Thessaloniki, Greece
持續時間: 22 9月 202025 9月 2020

指紋

深入研究「NCU-IISR: Using a Pre-trained Language Model and Logistic Regression Model for BioASQ Task 8b Phase B」主題。共同形成了獨特的指紋。

引用此