Integrate Candidate Answer Extraction with Re-Ranking for Chinese Machine Reading Comprehension
Open Access
- 8 March 2021
- Vol. 23 (3), 322
- https://doi.org/10.3390/e23030322
Abstract
Machine Reading Comprehension (MRC) research concerns how to endow machines with the ability to understand given passages and answer questions, which is a challenging problem in the field of natural language processing. To solve the Chinese MRC task efficiently, this paper proposes an Improved Extraction-based Reading Comprehension method with Answer Re-ranking (IERC-AR), consisting of a candidate answer extraction module and a re-ranking module. The candidate answer extraction module uses an improved pre-training language model, RoBERTa-WWM, to generate precise word representations, which can solve the problem of polysemy and is good for capturing Chinese word-level features. The re-ranking module re-evaluates candidate answers based on a self-attention mechanism, which can improve the accuracy of predicting answers. Traditional machine-reading methods generally integrate different modules into a pipeline system, which leads to re-encoding problems and inconsistent data distribution between the training and testing phases; therefore, this paper proposes an end-to-end model architecture for IERC-AR to reasonably integrate the candidate answer extraction and re-ranking modules. The experimental results on the Les MMRC dataset show that IERC-AR outperforms state-of-the-art MRC approaches.Keywords
Funding Information
- National Science Foundation of Hunan Province (2017JJ3371)
This publication has 5 references indexed in Scilit:
- Improving Neural Response Diversity with Frequency-Aware Cross-Entropy LossPublished by Association for Computing Machinery (ACM) ,2019
- Gated Self-Matching Networks for Reading Comprehension and Question AnsweringPublished by Association for Computational Linguistics (ACL) ,2017
- Get To The Point: Summarization with Pointer-Generator NetworksPublished by Association for Computational Linguistics (ACL) ,2017
- Higher-order Lexical Semantic Models for Non-factoid Answer RerankingTransactions of the Association for Computational Linguistics, 2015
- Long Short-Term MemoryNeural Computation, 1997