메뉴 건너뛰기
Library Notice
Institutional Access
If you certify, you can access the articles for free.
Check out your institutions.
ex)Hankuk University, Nuri Motors
Log in Register Help KOR
Subject

Evaluating of Korean Machine Reading Comprehension Generalization Performance via Cross-, Blind and Open-Domain QA Dataset Assessment
Recommendations
Search
Questions

기계독해 말뭉치의 교차 평가, 블라인드 평가 및 오픈도메인 질의응답 환경 평가를 통한 한국어 기계독해의 일반화 성능 평가

논문 기본 정보

Type
Academic journal
Author
Joon-Ho Lim (한국전자통신연구원) Hyun-ki Kim (한국전자통신연구원)
Journal
Korean Institute of Information Scientists and Engineers Journal of KIISE Vol.48 No.3 KCI Excellent Accredited Journal
Published
2021.3
Pages
275 - 283 (9page)
DOI
10.5626/JOK.2021.48.3.275

Usage

cover
📌
Topic
📖
Background
🔬
Method
🏆
Result
Evaluating of Korean Machine Reading Comprehension Generalization Performance via Cross-, Blind and Open-Domain QA Dataset Assessment
Ask AI
Recommendations
Search
Questions

Abstract· Keywords

Report Errors
Machine reading comprehension (MRC) entails identification of the correct answer in a paragraph when a natural language question and paragraph are provided. Recently, fine-tuning based on a pre-trained language model yields the best performance. In this study, we evaluated the ability of machine-reading comprehension method to generalize question and paragraph pairs, rather than similar training sets. Towards this end, the cross-evaluation between datasets and blind evaluation was performed. The results showed a correlation between generalization performance and datasets such as answer length and overlap ratio between question and paragraph. As a result of blind evaluation, the evaluation dataset with the long answer and low lexical overlap between the questions and paragraphs resulted in less than 80% performance. Finally, the generalized performance of the MRC model under the open domain QA environment was evaluated, and the performance of the MRC using the searched paragraph was found to be degraded. According to the MRC task characteristics, the difficulty and differences in generalization performance depend on the relationship between the question and the answer, suggesting the need for analysis of different evaluation sets.

Contents

요약
Abstract
1. 서론
2. 관련 연구
3. 기계독해 말뭉치의 교차평가를 통한 일반화 성능 평가
4. 블라인드 평가를 통한 일반화 성능 평가
5. 오픈-도메인 질의응답 환경에서의 기계독해 일반화 성능 평가
6. 결론
References

References (15)

Add References

Recommendations

It is an article recommended by DBpia according to the article similarity. Check out the related articles!

Related Authors

Frequently Viewed Together

Recently viewed articles

Comments(0)

0

Write first comments.