ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2305.04737
8
3

SkillQG: Learning to Generate Question for Reading Comprehension Assessment

8 May 2023
Xiaoqiang Wang
Bang Liu
Siliang Tang
Lingfei Wu
ArXivPDFHTML
Abstract

We present \textbf{\texttt{SkillQG}}: a question generation framework with controllable comprehension types for assessing and improving machine reading comprehension models. Existing question generation systems widely differentiate questions by literal\textit{literal}literal information such as question words and answer types to generate semantically relevant questions for a given context. However, they rarely consider the comprehension\textit{comprehension}comprehension nature of questions, i.e. the different comprehension capabilities embodied by different questions. In comparison, our SkillQG\texttt{SkillQG}SkillQG is able to tailor a fine-grained assessment and improvement to the capabilities of question answering models built on it. Specifically, we first frame the comprehension type of questions based on a hierarchical skill-based schema, then formulate SkillQG\texttt{SkillQG}SkillQG as a skill-conditioned question generator. Furthermore, to improve the controllability of generation, we augment the input text with question focus and skill-specific knowledge, which are constructed by iteratively prompting the pre-trained language models. Empirical results demonstrate that SkillQG\texttt{SkillQG}SkillQG outperforms baselines in terms of quality, relevance, and skill-controllability while showing a promising performance boost in downstream question answering task.

View on arXiv
Comments on this paper