English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT
  Robust Training for Conversational Question Answering Models with Reinforced Reformulation Generation

Kaiser, M., Saha Roy, R., & Weikum, G. (2023). Robust Training for Conversational Question Answering Models with Reinforced Reformulation Generation. Retrieved from https://arxiv.org/abs/2310.13505.

Item is

Files

hide Files
:
arXiv:2310.13505.pdf (Preprint), 3MB
Name:
arXiv:2310.13505.pdf
Description:
File downloaded from arXiv at 2023-11-13 12:44 WSDM 2024 Research Paper, 11 pages
OA-Status:
Not specified
Visibility:
Public
MIME-Type / Checksum:
application/pdf / [MD5]
Technical Metadata:
Copyright Date:
-
Copyright Info:
-

Locators

show

Creators

hide
 Creators:
Kaiser, Magdalena1, Author           
Saha Roy, Rishiraj1, Author           
Weikum, Gerhard1, Author           
Affiliations:
1Databases and Information Systems, MPI for Informatics, Max Planck Society, ou_24018              

Content

hide
Free keywords: Computer Science, Computation and Language, cs.CL,Computer Science, Artificial Intelligence, cs.AI,Computer Science, Information Retrieval, cs.IR
 Abstract: Models for conversational question answering (ConvQA) over knowledge graphs
(KGs) are usually trained and tested on benchmarks of gold QA pairs. This
implies that training is limited to surface forms seen in the respective
datasets, and evaluation is on a small set of held-out questions. Through our
proposed framework REIGN, we take several steps to remedy this restricted
learning setup. First, we systematically generate reformulations of training
questions to increase robustness of models to surface form variations. This is
a particularly challenging problem, given the incomplete nature of such
questions. Second, we guide ConvQA models towards higher performance by feeding
it only those reformulations that help improve their answering quality, using
deep reinforcement learning. Third, we demonstrate the viability of training
major model components on one benchmark and applying them zero-shot to another.
Finally, for a rigorous evaluation of robustness for trained models, we use and
release large numbers of diverse reformulations generated by prompting GPT for
benchmark test sets (resulting in 20x increase in sizes). Our findings show
that ConvQA models with robust training via reformulations, significantly
outperform those with standard training from gold QA pairs only.

Details

hide
Language(s): eng - English
 Dates: 2023-10-202023-11-062023
 Publication Status: Published online
 Pages: 11 p.
 Publishing info: -
 Table of Contents: -
 Rev. Type: -
 Identifiers: arXiv: 2310.13505
BibTex Citekey: Kaiser2310.13505
URI: https://arxiv.org/abs/2310.13505
 Degree: -

Event

show

Legal Case

show

Project information

show

Source

show