Factoid question answering for spoken documents

  1. Comas Umbert, Pere Ramon
unter der Leitung von:
  1. Jordi Turmo Doktorvater/Doktormutter
  2. Lluís Márquez Villodre Doktorvater/Doktormutter

Universität der Verteidigung: Universitat Politècnica de Catalunya (UPC)

Fecha de defensa: 12 von Juni von 2012

Gericht:
  1. Horacio Rodríguez Hontoria Präsident/in
  2. Lluís Padró Cirera Sekretär/in
  3. Maarten de Rijke Vocal
  4. José-Luis Vicedo González Vocal
  5. Sophie Rosset Vocal

Art: Dissertation

Teseo: 114066 DIALNET

Zusammenfassung

In this dissertation, we present a factoid question answering system, specifically tailored for Question Answering (QA) on spoken documents. This work explores, for the first time, which techniques can be robustly adapted from the usual QA on written documents to the more difficult spoken documents scenario. More specifically, we study new information retrieval (IR) techniques designed for speech, and utilize several levels of linguistic information for the speech-based QA task. These include named-entity detection with phonetic information, syntactic parsing applied to speech transcripts, and the use of coreference resolution. Our approach is largely based on supervised machine learning techniques, with special focus on the answer extraction step, and makes little use of handcrafted knowledge. Consequently, it should be easily adaptable to other domains and languages. In the work resulting of this Thesis, we have impulsed and coordinated the creation of an evaluation framework for the task of QA on spoken documents. The framework, named QAst, provides multi-lingual corpora, evaluation questions, and answers key. These corpora have been used in the QAst evaluation that was held in the CLEF workshop for the years 2007, 2008 and 2009, thus helping the developing of state-of-the-art techniques for this particular topic. The presentend QA system and all its modules are extensively evaluated on the European Parliament Plenary Sessions English corpus composed of manual transcripts and automatic transcripts obtained by three different Automatic Speech Recognition (ASR) systems that exhibit significantly different word error rates. This data belongs to the CLEF 2009 track for QA on speech transcripts. The main results confirm that syntactic information is very useful for learning to rank question candidates, improving results on both manual and automatic transcripts unless the ASR quality is very low. Overall, the performance of our system is comparable or better than the state-of-the-art on this corpus, confirming the validity of our approach.