Question Answering with Lydia (TREC 2005 QA Track)
نویسندگان
چکیده
The goal of our participation in TREC 2005 was to determine how effectively our entity recognition/text analysis system, Lydia (http://www.textmap.com) [1–3] could be adapted to question answering. Indeed, our entire QA subsystem consists of only about 2000 additional lines of Perl code. Lydia detects every named entity mentioned in the AQUAINT corpus, and keeps a variety of information on named entities and documents in a relational database. We can collect candidate answers by means of information kept in the database. To produce a response for the main task or a ranked list of documents for the document ranking task, we rank the collected candidate answers or documents using syntactic and statistical analyses. A significant distinction from other question answering systems [4–6] presented earlier at TREC is that we do not use web sources such as Wikipedia and Google to generate candidate answers or answers. Rather, we only use syntactic and statistical features of the test set of questions and corpus provided. Our approach is independent of other sources, and finds answers from the text provided. We describe the design of Lydia and associated algorithms in Section 2, and focus on the design and algorithms of the QA system in Section 3. We then analyze the performance of the QA system in Section 4, and conclude this paper with discussion on future directions in Section 5.
منابع مشابه
The Evaluation of Question Answering Systems: Lessons Learned from the TREC QA Track
The TREC question answering (QA) track was the first large-scale evaluation of open-domain question answering systems. In addition to successfully fostering research on the QA task, the track has also been used to investigate appropriate evaluation methodologies for question answering systems. This paper gives a brief history of the TREC QA track, motivating the decisions made in its implementa...
متن کاملTREC 2005 Question Answering Experiments at Tokyo Institute of Technology
In this paper we describe Tokyo Institute of Technology’s submission to the TAC2008 question answering (QA) track. Keeping the same theoretical QA model as for the TREC2007 track, developed for factoid QA, we investigated the effects of retrieving blog data versus web data for rigid list questions. For squishy list questions we relied on sentence retrieval, similar to how we approached “other” ...
متن کاملOverview of the TREC 2006 Question Answering Track 99
The TREC 2006 question answering (QA) track contained two tasks: the main task and the complex, interactive question answering (ciQA) task. As in 2005, the main task consisted of series of factoid, list, and “Other” questions organized around a set of targets; in contrast to previous years, the evaluation of factoid and list responses distinguished between answers that were globally correct (wi...
متن کاملTowards a Multi-Stream Question Answering-As-XML-Retrieval Strategy
We describe our participation in the TREC 2005 Question Answering track; our main focus this year was on improving our multi-stream approach to question answering and on making a first step towards a question answering-as-XML retrieval strategy. We provide a detailed account of the ideas underlying our approaches to the QA task, report on our results, and give a summary of our findings.
متن کاملDalTREC 2005 QA System Jellyfish: Mark-and-Match Approach to Question Answering
This is the second year that Dalhousie University actively participated in TREC. Three runs were submitted for the Question Answering track. Our results are below the median; however, they’re signifantly larger than minimal, the lesson learnt will guide our future development of the system. Our approach was based on a mark-and-match methodology with regular expression rewriting.
متن کامل