Proceedings - Question Answering 2003¶
Overview of the TREC 2003 Question Answering Track¶
Ellen M. Voorhees
Abstract
The TREC 2003 question answering track contained two tasks, the passages task and the main task. In the passages task, systems returned a single text snippet in response to factoid questions; the evaluation metric was the number of snippets that contained a correct answer. The main task contained three separate types of questions, factoid questions, list questions, and definition questions. Each of the questions was tagged as to its type and the different question types were evaluated separately. The final score for a main task run was a combination of the scores for the separate question types. This paper defines the various tasks included in the track and reports the evaluation results. Since the TREC 2003 track was the first time for significant participation in the definition and list subtasks, the paper also examines the reliability of the evaluation for these tasks.
Bibtex
@inproceedings{DBLP:conf/trec/Voorhees03a,
author = {Ellen M. Voorhees},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {Overview of the {TREC} 2003 Question Answering Track},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {54--68},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/QA.OVERVIEW.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/Voorhees03a.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
AnswerFinder in TREC 2003¶
Diego Mollá Aliod
- Participant: macquarieu.molla
- Paper: http://trec.nist.gov/pubs/trec12/papers/macquarieu.qa.pdf
- Runs: answfind3 | answfind1 | answfind2
Abstract
In this our first participation in TREC we have focused on the passage task of the question answering track. The main aim of our participation was to test the impact of various types of linguistic information in a simple question answering system. In particular, we have tested various combinations of word overlap, grammatical relations overlap, and overlap of minimal logical forms in the final scoring module of the system. The results indicate a small increase of accuracy with respect to a baseline system based on word overlap. Overall, given the short time available for developing the system, the results are satisfactory and equal or surpass the median.
Bibtex
@inproceedings{DBLP:conf/trec/Aliod03,
author = {Diego Moll{\'{a}} Aliod},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {AnswerFinder in {TREC} 2003},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {392--398},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/macquarieu.qa.pdf},
timestamp = {Wed, 07 Jul 2021 16:44:22 +0200},
biburl = {https://dblp.org/rec/conf/trec/Aliod03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
A Hybrid Approach for QA Track Definitional Questions¶
Sasha Blair-Goldensohn, Kathleen R. McKeown, Andrew Hazen Schlaikjer
- Participant: ucolorado.ward
- Paper: http://trec.nist.gov/pubs/trec12/papers/columbiau.qa.pdf
- Runs: cuaqdef2003
Abstract
We present an overview of DefScriber, a system developed at Columbia University that combines knowledge-based and statistical methods to answer definitional questions of the form, “What is X?” We discuss how DefScriber was applied to the definition questions in the TREC 2003 QA track main task. We conclude with an analysis of our system's results on the definition questions.
Bibtex
@inproceedings{DBLP:conf/trec/Blair-GoldensohnMS03,
author = {Sasha Blair{-}Goldensohn and Kathleen R. McKeown and Andrew Hazen Schlaikjer},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {A Hybrid Approach for {QA} Track Definitional Questions},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {185--192},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/columbiau.qa.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/Blair-GoldensohnMS03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
MITRE's Qanda at TREC-12¶
John D. Burger
- Participant: mitre.burger
- Paper: http://trec.nist.gov/pubs/trec12/papers/mitre.qa.pdf
- Runs: MITRE2003A
Abstract
Qanda is MITRE's TREC-style question answering system. This year, we were able to apply only a small effort to the TREC QA activity, approximately one person-month. As well as some general improvements in Qanda's processing, we made some simple attempts to handle definition and list answers.
Bibtex
@inproceedings{DBLP:conf/trec/Burger03,
author = {John D. Burger},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {MITRE's Qanda at {TREC-12}},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {436--440},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/mitre.qa.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/Burger03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
TREC 2003 Question Answering Track at CAS-ICT¶
Yi Chang, Hongbo Xu, Shuo Bai
- Participant: cas-ict.bin
- Paper: http://trec.nist.gov/pubs/trec12/papers/chinese-acad-sci.qa.final.pdf
- Runs: ICTQA2003A | ICTQA2003B | ICTQA2003C
Abstract
In our system, we make use of Chunk information to analyze the question. A multilevel method is fulfilled to retrieve candidate Bi-sentences. As to answer selecting, we proposed a voting method. We figure out the performance of each module of our system, and our study shows that 65.54% information has lost in document retrieval and Bi-sentence retrieval.
Bibtex
@inproceedings{DBLP:conf/trec/ChangXB03,
author = {Yi Chang and Hongbo Xu and Shuo Bai},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {{TREC} 2003 Question Answering Track at {CAS-ICT}},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {147--151},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/chinese-acad-sci.qa.final.pdf},
timestamp = {Thu, 13 Aug 2020 01:00:00 +0200},
biburl = {https://dblp.org/rec/conf/trec/ChangXB03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
Bangor at TREC 2003: Q&A and Genomics Tracks¶
Terence Clifton, Alex Colquhoun, William John Teahan
- Participant: uwales.teahan
- Paper: http://trec.nist.gov/pubs/trec12/papers/ubangor-wales.qa.genomics.pdf
- Runs: uwbqitekat03
Abstract
This paper describes the participation of the School of Informatics, University of Wales, Bangor at TREC'2003 in the Q&A and Genomics Tracks. The paper is organized into three parts as follows. The first part provides a brief overview of the logic-based framework for Knowledgeable Agents that is currently being developed at Bangor. This was adopted as the basis for implementations used for both Tracks. The second part describes the Q&A system that was developed based on the framework, and the final part describes some experiments that were conducted within the Genomics Track at specifying context using GeneRIFs (for a Q&A system being developed for the BioMedical domain).
Bibtex
@inproceedings{DBLP:conf/trec/CliftonCT03,
author = {Terence Clifton and Alex Colquhoun and William John Teahan},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {Bangor at {TREC} 2003: Q{\&}A and Genomics Tracks},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {600--611},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/ubangor-wales.qa.genomics.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/CliftonCT03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
Multiple-Engine Question Answering in TextMap¶
Abdessamad Echihabi, Ulf Hermjakob, Eduard H. Hovy, Daniel Marcu, Eric Melz, Deepak Ravichandran
- Participant: usc-isi.hermjakob
- Paper: http://trec.nist.gov/pubs/trec12/papers/usc-isi.qa.pdf
- Runs: isi03a | isi03c | isi03b
Abstract
At TREC-2003, TextMap participated in the Main task, which encompassed answering the following types of questions: factoid questions; list questions; definition questions. In this paper, we overview the architecture of the TextMap system and report its performance, as evaluated by the NIST assessors, on each of these tracks.
Bibtex
@inproceedings{DBLP:conf/trec/EchihabiHHMMR03,
author = {Abdessamad Echihabi and Ulf Hermjakob and Eduard H. Hovy and Daniel Marcu and Eric Melz and Deepak Ravichandran},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {Multiple-Engine Question Answering in TextMap},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {772--781},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/usc-isi.qa.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/EchihabiHHMMR03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
Experiments in Novelty, Genes and Questions at the University of Iowa¶
David Eichmann, Padmini Srinivasan, Marc Light, Hudong Wang, Xin Ying Qiu, Robert J. Arens, Aditya Kumar Sehgal
- Participant: uiowa.eichmann
- Paper: http://trec.nist.gov/pubs/trec12/papers/uiowa.novelty.genomics.qa.pdf
- Runs: UIowaQA0301 | UIowaQA0302 | UIowaQA0303
Abstract
The University of Iowa participated in the novelty, genomics and question answering tracks of TREC-2003.
Bibtex
@inproceedings{DBLP:conf/trec/EichmannSLWQAS03,
author = {David Eichmann and Padmini Srinivasan and Marc Light and Hudong Wang and Xin Ying Qiu and Robert J. Arens and Aditya Kumar Sehgal},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {Experiments in Novelty, Genes and Questions at the University of Iowa},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {678--685},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/uiowa.novelty.genomics.qa.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/EichmannSLWQAS03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
The University of Sheffield's TREC 2003 Q&A Experiments¶
Robert J. Gaizauskas, Mark A. Greenwood, Mark Hepple, Ian Roberts, Horacio Saggion, Matthew Sargaison
- Participant: usheffield.gaizauskas
- Paper: http://trec.nist.gov/pubs/trec12/papers/usheffield.qa.pdf
- Runs: shef12simple | shef12okapi | shef12madcow
Abstract
The systems entered by the University of Sheffield in the question answering track of previous TRECs have been developments of the system first entered in TREC 8 (Humphreys et al., 1999). Although a range of improvements have been made to the system over the last four years (Scott and Gaizauskas, 2000; Greenwood et al., 2002), none has resulted in a significant performance increase. For this reason it was decided to approach the TREC 2003 evaluation more as a learning experience than as a forum in which to promote a particular approach to QA. We view this as the beginning of a process that will lead to much fuller appreciation of how to build more effective QA systems. [...]
Bibtex
@inproceedings{DBLP:conf/trec/GaizauskasGHRSS03,
author = {Robert J. Gaizauskas and Mark A. Greenwood and Mark Hepple and Ian Roberts and Horacio Saggion and Matthew Sargaison},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {The University of Sheffield's {TREC} 2003 Q{\&}A Experiments},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {782--790},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/usheffield.qa.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/GaizauskasGHRSS03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
TREC 2003 Robust, HARD and QA Track Experiments using PIRCS¶
Laszlo Grunfeld, Kui-Lam Kwok, Norbert Dinstl, Peter Deng
- Participant: queensc.kwok
- Paper: http://trec.nist.gov/pubs/trec12/papers/queens-college.robust.hard.qa.pdf
- Runs: pircsqa1 | pircsqa2 | pircsqa3
Abstract
We participated in the Robust, HARD and part of the QA tracks in TREC2003. For Robust track, a new way of doing ad-hoc retrieval based on web assistance was introduced. For HARD track, we followed the guideline to generate clarification forms for each topic so as to experiment with user feedback and metadata. In QA, we only did the factoid experiment. The approach to QA was similar to what we have used before, except that WWW searching was added as a front-end processing. These experiments are described in Sections 2, 3 and 4 respectively.
Bibtex
@inproceedings{DBLP:conf/trec/GrunfeldKDD03,
author = {Laszlo Grunfeld and Kui{-}Lam Kwok and Norbert Dinstl and Peter Deng},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {{TREC} 2003 Robust, {HARD} and {QA} Track Experiments using {PIRCS}},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {510--521},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/queens-college.robust.hard.qa.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/GrunfeldKDD03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
Answer Mining by Combining Extraction Techniques with Abductive Reasoning¶
Sanda M. Harabagiu, Dan I. Moldovan, Christine Clark, Mitchell Bowden, John Williams, Jeremy Bensley
- Participant: lcc.harabagiu
- Paper: http://trec.nist.gov/pubs/trec12/papers/lcc.qa.pdf
- Runs: LCCmainE03 | LCCpass03 | LCCmainS03
Abstract
Language Computer Corporation's Question Answering system combines the strengths of Information Extraction (IE) techniques with the vastness of axiomatic knowledge representations derived from Word-Net for justifying answers that are extracted from the AQUAINT text collection. CICERO LITE, the named entity recognizer employed in LCC's QA system was able to recognize precisely a large set of entities that ranged over an extended set of semantic categories. Similarly, the semantic hierarchy of answer types was also enhanced. To improve the precision of answer min-ing, the QA system also relied on a theorem prover that was able to produce abductive justifications of the answers when it had access to the axiomatic transformations of the WordNet glosses. This combination of techniques was successful and furthermore, produced little difference between the exact extractions and the paragraph extractions.
Bibtex
@inproceedings{DBLP:conf/trec/HarabagiuMCBWB03,
author = {Sanda M. Harabagiu and Dan I. Moldovan and Christine Clark and Mitchell Bowden and John Williams and Jeremy Bensley},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {Answer Mining by Combining Extraction Techniques with Abductive Reasoning},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {375--382},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/lcc.qa.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/HarabagiuMCBWB03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
UMass at TREC 2003: HARD and QA¶
Nasreen Abdul Jaleel, Andrés Corrada-Emmanuel, Qi Li, Xiaoyong Liu, Courtney Wade, James Allan
- Participant: umass.allan
- Paper: http://trec.nist.gov/pubs/trec12/papers/umass-amherst.hard.qa.pdf
- Runs: umassql
Abstract
The Center for Intelligent Information Retrieval (CIIR) at UMass Amherst participated in two tracks for TREC 2003: High Accuracy Retrieval from Documents (HARD) and Question Answering (QA). In the HARD track, we developed document metadata to correspond to query metadata requirements; implemented clarification forms based on query expansion, passage retrieval, and clustering; and retrieved variable length passages deemed most likely to be relevant. This work is discussed at length in Section 1. In the QA track, we focused on retrieving passages that were likely to contain the answer to the question.
Bibtex
@inproceedings{DBLP:conf/trec/JaleelCLLWA03,
author = {Nasreen Abdul Jaleel and Andr{\'{e}}s Corrada{-}Emmanuel and Qi Li and Xiaoyong Liu and Courtney Wade and James Allan},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {UMass at {TREC} 2003: {HARD} and {QA}},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {715--725},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/umass-amherst.hard.qa.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/JaleelCLLWA03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
The University of Amsterdam at the TREC 2003 Question Answering Track¶
Valentin Jijkoun, Gilad Mishne, Christof Monz, Maarten de Rijke, Stefan Schlobach, Oren Tsur
- Participant: uamsterdam.derijke
- Paper: http://trec.nist.gov/pubs/trec12/papers/uamsterdam.qa.pdf
- Runs: UAmsT03M2 | UAmsT03M1 | UAmsT03M3 | UAmsT03P1
Abstract
We describe our participation in the TREC 2003 Question Answering track. We explain the ideas underlying our approaches to the task, report on our results, provide an error analysis, and give a summary of our findings so far.
Bibtex
@inproceedings{DBLP:conf/trec/JijkounMMRST03,
author = {Valentin Jijkoun and Gilad Mishne and Christof Monz and Maarten de Rijke and Stefan Schlobach and Oren Tsur},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {The University of Amsterdam at the {TREC} 2003 Question Answering Track},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {586--593},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/uamsterdam.qa.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/JijkounMMRST03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
Integrating Web-based and Corpus-based Techniques for Question Answering¶
Boris Katz, Jimmy Lin, Daniel Loreto, Wesley Hildebrandt, Matthew W. Bilotti, Sue Felshin, Aaron Fernandes, Gregory Marton, Federico Mora
- Participant: mit.lin
- Paper: http://trec.nist.gov/pubs/trec12/papers/mit.qa.pdf
- Runs: MITCSAIL03a | MITCSAIL03b | MITCSAIL03c
Abstract
MIT CSAIL's entry in this year's TREC Question Answering track focused on integrating Web-based techniques with more traditional strategies based on document retrieval and named-entity detection. We believe that achieving high performance in the question answering task requires a combination of multiple strategies designed to capitalize on different characteristics of various resources. [...]
Bibtex
@inproceedings{DBLP:conf/trec/KatzLLHBFFMM03,
author = {Boris Katz and Jimmy Lin and Daniel Loreto and Wesley Hildebrandt and Matthew W. Bilotti and Sue Felshin and Aaron Fernandes and Gregory Marton and Federico Mora},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {Integrating Web-based and Corpus-based Techniques for Question Answering},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {426--435},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/mit.qa.pdf},
timestamp = {Fri, 27 Aug 2021 01:00:00 +0200},
biburl = {https://dblp.org/rec/conf/trec/KatzLLHBFFMM03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
ITC-irst at TREC 2003: the DIOGENE QA System¶
Milen Kouylekov, Bernardo Magnini, Matteo Negri, Hristo Tanev
- Participant: itcirst.magnini
- Paper: http://trec.nist.gov/pubs/trec12/papers/itc-irst.qa.pdf
- Runs: irstqa2003w | irstqa2003d | irstqa2003p
Abstract
This paper describes a new version of the DIOGENE Question Answering (QA) system developed at ITC-Irst. The recent updates here presented are targeted to the participation to TREC-2003 and meet the specific requirements of this year's QA main task. In particular, extending the backbone already developed for our participation to the last two editions of the QA track, special attention was paid to deal with the principal novelty factors of the new challenge, namely the introduction of the so-called definition and list questions. Moreover, we experimented with a first attempt to integrate parsing as a deeper linguistic analysis technique to find similarities between the syntactic structure of the input questions and the retrieved text passages. The outcome of such experiments, as well as the variations of the system's architecture and the results achieved at TREC-2003 will be presented in the following sections.
Bibtex
@inproceedings{DBLP:conf/trec/KouylekovMNT03,
author = {Milen Kouylekov and Bernardo Magnini and Matteo Negri and Hristo Tanev},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {ITC-irst at {TREC} 2003: the {DIOGENE} {QA} System},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {349--357},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/itc-irst.qa.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/KouylekovMNT03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
QED: The Edinburgh TREC-2003 Question Answering System¶
Jochen L. Leidner, Johan Bos, Tiphaine Dalmas, James R. Curran, Stephen Clark, Colin J. Bannard, Bonnie L. Webber, Mark Steedman
- Participant: uedinburgh.leidner
- Paper: http://trec.nist.gov/pubs/trec12/papers/uedinburgh.qa.ps
- Runs: EdinInf2003C | EdinInf2003B | EdinInf2003A
Abstract
This report describes a new open-domain answer retrieval system developed at the University of Edinburgh and gives results for the TREC-12 question answering track. Phrasal answers are identified by increasingly narrowing down the search space from a large text collection to a single phrase. The system uses document retrieval, query-based passage segmentation and ranking, semantic analysis from a wide-coverage parser, and a unification-like matching procedure to extract potential an-swers. A simple Web-based answer validation stage is also applied. The system is based on the Open Agent Architecture and has a parallel design so that multiple questions can be answered simultaneously on a Beowulf cluster.
Bibtex
@inproceedings{DBLP:conf/trec/LeidnerBDCCBWS03,
author = {Jochen L. Leidner and Johan Bos and Tiphaine Dalmas and James R. Curran and Stephen Clark and Colin J. Bannard and Bonnie L. Webber and Mark Steedman},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {{QED:} The Edinburgh {TREC-2003} Question Answering System},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {631--635},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/uedinburgh.qa.ps},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/LeidnerBDCCBWS03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
Use of Metadata for Question Answering and Novelty Tasks¶
Kenneth C. Litkowski
- Participant: clresearch
- Paper: http://trec.nist.gov/pubs/trec12/papers/clresearch.qa.novelty.pdf
- Runs: clr03m1 | clr03p2 | clr03p1
Abstract
CL Research's question-answering system for TREC 2003 was modified away from reliance on database technology to the core underlying technology of using massive XML-tagging for processing both questions and documents. This core technology was then extended to participate in the novelty task. This technology provides many opportuinities for experimenting with various approaches to question answering and novelty determination. For the QA track, we submitted one run and our overall main task score was 0.075, with scores of 0.070 for factoid questions, 0.000 for list questions, and 0.160 for definition questions. For the passage task, we submitted two runs, our better score was 0.119 for the factoid questions. These scores were all considerably below the medians for these tasks. We have implemented further routines since our official submission, improving our scores to 0.18 and 0.23 for the exact answer and passages tasks, respectively. For the Novelty track, we submitted four runs for task 1, one run for task 2, five runs for task 3, and one run for task 4; our submissions for tasks 2 and 4 were identical. For task 1, our best run received an F-score of 0.483 for relevant sentences and 0.410 for new sentences. For task 2, our F-score was 0.788 for new sentences. For task 3, our best F-score was 0.558 for relevant sentences and 0.419 for new sentences. For task 4, our F-score was 0.655 for new sentences. On average, our F-scores were somewhat above the medians on all tasks. We describe our system and examine our results from the perspective of exploiting the metadata in the XML tags.
Bibtex
@inproceedings{DBLP:conf/trec/Litkowski03,
author = {Kenneth C. Litkowski},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {Use of Metadata for Question Answering and Novelty Tasks},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {161--176},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/clresearch.qa.novelty.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/Litkowski03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
QA UdG-UPC System at TREC-12¶
Marc Massot, Horacio Rodríguez, Daniel Ferrés
- Participant: upc-udg.rodriguez
- Paper: http://trec.nist.gov/pubs/trec12/papers/upcudg.qa.pdf
- Runs: UPCUdGsys1
Abstract
This paper describes a prototype multilingual Q&A system that we have designed to participate in the Q&A Track of TREC-12. The system answer concrete responses, then we participate in the Q&A main task for factoid questions. The main areas of our system are: (1) Inductive Logic Programming to learn the question type, (2) Clustering of Named Entities to improve Information Retrieval and (3) Semantic relations and EuroWordNet synsets to perform a language-independent answer extraction.
Bibtex
@inproceedings{DBLP:conf/trec/MassotRF03,
author = {Marc Massot and Horacio Rodr{\'{\i}}guez and Daniel Ferr{\'{e}}s},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {{QA} UdG-UPC System at {TREC-12}},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {762--771},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/upcudg.qa.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/MassotRF03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
The JAVELIN Question-Answering System at TREC 2003: A Multi-Strategh Approach with Dynamic Planning¶
Eric Nyberg, Teruko Mitamura, James P. Callan, Jaime G. Carbonell, Robert E. Frederking, Kevyn Collins-Thompson, Laurie Hiyakumoto, Yifen Huang, Curtis Huttenhower, Scott Judy, Jeongwoo Ko, Anna Kupsc, Lucian Vlad Lita, Vasco Pedro, David Svoboda, Benjamin Van Durme
- Participant: cmu_javelin
- Paper: http://trec.nist.gov/pubs/trec12/papers/cmu.javelin.qa.pdf
- Runs: CMUJAV2003
Abstract
The JAVELIN system evaluated at TREC 2003 is an integrated architecture for open-domain question answering. JAVELIN employs a modular approach that addresses individual aspects of the QA task in an abstract manner. The System implements a planner that controls the execution and information flow, as well as a multiple answer seeking strategies used differently depending on the type of question.
Bibtex
@inproceedings{DBLP:conf/trec/NybergMCCFCHHHJKKLPSD03,
author = {Eric Nyberg and Teruko Mitamura and James P. Callan and Jaime G. Carbonell and Robert E. Frederking and Kevyn Collins{-}Thompson and Laurie Hiyakumoto and Yifen Huang and Curtis Huttenhower and Scott Judy and Jeongwoo Ko and Anna Kupsc and Lucian Vlad Lita and Vasco Pedro and David Svoboda and Benjamin Van Durme},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {The {JAVELIN} Question-Answering System at {TREC} 2003: {A} Multi-Strategh Approach with Dynamic Planning},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/cmu.javelin.qa.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/NybergMCCFCHHHJKKLPSD03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
Passage Scoring for Question Answering via Bayesian Inference on Lexical Relations¶
Deepa Paranjpe, Ganesh Ramakrishnan, Sumana Srinivasan
- Participant: iitb.ramakrishnan
- Paper: http://trec.nist.gov/pubs/trec12/papers/iit.qa.pdf
- Runs: IITBQA | IITBQA1
Abstract
Many researchers have used lexical networks and ontologies to mitigate synonymy and polysemy problems in Question Answering (QA), systems coupled with taggers, query classifiers, and answer extractors in complex and ad-hoc ways. We seek to make QA systems reproducible with shared and modest human effort, carefully separating knowledge from algorithms. To this end, we propose an aesthetically “clean” Bayesian inference scheme for exploiting lexical relations for passage-scoring for QA . The factors which contribute to the efficacy of Bayesian Inferencing on lexical relations are soft word sense disambiguation, parameter smoothing which ameliorates the data sparsity problem and estimation of joint probability over words which overcomes the deficiency of naive-bayes-like approaches.
Bibtex
@inproceedings{DBLP:conf/trec/ParanjpeRS03,
author = {Deepa Paranjpe and Ganesh Ramakrishnan and Sumana Srinivasan},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {Passage Scoring for Question Answering via Bayesian Inference on Lexical Relations},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {305--210},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/iit.qa.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/ParanjpeRS03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
IBM's PIQUANT in TREC2003¶
John M. Prager, Jennifer Chu-Carroll, Krzysztof Czuba, Christopher A. Welty, Abraham Ittycheriah, Ruchi Mahindru
- Participant: ibm.prager
- Paper: http://trec.nist.gov/pubs/trec12/papers/ibm-prager.qa.pdf
- Runs: IBM2003a | IBM2003c | IBM2003b
Abstract
For the most part, the system we used for TREC2003 was a smooth evolution of the one we ran in TREC2002 [Chu-Carroll et al, 2003b]. We continued to use our multi-source and multi-agent architecture. For Factoid questions we used all of our previous answering agents with an additional pattern-based agent, an enhanced answer resolution algorithm, and increased coverage of the Cyc sanity checker. We will devote a portion of this paper to performing a post-mortem of our experiences with Cyc this year. For List questions, which we did not attempt previously, we ran our Factoid system with different parameters. For Definition questions we took an entirely new approach, which we call QA-by-Dossier, and which will be the other focus of this paper. While we think that our system performed reasonably well in this subtask, the NIST evaluation results do not reflect this, raising some questions about the Definition subtask specification and evaluation.
Bibtex
@inproceedings{DBLP:conf/trec/PragerCCWIM03,
author = {John M. Prager and Jennifer Chu{-}Carroll and Krzysztof Czuba and Christopher A. Welty and Abraham Ittycheriah and Ruchi Mahindru},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {IBM's {PIQUANT} in {TREC2003}},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {283--292},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/ibm-prager.qa.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/PragerCCWIM03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
Question Answering using the DLT System at TREC 2003¶
Richard F. E. Sutcliffe, Igal Gabbay, Michael Mulcahy, Kieran White
- Participant: ulimerick.sutcliffe
- Paper: http://trec.nist.gov/pubs/trec12/papers/ulimerick.qa.pdf
- Runs: DLT03QA01 | DLT03QA02
Abstract
This article outlines our participation in the Question Answering Track of the Text REtrieval Conference organised by the National Institute of Standards and Technology. This was our second year in the track and we hoped to improve our performance relative to 2002. In the next section we outline the general strategy we adopted, the changes relative to last year and the approaches taken to the three question types, namely factoid, list and definition. Following this the individual system components are described in more detail. Thirdly, the runs we submitted are presented together with the results obtained. Finally, conclusions are drawn based on our findings.
Bibtex
@inproceedings{DBLP:conf/trec/SutcliffeGMW03,
author = {Richard F. E. Sutcliffe and Igal Gabbay and Michael Mulcahy and Kieran White},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {Question Answering using the {DLT} System at {TREC} 2003},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {686--698},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/ulimerick.qa.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/SutcliffeGMW03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
FDUQA on TREC2003 QA task¶
Lide Wu, Xuanjing Huang, Yaqian Zhou, Yongping Du, Lan You
- Participant: fudanu.lide
- Paper: http://trec.nist.gov/pubs/trec12/papers/fudanu.qa.pdf
- Runs: FDUT12QA2 | FDUT12QA3 | FDUT12QA1
Abstract
It is the fourth time that we take part in the QA track. Our system, FDUQA, is based on our previous system (Wu et al, 2002). FDUQA includes an offline part and an online part. We make great efforts on the online part while leaving the offline part unchanged. We have tried many natural language processing techniques, and incorporated many sources of world knowledge, including Web. A novel Query formulation technique has also been put forward. In addition, we've tried another attempt on answer extraction in this year's task. In the second section, we will describe the architecture of our QA system; and give a detailed description on the Query formulation for Web search in the third section; while in the fourth section, we will introduce our new attempt on answer extraction; and we will present our performance in the last section.
Bibtex
@inproceedings{DBLP:conf/trec/WuHZDY03,
author = {Lide Wu and Xuanjing Huang and Yaqian Zhou and Yongping Du and Lan You},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {{FDUQA} on {TREC2003} {QA} task},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {246--253},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/fudanu.qa.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/WuHZDY03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
Questioning Answering By Pattern Matching, Web-Proofing, Semantic Form Proofing¶
Min Wu, Xiaoyu Zheng, Michelle Duan, Ting Liu, Tomek Strzalkowski
- Participant: suny-albany.liu
- Paper: http://trec.nist.gov/pubs/trec12/papers/ualbany-suny.qa.pdf
- Runs: Albany03I2 | Albany03I3 | Albany03I4
Abstract
In this paper, we introduce the University at Albany's question answering system, ILQUA. It is developed on the following methods: pattern matching over annotated text, web-proofing and semantic form proofing. These methods are currently used in other QA systems, however, we revised them to work together in our QA system.
Bibtex
@inproceedings{DBLP:conf/trec/WuZDLS03,
author = {Min Wu and Xiaoyu Zheng and Michelle Duan and Ting Liu and Tomek Strzalkowski},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {Questioning Answering By Pattern Matching, Web-Proofing, Semantic Form Proofing},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {578--585},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/ualbany-suny.qa.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/WuZDLS03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
TREC 2003 QA at BBN: Answering Definitional Questions¶
Jinxi Xu, Ana Licuanan, Ralph M. Weischedel
- Participant: bbn.xu
- Paper: http://trec.nist.gov/pubs/trec12/papers/bbn.qa.pdf
- Runs: BBN2003C | BBN2003A | BBN2003B
Abstract
In TREC 2003, we focused on definitional questions. For factoid and list questions, we simply re-used our TREC 2002 system with some modifications. For definitional QA, we adopted a hybrid approach that combines several complementary technology components. Information retrieval (IR) was used to retrieve from the corpus the relevant documents for each question. Various linguistic and extraction tools were used to analyze the retrieved texts and to extract various types of kernel facts from which the answer to the question is generated. These tools include name finding, parsing, co-reference resolution, proposition extraction, relation extraction and extraction of structured patterns. All text analysis functions except structured pattern extraction were carried out by Serif, a state of the art information extraction engine (Ramshaw, et al, 2001) from BBN. Section 2 summarizes our submission for factoid and list qeustion answering (QA). The rest of the paper focuses on defintional questions. Section 4 concludes this work.
Bibtex
@inproceedings{DBLP:conf/trec/XuLW03,
author = {Jinxi Xu and Ana Licuanan and Ralph M. Weischedel},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {{TREC} 2003 {QA} at {BBN:} Answering Definitional Questions},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {98--106},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/bbn.qa.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/XuLW03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
QUALIFIER In TREC-12 QA Main Task¶
Hui Yang, Hang Cui, Mstislav Maslennikov, Long Qiu, Min-Yen Kan, Tat-Seng Chua
- Participant: nus.yang
- Paper: http://trec.nist.gov/pubs/trec12/papers/nus-yang.qa.pdf
- Runs: nusmml03r1 | nusmml03r2 | nusmml03r3
Abstract
This paper describes a question answering system and its various modules to solve definition, factoid and list questions defined in the TREC12 Main task. In particular, we tackle the factoid QA task by Event-based Question Answering. Each QA event comprises of elements describing different facets like time, location, object, action etc. By analyzing the external knowledge from pre-retrieved TREC documents, Web documents, WordNet and Ontology to discover the QA event structure, we explore the inherent associations among QA elements and then obtain the answers. There are three subsystems working parallel to handle definition, factoid, and list questions separately. We highlight the shared modules, fine-grained named entity recognition, anaphora resolution and canonicalization co-reference resolution, among the three subsystems as well.
Bibtex
@inproceedings{DBLP:conf/trec/YangCMQKC03,
author = {Hui Yang and Hang Cui and Mstislav Maslennikov and Long Qiu and Min{-}Yen Kan and Tat{-}Seng Chua},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {{QUALIFIER} In {TREC-12} {QA} Main Task},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {480--488},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/nus-yang.qa.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/YangCMQKC03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
A Language Modeling Approach to Passage Question Answering¶
Dell Zhang, Wee Sun Lee
- Participant: nus.sun
- Paper: http://trec.nist.gov/pubs/trec12/papers/nus-zhang.qa.pdf
- Runs: nuslamp03a | nuslamp03b | nuslamp03
Abstract
This paper reports our efforts on developing a language modeling approach to passage question answering. In particular, we address the following two problems: (i) generalized language modeling for question classification; (ii) constrained language modeling for passage retrieval.
Bibtex
@inproceedings{DBLP:conf/trec/ZhangL03,
author = {Dell Zhang and Wee Sun Lee},
editor = {Ellen M. Voorhees and Lori P. Buckland},
title = {A Language Modeling Approach to Passage Question Answering},
booktitle = {Proceedings of The Twelfth Text REtrieval Conference, {TREC} 2003, Gaithersburg, Maryland, USA, November 18-21, 2003},
series = {{NIST} Special Publication},
volume = {500-255},
pages = {489--495},
publisher = {National Institute of Standards and Technology {(NIST)}},
year = {2003},
url = {http://trec.nist.gov/pubs/trec12/papers/nus-zhang.qa.pdf},
timestamp = {Thu, 12 Mar 2020 00:00:00 +0100},
biburl = {https://dblp.org/rec/conf/trec/ZhangL03.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}