A Critical Survey of the Methodology for IE Evaluation


Alberto Lavelli (1), Mary Elaine Califf (2), Fabio Ciravegna (3), Dayne Freitag (4), Claudio Giuliano (1), Nicholas Kushmerick (5), Lorenza Romano (1)

(1) ITC-irst, Trento, Italy; (2) Department of Applied Computer Science, Illinois State University, Illinois, USA; (3) Computer Science Department, University of Sheffield, UK; (4) Fair Isaac Corporation, California, USA; (5) Computer Science Department, University College Dublin, Ireland




We survey the evaluation methodology adopted in Information Extraction (IE), as defined in the MUC conferences and in later independent efforts applying machine learning to IE. We point out a number of problematic issues that may hamper the comparison between results obtained by different researchers. Some of them are common to other NLP tasks: e.g., the difficulty of exactly identifying the effects on performance of the data (sample selection and sample size), of the domain theory (features selected), and of algorithm parameter settings. Issues specific to IE evaluation include: how leniently to assess inexact identification of filler boundaries, the possibility of multiple fillers for a slot, and how the counting is performed. We argue that, when specifying an information extraction task, a number of characteristics should be clearly defined. However, in the papers only a few of them are usually explicitly specified. Our aim is to elaborate a clear and detailed experimental methodology and propose it to the IE community. The goal is to reach a widespread agreement on such proposal so that future IE evaluations will adopt the proposed methodology, making comparisons between algorithms fair and reliable. In order to achieve this goal, we will develop and make available to the community a set of tools and resources that incorporate a standardized IE methodology.


Information Extraction, evaluation methodology



Full Paper