Share Email Print

Proceedings Paper

Using specific evaluation for comparing and combining competing algorithms: applying it to table column detection
Author(s): Ana Costa e Silva
Format Member Price Non-Member Price
PDF $17.00 $21.00

Paper Abstract

It is a commonly used evaluation strategy to run competing algorithms on a test dataset and state which performs better in average on the whole set. We call this generic evaluation. Although it is important, we believe this type of evaluation is incomplete. In this paper, we propose a methodology for algorithm comparison, which we call specific evaluation. This approach attempts to identify subsets of the data where one algorithm is better than the other. This allows not only knowing each algorithm's strengths and weaknesses better but also constitutes a simple way to develop a combination policy that allows enjoying the best of both. We shall be applying specific evaluation to an experiment that aims at grouping pre-obtained table cells into columns; we demonstrate how it identifies a subset of data for which the on-average least good but faster algorithm is equivalent or better, and how it then manages to create a policy for combining the two competing table column delimitation algorithms.

Paper Details

Date Published: 23 January 2012
PDF: 8 pages
Proc. SPIE 8297, Document Recognition and Retrieval XIX, 82970C (23 January 2012); doi: 10.1117/12.910544
Show Author Affiliations
Ana Costa e Silva, LIAAD-INESC Porto (Portugal)

Published in SPIE Proceedings Vol. 8297:
Document Recognition and Retrieval XIX
Christian Viard-Gaudin; Richard Zanibbi, Editor(s)

© SPIE. Terms of Use
Back to Top