Resource title

Deriving Consensus Rankings from Benchmarking Experiments

Resource image

image for OpenScout resource :: Deriving Consensus Rankings from Benchmarking Experiments

Resource description

Whereas benchmarking experiments are very frequently used to investigate the performance of statistical or machine learning algorithms for supervised and unsupervised learning tasks, overall analyses of such experiments are typically only carried out on a heuristic basis, if at all. We suggest to determine winners, and more generally, to derive a consensus ranking of the algorithms, as the linear order on the algorithms which minimizes average symmetric distance (Kemeny-Snell distance) to the performance relations on the individual benchmark data sets. This leads to binary programming problems which can typically be solved reasonably efficiently. We apply the approach to a medium-scale benchmarking experiment to assess the performance of Support Vector Machines in regression and classification problems, and compare the obtained consensus ranking with rankings obtained by simple scoring and Bradley-Terry modeling. (author's abstract) ; Series: Research Report Series / Department of Statistics and Mathematics

Resource author

Kurt Hornik, David Meyer

Resource publisher

Resource publish date

Resource language


Resource content type


Resource resource URL

Resource license

Adapt according to the license agreement. Always reference the original source and author.