Comparison of classifiers in challenge scheme

05/16/2023
by   Sergio Nava-Muñoz, et al.
8

In recent decades, challenges have become very popular in scientific research as these are crowdsourcing schemes. In particular, challenges are essential for developing machine learning algorithms. For the challenges settings, it is vital to establish the scientific question, the dataset (with adequate quality, quantity, diversity, and complexity), performance metrics, as well as a way to authenticate the participants' results (Gold Standard). This paper addresses the problem of evaluating the performance of different competitors (algorithms) under the restrictions imposed by the challenge scheme, such as the comparison of multiple competitors with a unique dataset (with fixed size), a minimal number of submissions and, a set of metrics chosen to assess performance. The algorithms are sorted according to the performance metric. Still, it is common to observe performance differences among competitors as small as hundredths or even thousandths, so the question is whether the differences are significant. This paper analyzes the results of the MeOffendEs@IberLEF 2021 competition and proposes to make inference through resampling techniques (bootstrap) to support Challenge organizers' decision-making.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/07/2018

Not quite unreasonable effectiveness of machine learning algorithms

State-of-the-art machine learning algorithms demonstrate close to absolu...
research
12/13/2021

Beyond Ads: Sequential Decision-Making Algorithms in Public Policy

We explore the promises and challenges of employing sequential decision-...
research
01/16/2023

LYSTO: The Lymphocyte Assessment Hackathon and Benchmark Dataset

We introduce LYSTO, the Lymphocyte Assessment Hackathon, which was held ...
research
05/12/2014

The Metrics Matter! On the Incompatibility of Different Flavors of Replanning

When autonomous agents are executing in the real world, the state of the...
research
02/09/2021

k-Anonymity in Practice: How Generalisation and Suppression Affect Machine Learning Classifiers

The protection of private information is a crucial issue in data-driven ...
research
10/09/2020

Evaluating and Characterizing Human Rationales

Two main approaches for evaluating the quality of machine-generated rati...

Please sign up or login with your details

Forgot password? Click here to reset