The Importance of the Test Set Size in Quantification Assessment

The Importance of the Test Set Size in Quantification Assessment

André Maletzke, Waqar Hassan, Denis dos Reis, Gustavo Batista

Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence
Main track. Pages 2640-2646. https://doi.org/10.24963/ijcai.2020/366

Quantification is a task similar to classification in the sense that it learns from a labeled training set. However, quantification is not interested in predicting the class of each observation, but rather measure the class distribution in the test set. The community has developed performance measures and experimental setups tailored to quantification tasks. Nonetheless, we argue that a critical variable, the size of the test sets, remains ignored. Such disregard has three main detrimental effects. First, it implicitly assumes that quantifiers will perform equally well for different test set sizes. Second, it increases the risk of cherry-picking by selecting a test set size for which a particular proposal performs best. Finally, it disregards the importance of designing methods that are suitable for different test set sizes. We discuss these issues with the support of one of the broadest experimental evaluations ever performed, with three main outcomes. (i) We empirically demonstrate the importance of the test set size to assess quantifiers. (ii) We show that current quantifiers generally have a mediocre performance on the smallest test sets. (iii) We propose a metalearning scheme to select the best quantifier based on the test size that can outperform the best single quantification method.
Keywords:
Machine Learning: Other
Data Mining: Other