Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2013:2013:245-251.
doi: 10.1109/ACII.2013.47.

Facing Imbalanced Data Recommendations for the Use of Performance Metrics

Affiliations
Free PMC article

Facing Imbalanced Data Recommendations for the Use of Performance Metrics

László A Jeni et al. Int Conf Affect Comput Intell Interact Workshops. 2013.
Free PMC article

Abstract

Recognizing facial action units (AUs) is important for situation analysis and automated video annotation. Previous work has emphasized face tracking and registration and the choice of features classifiers. Relatively neglected is the effect of imbalanced data for action unit detection. While the machine learning community has become aware of the problem of skewed data for training classifiers, little attention has been paid to how skew may bias performance metrics. To address this question, we conducted experiments using both simulated classifiers and three major databases that differ in size, type of FACS coding, and degree of skew. We evaluated influence of skew on both threshold metrics (Accuracy, F-score, Cohen's kappa, and Krippendorf's alpha) and rank metrics (area under the receiver operating characteristic (ROC) curve and precision-recall curve). With exception of area under the ROC curve, all were attenuated by skewed distributions, in many cases, dramatically so. While ROC was unaffected by skew, precision-recall curves suggest that ROC may mask poor performance. Our findings suggest that skew is a critical factor in evaluating performance metrics. To avoid or minimize skew-biased estimates of performance, we recommend reporting skew-normalized scores along with the obtained ones.

PubMed Disclaimer

Figures

Figure 1
Figure 1
The behaviour of different metrics using simulated classifiers. The horizontal axis depicts the skew ratio (Skew=NegativeexamplesPositiveexamples), while the vertical axis shows the given metric score. The metrics are (a): Accuracy, (b): Cohen's kappa, (c) Area Under ROC, (d) F1 score, (e) Krippendorff's alpha, (f) Area Under PR Curve. The different lines show the relative misclassification rates of the simulated classifiers.

Similar articles

Cited by

References

    1. Abe S. Support vector machines for pattern classification. Springer; 2010. 3.
    1. Akbani R, Kwek S, Japkowicz N. Machine Learning: ECML 2004. Springer; Berlin Heidelberg: 2004. Applying support vector machines to imbalanced datasets. pp. 39–50. 1.
    1. Ashraf AB, Lucey S, Cohn JF, Chen T, Ambadar Z, Prkachin KM, Solomon PE. The painful face Pain expression recognition using active appearance models. Image and Vision Computing. 2009;27(12):1788–1796. 2. - PMC - PubMed
    1. Chawla NV, Japkowicz N, Kotcz A. Editorial: special issue on learning from imbalanced data sets. SIGKDD Explor. Newsl. 2004 Jun;6(1):1–6. (2004) 1.
    1. Chew SW, Lucey PJ, Lucey S, Saragih J, Sridharan J. F. Cohn S. Person-independent facial expression detection using constrained local models.. Proceedings of FG 2011 Facial Expression Recognition and Analysis Challenge; Santa Barbara, CA. 2011; 5.

LinkOut - more resources