Printable PDF
Department of Mathematics,
University of California San Diego

****************************

Statistics Seminar

Xin Tong

Marshall School of Business, University of Southern California

Neyman-Pearson classification

Abstract:

In many binary classification applications, such as disease diagnosis and spam detection, practitioners commonly face the need to limit type I error (that is, the conditional probability of misclassifying a class 0 observation as class 1) so that it remains below a desired threshold. To address this need, the Neyman-Pearson (NP) classification paradigm is a natural choice; it minimizes type II error (that is, the conditional probability of misclassifying a class 1 observation as class 0) while enforcing an upper bound, alpha, on the type I error. Although the NP paradigm has a century-long history in hypothesis testing, it has not been well recognized and implemented in classification schemes. Common practices that directly limit the empirical type I error to no more than alpha do not satisfy the type I error control objective because the resulting classifiers are still likely to have type I errors much larger than alpha. This talk introduces the speaker's work on NP classification algorithms and their applications and raises current challenges under the NP paradigm.

Host: Jelena Bradic

October 5, 2018

10:00 AM

AP&M 7321

****************************