Timezone: »
Given an unsupervised novelty detection task on a new dataset, how can we automatically select a ''best'' detection model while simultaneously controlling the error rate of the best model? For novelty detection analysis, numerous detectors have been proposed to detect outliers on a new unseen dataset based on a score function trained on available clean data. However, due to the absence of labeled data for model evaluation and comparison, there is a lack of systematic approaches that are able to select a ''best'' model/detector (i.e., the algorithm as well as its hyperparameters) and achieve certain error rate control simultaneously. In this paper, we introduce a unified data-driven procedure to address this issue. The key idea is to maximize the number of detected outliers while controlling the false discovery rate (FDR) with the help of Jackknife prediction. We establish non-asymptotic bounds for the false discovery proportions and show that the proposed procedure yields valid FDR control under some mild conditions. Numerical experiments on both synthetic and real data validate the theoretical results and demonstrate the effectiveness of our proposed AutoMS method. The code is available at https://github.com/ZhangYifan1996/AutoMS.
Author Information
Yifan Zhang (Nankai University)

Zhang Yifan is a Ph.D. student in Statistics, from School of Statistics and Data Science, Nankai University, China. Her Ph.D. supervisor is [Prof. Changliang Zou](http://web.stat.nankai.edu.cn/chlzou/). Her research interests include change-point analysis, outlier detection, and statistical process control. She was awarded university scholarships in 2014, 2015, 2017, 2019, 2021. If you are interested in her research topics, please feel free to contact her via email: yfzhang_stat@mail.nankai.edu.cn.
Haiyan Jiang (MBZUAI)
Haojie Ren (Shanghai Jiaotong University)
Changliang Zou
Dejing Dou (Baidu)
More from the Same Authors
-
2022 Poster: Generative Time Series Forecasting with Diffusion, Denoise, and Disentanglement »
Yan Li · Xinjiang Lu · Yaqing Wang · Dejing Dou -
2022 Poster: InterpretDL: Explaining Deep Models in PaddlePaddle »
Xuhong Li · Haoyi Xiong · Xingjian Li · Xuanyu Wu · Zeyu Chen · Dejing Dou -
2022 : SMILE: Sample-to-feature MIxup for Efficient Transfer LEarning »
Xingjian Li · Haoyi Xiong · Cheng-Zhong Xu · Dejing Dou -
2022 : A Simple Framework for Active Learning to Rank »
Qingzhong Wang · Haifang Li · Haoyi Xiong · Wen Wang · Jiang Bian · Yu Lu · Shuaiqiang Wang · zhicong cheng · Dawei Yin · Dejing Dou -
2022 : A Comparative Survey of Deep Active Learning »
Xueying Zhan · Qingzhong Wang · Kuan-Hao Huang · Haoyi Xiong · Dejing Dou · Antoni Chan -
2022 Spotlight: InterpretDL: Explaining Deep Models in PaddlePaddle »
Xuhong Li · Haoyi Xiong · Xingjian Li · Xuanyu Wu · Zeyu Chen · Dejing Dou -
2022 Spotlight: Lightning Talks 1A-1 »
Siba Smarak Panigrahi · Xuhong Li · Mikhail Usvyatsov · Shaohan Chen · Sohan Patnaik · Haoyi Xiong · Nikolaos V Sahinidis · Rafael Ballester-Ripoll · Chuanhou Gao · Xingjian Li · Konrad Schindler · Xuanyu Wu · Zeyu Chen · Dejing Dou -
2021 : [O6] Explaining Information Flow Inside Vision Transformers Using Markov Chain »
Tingyi Yuan · Xuhong Li · Haoyi Xiong · Dejing Dou -
2020 Poster: Discriminative Sounding Objects Localization via Self-supervised Audiovisual Matching »
Di Hu · Rui Qian · Minyue Jiang · Xiao Tan · Shilei Wen · Errui Ding · Weiyao Lin · Dejing Dou