Timezone: »
Our goal is to evaluate the accuracy of a black-box classification model, not as a single aggregate on a given test data distribution, but as a surface over a large number of combinations of attributes characterizing multiple test data distributions. Such attributed accuracy measures become important as machine learning models get deployed as a service, where the training data distribution is hidden from clients, and different clients may be interested in diverse regions of the data distribution. We present Attributed Accuracy Assay (AAA) --- a Gaussian Process (GP)-based probabilistic estimator for such an accuracy surface. Each attribute combination, called an 'arm' is associated with a Beta density from which the service's accuracy is sampled. We expect the GP to smooth the parameters of the Beta density over related arms to mitigate sparsity. We show that obvious application of GPs cannot address the challenge of heteroscedastic uncertainty over a huge attribute space that is sparsely and unevenly populated. In response, we present two enhancements: pooling sparse observations, and regularizing the scale parameter of the Beta densities. After introducing these innovations, we establish the effectiveness of AAA both in terms of its estimation accuracy and exploration efficiency, through extensive experiments and analysis.
Author Information
Vihari Piratla (IIT Bombay)
Soumen Chakrabarti (Indian Institute of Technology Bombay)
Sunita Sarawagi (IIT Bombay)
More from the Same Authors
-
2021 Spotlight: Redesigning the Transformer Architecture with Insights from Multi-particle Dynamical Systems »
Subhabrata Dutta · Tanya Gautam · Soumen Chakrabarti · Tanmoy Chakraborty -
2022 : Learning Cross-Database Transfer of Text-queries for Adapting Text-to-SQL Parsers »
Abhijeet Awasthi · Ashutosh Sathe · Sunita Sarawagi -
2023 Poster: Locality Sensitive Hashing in Fourier Frequency Domain For Soft Set Containment Search »
Indradyumna Roy · Rishi Agarwal · Soumen Chakrabarti · Anirban Dasgupta · Abir De -
2023 Poster: Use perturbations when learning from explanations »
Juyeon Heo · Vihari Piratla · Matthew Wicker · Adrian Weller -
2023 Poster: Certification of Distributional Individual Fairness »
Matthew Wicker · Vihari Piratla · Adrian Weller -
2022 Spotlight: Neural Estimation of Submodular Functions with Applications to Differentiable Subset Selection »
Abir De · Soumen Chakrabarti -
2022 Spotlight: Maximum Common Subgraph Guided Graph Retrieval: Late and Early Interaction Networks »
Indradyumna Roy · Soumen Chakrabarti · Abir De -
2022 Spotlight: Learning Recourse on Instance Environment to Enhance Prediction Accuracy »
Lokesh N · Guntakanti Sai Koushik · Abir De · Sunita Sarawagi -
2022 Poster: Neural Estimation of Submodular Functions with Applications to Differentiable Subset Selection »
Abir De · Soumen Chakrabarti -
2022 Poster: Learning Recourse on Instance Environment to Enhance Prediction Accuracy »
Lokesh N · Guntakanti Sai Koushik · Abir De · Sunita Sarawagi -
2022 Poster: Maximum Common Subgraph Guided Graph Retrieval: Late and Early Interaction Networks »
Indradyumna Roy · Soumen Chakrabarti · Abir De -
2021 Poster: Redesigning the Transformer Architecture with Insights from Multi-particle Dynamical Systems »
Subhabrata Dutta · Tanya Gautam · Soumen Chakrabarti · Tanmoy Chakraborty -
2021 : Live Q&A with Sunita Sarawagi »
Sunita Sarawagi -
2021 : Invited talk - Machine Learning as a Service: The Challenges of Serving diverse client Distributions, Sunita Sarawagi »
Sunita Sarawagi -
2021 Poster: Training for the Future: A Simple Gradient Interpolation Loss to Generalize Along Time »
Anshul Nasery · Soumyadeep Thakur · Vihari Piratla · Abir De · Sunita Sarawagi -
2010 Oral: MAP estimation in Binary MRFs via Bipartite Multi-cuts »
Sashank Jakkam Reddi · Sunita Sarawagi · Sundar Vishwanathan -
2010 Poster: MAP estimation in Binary MRFs via Bipartite Multi-cuts »
Sashank Jakkam Reddi · Sunita Sarawagi · Sundar Vishwanathan