Timezone: »
Interpretable and explainable machine learning has seen a recent surge of interest. We focus on safety as a key motivation behind the surge and make the relationship between interpretability and safety more quantitative. Toward assessing safety, we introduce the concept of maximum deviation via an optimization problem to find the largest deviation of a supervised learning model from a reference model regarded as safe. We then show how interpretability facilitates this safety assessment. For models including decision trees, generalized linear and additive models, the maximum deviation can be computed exactly and efficiently. For tree ensembles, which are not regarded as interpretable, discrete optimization techniques can still provide informative bounds. For a broader class of piecewise Lipschitz functions, we leverage the multi-armed bandit literature to show that interpretability produces tighter (regret) bounds on the maximum deviation. We present case studies, including one on mortgage approval, to illustrate our methods and the insights about models that may be obtained from deviation maximization.
Author Information
Dennis Wei (IBM Research)
Rahul Nair (IBM Research Europe)
Amit Dhurandhar (IBM Research)
Kush Varshney (IBM Research)
Elizabeth Daly (IBM Research)
Moninder Singh (IBM Research AI)
More from the Same Authors
-
2021 : Accurate Multi-Endpoint Molecular Toxicity Predictions in Humans with Contrastive Explanations »
Bhanushee Sharma · Vijil Chenthamarakshan · Amit Dhurandhar · James Hendler · Jonathan S. Dordick · Payel Das -
2023 Poster: Locally Invariant Explanations: Towards Stable and Unidirectional Explanations through Local Invariant Learning »
Amit Dhurandhar · Karthikeyan Natesan Ramamurthy · Kartik Ahuja · Vijay Arya -
2023 Poster: Cookie Consent Has Disparate Impact on Estimation Accuracy »
Erik Miehling · Rahul Nair · Elizabeth Daly · Karthikeyan Natesan Ramamurthy · Robert Redmond -
2023 Poster: Effective Human-AI Teams via Learned Natural Language Rules and Onboarding »
Hussein Mozannar · Jimin Lee · Dennis Wei · Prasanna Sattigeri · Subhro Das · David Sontag -
2022 Poster: Is this the Right Neighborhood? Accurate and Query Efficient Model Agnostic Explanations »
Amit Dhurandhar · Karthikeyan Natesan Ramamurthy · Karthikeyan Shanmugam -
2022 Poster: Fair Infinitesimal Jackknife: Mitigating the Influence of Biased Training Data Points Without Refitting »
Prasanna Sattigeri · Soumya Ghosh · Inkit Padhi · Pierre Dognin · Kush Varshney -
2021 Poster: CoFrNets: Interpretable Neural Architecture Inspired by Continued Fractions »
Isha Puri · Amit Dhurandhar · Tejaswini Pedapati · Karthikeyan Shanmugam · Dennis Wei · Kush Varshney -
2021 : Prospective Explanations: An Interactive Mechanism for Model Understanding »
Rahul Nair · Pierpaolo Tommasi -
2021 : AIMEE: Interactive model maintenance with rule-based surrogates »
Owen Cornec · Rahul Nair · Oznur Alkan · Dennis Wei · Elizabeth Daly -
2020 Poster: DAGs with No Fears: A Closer Look at Continuous Optimization for Learning Bayesian Networks »
Dennis Wei · Tian Gao · Yue Yu -
2020 Spotlight: DAGs with No Fears: A Closer Look at Continuous Optimization for Learning Bayesian Networks »
Dennis Wei · Tian Gao · Yue Yu -
2020 Poster: Model Agnostic Multilevel Explanations »
Karthikeyan Natesan Ramamurthy · Bhanukiran Vinzamuri · Yunfeng Zhang · Amit Dhurandhar -
2020 Poster: Learning Global Transparent Models consistent with Local Contrastive Explanations »
Tejaswini Pedapati · Avinash Balakrishnan · Karthikeyan Shanmugam · Amit Dhurandhar -
2019 : Coffee Break and Poster Session »
Rameswar Panda · Prasanna Sattigeri · Kush Varshney · Karthikeyan Natesan Ramamurthy · Harvineet Singh · Vishwali Mhasawade · Shalmali Joshi · Laleh Seyyed-Kalantari · Matthew McDermott · Gal Yona · James Atwood · Hansa Srinivasan · Yonatan Halpern · D. Sculley · Behrouz Babaki · Margarida Carvalho · Josie Williams · Narges Razavian · Haoran Zhang · Amy Lu · Irene Y Chen · Xiaojie Mao · Angela Zhou · Nathan Kallus -
2019 : Poster Session »
Nathalie Baracaldo · Seth Neel · Tuyen Le · Dan Philps · Suheng Tao · Sotirios Chatzis · Toyo Suzumura · Wei Wang · WENHANG BAO · Solon Barocas · Manish Raghavan · Samuel Maina · Reginald Bryant · Kush Varshney · Skyler D. Speakman · Navdeep Gill · Nicholas Schmidt · Kevin Compher · Naveen Sundar Govindarajulu · Vivek Sharma · Praneeth Vepakomma · Tristan Swedish · Jayashree Kalpathy-Cramer · Ramesh Raskar · Shihao Zheng · Mykola Pechenizkiy · Marco Schreyer · Li Ling · Chirag Nagpal · Robert Tillman · Manuela Veloso · Hanjie Chen · Xintong Wang · Michael Wellman · Matthew van Adelsberg · Ben Wood · Hans Buehler · Mahmoud Mahfouz · Antonios Alexos · Megan Shearer · Antigoni Polychroniadou · Lucia Larise Stavarache · Dmitry Efimov · Johnston P Hall · Yukun Zhang · Emily Diana · Sumitra Ganesh · Vineeth Ravi · · Swetasudha Panda · Xavier Renard · Matthew Jagielski · Yonadav Shavit · Joshua Williams · Haoran Wei · Shuang (Sophie) Zhai · Xinyi Li · Hongda Shen · Daiki Matsunaga · Jaesik Choi · Alexis Laignelet · Batuhan Guler · Jacobo Roa Vicens · Ajit Desai · Jonathan Aigrain · Robert Samoilescu -
2018 Poster: Boolean Decision Rules via Column Generation »
Sanjeeb Dash · Oktay Gunluk · Dennis Wei -
2018 Spotlight: Boolean Decision Rules via Column Generation »
Sanjeeb Dash · Oktay Gunluk · Dennis Wei -
2018 Poster: Improving Simple Models with Confidence Profiles »
Amit Dhurandhar · Karthikeyan Shanmugam · Ronny Luss · Peder A Olsen -
2018 Poster: Explanations based on the Missing: Towards Contrastive Explanations with Pertinent Negatives »
Amit Dhurandhar · Pin-Yu Chen · Ronny Luss · Chun-Chen Tu · Paishun Ting · Karthikeyan Shanmugam · Payel Das -
2017 Poster: Scalable Demand-Aware Recommendation »
Jinfeng Yi · Cho-Jui Hsieh · Kush Varshney · Lijun Zhang · Yao Li -
2017 Poster: Optimized Pre-Processing for Discrimination Prevention »
Flavio Calmon · Dennis Wei · Bhanukiran Vinzamuri · Karthikeyan Natesan Ramamurthy · Kush Varshney -
2016 Poster: A Constant-Factor Bi-Criteria Approximation Guarantee for k-means++ »
Dennis Wei