( events)   Timezone: »  
Tue Dec 14 01:20 AM -- 02:40 PM (PST)
Privacy in Machine Learning (PriML) 2021
Yu-Xiang Wang · Borja Balle · Giovanni Cherubin · Kamalika Chaudhuri · Antti Honkela · Jonathan Lebensold · Casey Meehan · Mi Jung Park · Adrian Weller · Yuqing Zhu

The goal of our workshop is to bring together privacy experts working in academia and industry to discuss the present and future of technologies that enable machine learning with privacy. The workshop will focus on the technical aspects of privacy research and deployment with invited and contributed talks by distinguished researchers in the area. By design, the workshop should serve as a meeting point for regular NeurIPS attendees interested/working on privacy to meet other parts of the privacy community (security researchers, legal scholars, industry practitioners). The focus this year will include emerging problems such as machine unlearning, privacy-fairness tradeoffs and legal challenges in recent deployments of differential privacy (e.g. that of the US Census Bureau). We will conclude the workshop with a panel discussion titled: “Machine Learning and Privacy in Practice: Challenges, Pitfalls and Opportunities”. A diverse set of panelists will address the challenges faced applying these technologies to the real world. The programme of the workshop will emphasize the diversity of points of view on the problem of privacy. We will also ensure that there is ample time for discussions that encourage networking between researchers, which should result in mutually beneficial new long-term collaborations.

Introduction (Opening)
Invited talk: Emiliano de Cristofaro (University College London) --- Privacy in Machine Learning -- It's Complicated (Invited talk)
Emiliano Q&A (Q&A)
Coffee break (coffee break)
Differential Privacy via Group Shuffling (Contributed talk)
SoK: Privacy-preserving Clustering (Extended Abstract) (Contributed talk)
Contributed talk Q&A (Q&A)
Coffee Break (Break)
Poster Session (Poster)
Introduction (Opening)
Invited talk: Helen Nissenbaum (Cornell Tech) --- Practical Privacy, Fairness, Ethics, Policy (Invited talk)
Invited talk: Aaron Roth (UPenn / Amazon): Machine Unlearning. (Invited talk)
Q&A for Helen and Aaron (Q&A)
Coffee break (Break)
Poster Session (Gather.Town)
Coffee break (Break)
Invited talk: Kristin Lauter (Facebook AI Research): ML on Encrypted Data. (Invited talk)
Q&A for Kristin (Q&A)
Privacy-Aware Rejection Sampling (Contributed talk)
Population Level Privacy Leakage in Binary Classification wtih Label Noise (Contributed talk)
Simple Baselines Are Strong Performers for Differentially Private Natural Language Processing (Contributed talk)
Canonical Noise Distributions and Private Hypothesis Tests (Contributed talk)
Q&A for four contributed talks (Q&A)
Closing (closing)
Differential Privacy via Group Shuffling (Poster)
Unsupervised Membership Inference Attacks Against Machine Learning Models (Poster)
Feature-level privacy loss modelling in differentially private machine learning (Poster)
Interaction data are identifiable even across long periods of time (Poster)
Communication Efficient Federated Learning with Secure Aggregation and Differential Privacy (Poster)
Combining Public and Private Data (Poster)
Differentially Private Hamiltonian Monte Carlo (Poster)
Simple Baselines Are Strong Performers for Differentially Private Natural Language Processing (Poster)
Basil: A Fast and Byzantine-Resilient Approach for Decentralized Training (Poster)
An automatic differentiation system for the age of differential privacy (Poster)
Realistic Face Reconstruction from Deep Embeddings (Poster)
Tight Accounting in the Shuffle Model of Differential Privacy (Poster)
A Novel Self-Distillation Architecture to Defeat Membership Inference Attacks (Poster)
Efficient passive membership inference attack in federated learning (Poster)
SoK: Privacy-preserving Clustering (Extended Abstract) (Poster)
Canonical Noise Distributions and Private Hypothesis Tests (Poster)
A Generic Hybrid 2PC Framework with Application to Private Inference of Unmodified Neural Networks (Extended Abstract) (Poster)
Population Level Privacy Leakage in Binary Classification wtih Label Noise (Poster)
Private Confidence Sets (Poster)
DP-SEP: Differentially private stochastic expectation propagation (Poster)
Opacus: User-Friendly Differential Privacy Library in PyTorch (Poster)
ABY2.0: New Efficient Primitives for STPC with Applications to Privacy in Machine Learning (Extended Abstract) (Poster)
Label Private Deep Learning Training based on Secure Multiparty Computation and Differential Privacy (Poster)
Membership Inference Attacks Against NLP Classification Models (Poster)
Reconstructing Test Labels from Noisy Loss Scores (Extended Abstract) (Poster)
A Joint Exponential Mechanism for Differentially Private Top-k Set (Poster)
Understanding Training-Data Leakage from Gradients in Neural Networks for ImageClassifications (Poster)
Mean Estimation with User-level Privacy under Data Heterogeneity (Poster)
Sample-and-threshold differential privacy: Histograms and applications (Poster)
Enforcing fairness in private federated learning via the modified method of differential multipliers (Poster)
Characterizing and Improving MPC-based Private Inference for Transformer-based Models (Poster)
Certified Predictions using MPC-Friendly Publicly Verifiable Covertly Secure Commitments (Poster)
Privacy-Aware Rejection Sampling (Poster)
Reconstructing Training Data with Informed Adversaries (Poster)
Adversarial Detection Avoidance Attacks: Evaluating the robustness of perceptual hashing-based client-side scanning (Poster)
SSSE: Efficiently Erasing Samples from Trained Machine Learning Models (Poster)
Architecture Matters: Investigating the Influence of Differential Privacy on Neural Network Design (Poster)
Iterative Methods for Private Synthetic Data: Unifying Framework and New Methods (Poster)
Zero Knowledge Arguments for Verifiable Sampling (Poster)