Timezone: »

Counting Distinct Elements in the Turnstile Model with Differential Privacy under Continual Observation
Palak Jain · Iden Kalemaj · Sofya Raskhodnikova · Satchit Sivakumar · Adam Smith

Tue Dec 12 03:15 PM -- 05:15 PM (PST) @ Great Hall & Hall B1+B2 #1605
Privacy is a central challenge for systems that learn from sensitive data sets, especially when a system's outputs must be continuously updated to reflect changing data. We consider the achievable error for differentially private continual release of a basic statistic---the number of distinct items---in a stream where items may be both inserted and deleted (the turnstile model). With only insertions, existing algorithms have additive error just polylogarithmic in the length of the stream $T$. We uncover a much richer landscape in the turnstile model, even without considering memory restrictions. We show that every differentially private mechanism that handles insertions and deletions has worst-case additive error at least $T^{1/4}$ even under a relatively weak, event-level privacy definition. Then, we identify a parameter of the input stream, its maximum flippancy, that is low for natural data streams and for which we give tight parameterized error guarantees. Specifically, the maximum flippancy is the largest number of times that the contribution of a single item to the distinct elements count changes over the course of the stream. We present an item-level differentially private mechanism that, for all turnstile streams with maximum flippancy $w$, continually outputs the number of distinct elements with an $O(\sqrt{w} \cdot \mathsf{poly}\log T)$ additive error, without requiring prior knowledge of $w$. We prove that this is the best achievable error bound that depends only on $w$, for a large range of values of $w$. When $w$ is small, the error of our mechanism is similar to the polylogarithmic in $T$ error in the insertion-only setting, bypassing the hardness in the turnstile model.

Author Information

Palak Jain (Boston University)
Palak Jain

I’m a Ph.D. candidate in Theoretical Computer Science at Boston University advised by professor Adam Smith. My research uses the lenses of cryptography and differential privacy to design privacy-respecting systems and understand the downstream effects of those technologies on the individuals they intend to protect.

Iden Kalemaj (Boston University)
Sofya Raskhodnikova (Boston University)
Satchit Sivakumar (Boston University)
Adam Smith (Boston University)

More from the Same Authors