Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
60 tokens/sec
GPT-4o
12 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Lower Bounds for Differential Privacy Under Continual Observation and Online Threshold Queries (2403.00028v2)

Published 28 Feb 2024 in cs.CR and cs.LG

Abstract: One of the most basic problems for studying the "price of privacy over time" is the so called private counter problem, introduced by Dwork et al. (2010) and Chan et al. (2010). In this problem, we aim to track the number of events that occur over time, while hiding the existence of every single event. More specifically, in every time step $t\in[T]$ we learn (in an online fashion) that $\Delta_t\geq 0$ new events have occurred, and must respond with an estimate $n_t\approx\sum_{j=1}t \Delta_j$. The privacy requirement is that all of the outputs together, across all time steps, satisfy event level differential privacy. The main question here is how our error needs to depend on the total number of time steps $T$ and the total number of events $n$. Dwork et al. (2015) showed an upper bound of $O\left(\log(T)+\log2(n)\right)$, and Henzinger et al. (2023) showed a lower bound of $\Omega\left(\min{\log n, \log T}\right)$. We show a new lower bound of $\Omega\left(\min{n,\log T}\right)$, which is tight w.r.t. the dependence on $T$, and is tight in the sparse case where $\log2 n=O(\log T)$. Our lower bound has the following implications: $\bullet$ We show that our lower bound extends to the "online thresholds problem", where the goal is to privately answer many "quantile queries" when these queries are presented one-by-one. This resolves an open question of Bun et al. (2017). $\bullet$ Our lower bound implies, for the first time, a separation between the number of mistakes obtainable by a private online learner and a non-private online learner. This partially resolves a COLT'22 open question published by Sanyal and Ramponi. $\bullet$ Our lower bound also yields the first separation between the standard model of private online learning and a recently proposed relaxed variant of it, called private online prediction.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (18)
  1. Differentially private release and learning of threshold functions. In FOCS, 2015.
  2. Simultaneous private learning of multiple concepts. In ITCS, 2016.
  3. Make up your mind: The price of online queries in differential privacy. In SODA, 2017.
  4. Private and continual release of statistics. In ICALP, 2010.
  5. Differentially private weighted sampling. In AISTATS, 2021.
  6. Õptimal differentially private learning of thresholds and quasi-concave optimization. In STOC, 2023.
  7. Calibrating noise to sensitivity in private data analysis. In TCC, 2006.
  8. On the complexity of differentially private data release: efficient algorithms and hardness results. In STOC, 2009.
  9. Differential privacy under continual observation. In STOC, 2010a.
  10. Boosting and differential privacy. In FOCS, 2010b.
  11. Pure differential privacy for rectangle queries via private partitions. In ASIACRYPT, 2015.
  12. Littlestone classes are privately online learnable. In NeurIPS, 2021.
  13. Almost tight error bounds on differentially private continual counting. In SODA, 2023.
  14. Svante Janson. Tail bounds for sums of geometric and exponential variables. Statistics & Probability Letters, 135:1–6, 2018.
  15. Privately learning thresholds: Closing the exponential gap. In COLT, 2020.
  16. Black-box differential privacy for interactive ml. In NeurIPS, 2023.
  17. Robust mediators in large games. CoRR, abs/1512.02698, 2015.
  18. Do you pay for privacy in online learning? In Conference on Learning Theory, pages 5633–5637. PMLR, 2022.
Citations (3)

Summary

We haven't generated a summary for this paper yet.