Lower Bounds for Differential Privacy Under Continual Observation and Online Threshold Queries (2403.00028v2)
Abstract: One of the most basic problems for studying the "price of privacy over time" is the so called private counter problem, introduced by Dwork et al. (2010) and Chan et al. (2010). In this problem, we aim to track the number of events that occur over time, while hiding the existence of every single event. More specifically, in every time step $t\in[T]$ we learn (in an online fashion) that $\Delta_t\geq 0$ new events have occurred, and must respond with an estimate $n_t\approx\sum_{j=1}t \Delta_j$. The privacy requirement is that all of the outputs together, across all time steps, satisfy event level differential privacy. The main question here is how our error needs to depend on the total number of time steps $T$ and the total number of events $n$. Dwork et al. (2015) showed an upper bound of $O\left(\log(T)+\log2(n)\right)$, and Henzinger et al. (2023) showed a lower bound of $\Omega\left(\min{\log n, \log T}\right)$. We show a new lower bound of $\Omega\left(\min{n,\log T}\right)$, which is tight w.r.t. the dependence on $T$, and is tight in the sparse case where $\log2 n=O(\log T)$. Our lower bound has the following implications: $\bullet$ We show that our lower bound extends to the "online thresholds problem", where the goal is to privately answer many "quantile queries" when these queries are presented one-by-one. This resolves an open question of Bun et al. (2017). $\bullet$ Our lower bound implies, for the first time, a separation between the number of mistakes obtainable by a private online learner and a non-private online learner. This partially resolves a COLT'22 open question published by Sanyal and Ramponi. $\bullet$ Our lower bound also yields the first separation between the standard model of private online learning and a recently proposed relaxed variant of it, called private online prediction.
- Differentially private release and learning of threshold functions. In FOCS, 2015.
- Simultaneous private learning of multiple concepts. In ITCS, 2016.
- Make up your mind: The price of online queries in differential privacy. In SODA, 2017.
- Private and continual release of statistics. In ICALP, 2010.
- Differentially private weighted sampling. In AISTATS, 2021.
- Õptimal differentially private learning of thresholds and quasi-concave optimization. In STOC, 2023.
- Calibrating noise to sensitivity in private data analysis. In TCC, 2006.
- On the complexity of differentially private data release: efficient algorithms and hardness results. In STOC, 2009.
- Differential privacy under continual observation. In STOC, 2010a.
- Boosting and differential privacy. In FOCS, 2010b.
- Pure differential privacy for rectangle queries via private partitions. In ASIACRYPT, 2015.
- Littlestone classes are privately online learnable. In NeurIPS, 2021.
- Almost tight error bounds on differentially private continual counting. In SODA, 2023.
- Svante Janson. Tail bounds for sums of geometric and exponential variables. Statistics & Probability Letters, 135:1–6, 2018.
- Privately learning thresholds: Closing the exponential gap. In COLT, 2020.
- Black-box differential privacy for interactive ml. In NeurIPS, 2023.
- Robust mediators in large games. CoRR, abs/1512.02698, 2015.
- Do you pay for privacy in online learning? In Conference on Learning Theory, pages 5633–5637. PMLR, 2022.