Nearly-Linear Time Private Hypothesis Selection with the Optimal Approximation Factor

Jun 1, 2025·
Maryam Aliakbarpour
,
Zhan Shi
Ria Stevens
Ria Stevens
,
Vincent X. Wang
· 0 min read
Abstract
Estimating the density of a distribution from its samples is a fundamental problem in statistics. Hypothesis selection addresses the setting where, in addition to a sample set, we are given n candidate distributions-referred to as hypotheses-and the goal is to determine which one best describes the underlying data distribution. This problem is known to be solvable very efficiently, requiring roughly $O(\log n)$ samples and running in $\tilde{O}(n)$ time. The quality of the output is measured via the total variation distance to the unknown distribution, and the approximation factor of the algorithm determines how large this distance is compared to the optimal distance achieved by the best candidate hypothesis. It is known that $\alpha=3$ is the optimal approximation factor for this problem. We study hypothesis selection under the constraint of differential privacy. We propose a differentially private algorithm in the central model that runs in nearly-linear time with respect to the number of hypotheses, achieves the optimal approximation factor, and incurs only a modest increase in sample complexity, which remains polylogarithmic in $n$. This resolves an open question posed by [Bun, Kamath, Steinke, Wu, NeurIPS 2019]. Prior to our work, existing upper bounds required quadratic time.
Publication
In the Thirty-Ninth Annual Conference on Neural Information Processing Systems (NeurIPS 2025)
publications
Authors
Ria Stevens
Authors
Ria Stevens (she/her)
PhD Candidate

I am a fourth-year Computer Science PhD student at Rice University, advised by Dr. Maryam Aliakbarpour and Dr. Tasos Kyrillidis. My research interests lie in theoretical computer science and statistical learning theory. I am particularly interested in problems involving differential privacy. For the winter 2026 semester, I am a visiting graduate student in the Federated and Collaborative Learning program at the Simons Institute in UC Berkeley. I am also passionate about teaching and recently developed and taught a discrete math course for incoming freshmen in Rice’s CS-RESP program.

Before coming to Rice, I completed my Bachelor’s in Computer Science and Statistics at McGill University, where I was fortunate to be advised by Drs. Elliot and Courtney Paquette.

Outside of my research, I enjoy playing sports, travelling and discovering new coffee shops. I play ultimate frisbee for Torque, Rice’s women’s club team, and I previously played varsity ice hockey for the McGill Martlets.