A Moment-Matching Approach to Testable Learning and a New Characterization of Rademacher Complexity

11/23/2022
by   Aravind Gollakota, et al.
0

A remarkable recent paper by Rubinfeld and Vasilyan (2022) initiated the study of testable learning, where the goal is to replace hard-to-verify distributional assumptions (such as Gaussianity) with efficiently testable ones and to require that the learner succeed whenever the unknown distribution passes the corresponding test. In this model, they gave an efficient algorithm for learning halfspaces under testable assumptions that are provably satisfied by Gaussians. In this paper we give a powerful new approach for developing algorithms for testable learning using tools from moment matching and metric distances in probability. We obtain efficient testable learners for any concept class that admits low-degree sandwiching polynomials, capturing most important examples for which we have ordinary agnostic learners. We recover the results of Rubinfeld and Vasilyan as a corollary of our techniques while achieving improved, near-optimal sample complexity bounds for a broad range of concept classes and distributions. Surprisingly, we show that the information-theoretic sample complexity of testable learning is tightly characterized by the Rademacher complexity of the concept class, one of the most well-studied measures in statistical learning theory. In particular, uniform convergence is necessary and sufficient for testable learning. This leads to a fundamental separation from (ordinary) distribution-specific agnostic learning, where uniform convergence is sufficient but not necessary.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/05/2023

Optimal lower bounds for Quantum Learning via Information Theory

Although a concept class may be learnt more efficiently using quantum sa...
research
03/09/2022

Metric Entropy Duality and the Sample Complexity of Outcome Indistinguishability

We give the first sample complexity characterizations for outcome indist...
research
04/14/2022

Testing distributional assumptions of learning algorithms

There are many important high dimensional function classes that have fas...
research
10/26/2022

Learning versus Refutation in Noninteractive Local Differential Privacy

We study two basic statistical tasks in non-interactive local differenti...
research
02/09/2022

An Exploration of Multicalibration Uniform Convergence Bounds

Recent works have investigated the sample complexity necessary for fair ...
research
12/04/2020

Near-Optimal Model Discrimination with Non-Disclosure

Let θ_0,θ_1 ∈ℝ^d be the population risk minimizers associated to some lo...
research
02/13/2023

Do PAC-Learners Learn the Marginal Distribution?

We study a foundational variant of Valiant and Vapnik and Chervonenkis' ...

Please sign up or login with your details

Forgot password? Click here to reset