Approximability and Generalisation

03/15/2022
by   Andrew J. Turner, et al.
0

Approximate learning machines have become popular in the era of small devices, including quantised, factorised, hashed, or otherwise compressed predictors, and the quest to explain and guarantee good generalisation abilities for such methods has just begun. In this paper we study the role of approximability in learning, both in the full precision and the approximated settings of the predictor that is learned from the data, through a notion of sensitivity of predictors to the action of the approximation operator at hand. We prove upper bounds on the generalisation of such predictors, yielding the following main findings, for any PAC-learnable class and any given approximation operator. 1) We show that under mild conditions, approximable target concepts are learnable from a smaller labelled sample, provided sufficient unlabelled data. 2) We give algorithms that guarantee a good predictor whose approximation also enjoys the same generalisation guarantees. 3) We highlight natural examples of structure in the class of sensitivities, which reduce, and possibly even eliminate the otherwise abundant requirement of additional unlabelled data, and henceforth shed new light onto what makes one problem instance easier to learn than another. These results embed the scope of modern model compression approaches into the general goal of statistical learning theory, which in return suggests appropriate algorithms through minimising uniform bounds.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/12/2019

VC Classes are Adversarially Robustly Learnable, but Only Improperly

We study the question of learning an adversarially robust predictor. We ...
research
08/21/2023

Fat Shattering, Joint Measurability, and PAC Learnability of POVM Hypothesis Classes

We characterize learnability for quantum measurement classes by establis...
research
09/21/2022

Learning-Augmented Algorithms for Online Linear and Semidefinite Programming

Semidefinite programming (SDP) is a unifying framework that generalizes ...
research
03/30/2023

Online Learning and Disambiguations of Partial Concept Classes

In a recent article, Alon, Hanneke, Holzman, and Moran (FOCS '21) introd...
research
01/26/2022

Competition over data: how does data purchase affect users?

As machine learning (ML) is deployed by many competing service providers...
research
02/23/2020

De-randomized PAC-Bayes Margin Bounds: Applications to Non-convex and Non-smooth Predictors

In spite of several notable efforts, explaining the generalization of de...
research
08/11/2021

Repeated undersampling in PrInDT (RePrInDT): Variation in undersampling and prediction, and ranking of predictors in ensembles

In this paper, we extend our PrInDT method (Weihs Buschfeld 2021a) t...

Please sign up or login with your details

Forgot password? Click here to reset