Agnostic Learning of Halfspaces with Gradient Descent via Soft Margins

10/01/2020
by   Spencer Frei, et al.
1

We analyze the properties of gradient descent on convex surrogates for the zero-one loss for the agnostic learning of linear halfspaces. If 𝖮𝖯𝖳 is the best classification error achieved by a halfspace, by appealing to the notion of soft margins we are able to show that gradient descent finds halfspaces with classification error Õ(𝖮𝖯𝖳^1/2) + ε in poly(d,1/ε) time and sample complexity for a broad class of distributions that includes log-concave isotropic distributions as a subclass. Along the way we answer a question recently posed by Ji et al. (2020) on how the tail behavior of a loss function can affect sample complexity and runtime guarantees for gradient descent.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/14/2018

Secondary gradient descent in higher codimension

In this paper, we analyze discrete gradient descent and ϵ-noisy gradient...
research
02/13/2019

Learning Ising Models with Independent Failures

We give the first efficient algorithm for learning the structure of an I...
research
05/29/2020

Agnostic Learning of a Single Neuron with Gradient Descent

We consider the problem of learning the best-fitting single neuron as me...
research
08/04/2022

Agnostic Learning of General ReLU Activation Using Gradient Descent

We provide a convergence analysis of gradient descent for the problem of...
research
07/03/2018

On the Computational Power of Online Gradient Descent

We prove that the evolution of weight vectors in online gradient descent...
research
06/22/2021

Robust Regression Revisited: Acceleration and Improved Estimation Rates

We study fast algorithms for statistical regression problems under the s...
research
05/30/2022

Agnostic Physics-Driven Deep Learning

This work establishes that a physical system can perform statistical lea...

Please sign up or login with your details

Forgot password? Click here to reset