Testing noisy linear functions for sparsity

11/03/2019
by   Xue Chen, et al.
0

We consider the following basic inference problem: there is an unknown high-dimensional vector w ∈R^n, and an algorithm is given access to labeled pairs (x,y) where x ∈R^n is a measurement and y = w · x + noise. What is the complexity of deciding whether the target vector w is (approximately) k-sparse? The recovery analogue of this problem — given the promise that w is sparse, find or approximate the vector w— is the famous sparse recovery problem, with a rich body of work in signal processing, statistics, and computer science. We study the decision version of this problem (i.e. deciding whether the unknown w is k-sparse) from the vantage point of property testing. Our focus is on answering the following high-level question: when is it possible to efficiently test whether the unknown target vector w is sparse versus far-from-sparse using a number of samples which is completely independent of the dimension n? We consider the natural setting in which x is drawn from a i.i.d. product distribution D over R^n and the noise process is independent of the input x. As our main result, we give a general algorithm which solves the above-described testing problem using a number of samples which is completely independent of the ambient dimension n, as long as D is not a Gaussian. In fact, our algorithm is fully noise tolerant, in the sense that for an arbitrary w, it approximately computes the distance of w to the closest k-sparse vector. To complement this algorithmic result, we show that weakening any of our condition makes it information-theoretically impossible for any algorithm to solve the testing problem with fewer than essentially log n samples.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/24/2020

Robust testing of low-dimensional functions

A natural problem in high-dimensional inference is to decide if a classi...
research
05/07/2019

The algorithm for the recovery of integer vector via linear measurements

In this paper we continue the studies on the integer sparse recovery pro...
research
08/02/2019

Efficient Truncated Statistics with Unknown Truncation

We study the problem of estimating the parameters of a Gaussian distribu...
research
05/16/2022

Robust Testing in High-Dimensional Sparse Models

We consider the problem of robustly testing the norm of a high-dimension...
research
04/02/2013

Sparse Signal Processing with Linear and Nonlinear Observations: A Unified Shannon-Theoretic Approach

We derive fundamental sample complexity bounds for recovering sparse and...
research
04/21/2016

Robust Estimators in High Dimensions without the Computational Intractability

We study high-dimensional distribution learning in an agnostic setting w...
research
08/05/2019

Imaging with highly incomplete and corrupted data

We consider the problem of imaging sparse scenes from a few noisy data u...

Please sign up or login with your details

Forgot password? Click here to reset