Finite-sample Guarantees for Winsorized Importance Sampling

10/25/2018
by   Paulo Orenstein, et al.
0

Importance sampling is a widely used technique to estimate the properties of a distribution. The resulting estimator is always unbiased, but may sometimes incur huge variance. This paper investigates trading-off some bias for variance by winsorizing the importance sampling estimator. The threshold level at which to winsorize is determined by a concrete version of the Balancing Principle, also known as Lepski's Method, which may be of independent interest. The procedure adaptively chooses a threshold level among a pre-defined set by roughly balancing the bias and variance of the estimator when winsorized at different levels. As a consequence, it provides a principled way to perform winsorization, with finite-sample optimality guarantees. The empirical performance of the winsorized estimator is considered in various examples, both real and synthetic. The estimator outperforms the usual importance sampling estimator in high-variance settings, and remains competitive when the variance of the importance sampling weights is low.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/10/2016

Importance Sampling with Unequal Support

Importance sampling is often used in machine learning when training and ...
research
07/13/2022

BR-SNIS: Bias Reduced Self-Normalized Importance Sampling

Importance Sampling (IS) is a method for approximating expectations unde...
research
01/23/2019

Coupling the reduced-order model and the generative model for an importance sampling estimator

In this work, we develop an importance sampling estimator by coupling th...
research
06/04/2019

Robust Mean Estimation with the Bayesian Median of Means

The sample mean is often used to aggregate different unbiased estimates ...
research
09/13/2021

Low-Shot Validation: Active Importance Sampling for Estimating Classifier Performance on Rare Categories

For machine learning models trained with limited labeled training data, ...
research
10/20/2019

Amortized Rejection Sampling in Universal Probabilistic Programming

Existing approaches to amortized inference in probabilistic programs wit...
research
03/02/2018

Not All Samples Are Created Equal: Deep Learning with Importance Sampling

Deep neural network training spends most of the computation on examples ...

Please sign up or login with your details

Forgot password? Click here to reset