An Easy Implementation of CV-TMLE

11/12/2018
by   Jonathan Levy, et al.
0

In the world of targeted learning, cross-validated targeted maximum likelihood estimators, CV-TMLE Zheng:2010aa, has a distinct advantage over TMLE Laan:2006aa in that one less condition is required of CV-TMLE in order to achieve asymptotic efficiency in the nonparametric or semiparametric settings. CV-TMLE as originally formulated, consists of averaging usually 10 (for 10-fold cross-validation) parameter estimates, each of which is performed on a validation set separate from where the initial fit was trained. The targeting step is usually performed as a pooled regression over all validation folds but in each fold, we separately evaluate any means as well as the parameter estimate. One nice thing about CV-TMLE, is that we average 10 plug-in estimates so the plug-in quality of preserving the natural parameter bounds is respected. Our adjustment of this procedure also preserves the plug-in characteristic as well as avoids the donsker condtion. The advantage of our procedure is the implementation of the targeting is identical to that of a regular TMLE, once all the validation set initial predictions have been formed. In short, we stack the validation set predictions and pretend as if we have a regular TMLE, which is not necessarily quite a plug-in estimator on each fold but overall will perform asymptotically the same and might have some slight advantage, a subject for future research. In the case of average treatment effect, treatment specific mean and mean outcome under a stochastic intervention, the procedure overlaps exactly with the originally formulated CV-TMLE with a pooled regression for the targeting.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/29/2020

Model selection for estimation of causal parameters

A popular technique for selecting and tuning machine learning estimators...
research
09/28/2021

Evaluating the Robustness of Targeted Maximum Likelihood Estimators via Realistic Simulations in Nutrition Intervention Trials

Several recently developed methods have the potential to harness machine...
research
01/08/2021

Fast calculation of Gaussian Process multiple-fold cross-validation residuals and their covariances

We generalize fast Gaussian process leave-one-out formulae to multiple-f...
research
04/21/2021

On the Asymptotic Optimality of Cross-Validation based Hyper-parameter Estimators for Regularized Least Squares Regression Problems

The asymptotic optimality (a.o.) of various hyper-parameter estimators w...
research
06/09/2021

Robust Prediction Interval estimation for Gaussian Processes by Cross-Validation method

Probabilistic regression models typically use the Maximum Likelihood Est...
research
11/03/2018

Canonical Least Favorable Submodels:A New TMLE Procedure for Multidimensional Parameters

This paper is a fundamental addition to the world of targeted maximum li...
research
10/21/2019

Generalised learning of time-series: Ornstein-Uhlenbeck processes

In machine learning, statistics, econometrics and statistical physics, k...

Please sign up or login with your details

Forgot password? Click here to reset