Generalization error bounds for kernel matrix completion and extrapolation

06/20/2019
by   Pere Giménez-Febrer, et al.
4

Prior information can be incorporated in matrix completion to improve estimation accuracy and extrapolate the missing entries. Reproducing kernel Hilbert spaces provide tools to leverage the said prior information, and derive more reliable algorithms. This paper analyzes the generalization error of such approaches, and presents numerical tests confirming the theoretical results.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/01/2018

Matrix completion and extrapolation via kernel regression

Matrix completion and extrapolation (MCEX) are dealt with here over repr...
research
02/14/2017

Mutual Kernel Matrix Completion

With the huge influx of various data nowadays, extracting knowledge from...
research
04/17/2018

Parametric Models for Mutual Kernel Matrix Completion

Recent studies utilize multiple kernel learning to deal with incomplete-...
research
03/19/2019

State-Building through Public Land Disposal? An Application of Matrix Completion for Counterfactual Prediction

How would the frontier have evolved in the absence of homestead policies...
research
06/17/2019

Online Matrix Completion with Side Information

We give an online algorithm and prove novel mistake and regret bounds fo...
research
12/16/2022

Generalization Bounds for Inductive Matrix Completion in Low-noise Settings

We study inductive matrix completion (matrix completion with side inform...
research
03/29/2017

Efficient Spatial Variation Characterization via Matrix Completion

In this paper, we propose a novel method to estimate and characterize sp...

Please sign up or login with your details

Forgot password? Click here to reset