Bayesian Alignments of Warped Multi-Output Gaussian Processes

10/08/2017
by   Markus Kaiser, et al.
0

We present a Bayesian extension to convolution processes which defines a representation between multiple functions by an embedding in a shared latent space. The proposed model allows for both arbitrary alignments of the inputs and and also non-parametric output warpings to transform the observations. This gives rise to multiple deep Gaussian process models connected via latent generating processes. We derive an efficient variational approximation based on nested variational compression and show how the model can be used to extract shared information between dependent time series, recovering an interpretable functional decomposition of the learning problem.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset