2014-11-21 12:10:50 +00:00
|
|
|
# Copyright (c) 2012 - 2014 the GPy Austhors (see AUTHORS.txt)
|
2013-06-05 14:11:49 +01:00
|
|
|
# Licensed under the BSD 3-clause license (see LICENSE.txt)
|
|
|
|
|
|
|
|
|
|
import numpy as np
|
2013-12-16 13:45:24 +00:00
|
|
|
from .. import kern
|
2014-09-22 18:12:19 +01:00
|
|
|
from ..core.sparse_gp_mpi import SparseGP_MPI
|
2013-12-16 13:45:24 +00:00
|
|
|
from ..likelihoods import Gaussian
|
2014-10-16 14:39:26 +01:00
|
|
|
from ..core.parameterization.variational import NormalPosterior, NormalPrior
|
|
|
|
|
from ..inference.latent_function_inference.var_dtc_parallel import VarDTC_minibatch
|
2014-05-21 16:32:06 +01:00
|
|
|
import logging
|
2013-06-05 14:11:49 +01:00
|
|
|
|
2014-09-22 18:12:19 +01:00
|
|
|
class BayesianGPLVM(SparseGP_MPI):
|
2013-06-05 14:11:49 +01:00
|
|
|
"""
|
|
|
|
|
Bayesian Gaussian Process Latent Variable Model
|
|
|
|
|
|
|
|
|
|
:param Y: observed data (np.ndarray) or GPy.likelihood
|
|
|
|
|
:type Y: np.ndarray| GPy.likelihood instance
|
|
|
|
|
:param input_dim: latent dimensionality
|
|
|
|
|
:type input_dim: int
|
|
|
|
|
:param init: initialisation method for the latent space
|
|
|
|
|
:type init: 'PCA'|'random'
|
|
|
|
|
|
|
|
|
|
"""
|
2014-02-10 15:12:49 +00:00
|
|
|
def __init__(self, Y, input_dim, X=None, X_variance=None, init='PCA', num_inducing=10,
|
2014-10-08 12:03:51 +01:00
|
|
|
Z=None, kernel=None, inference_method=None, likelihood=None,
|
|
|
|
|
name='bayesian gplvm', mpi_comm=None, normalizer=None,
|
2015-04-10 15:24:28 +01:00
|
|
|
missing_data=False, stochastic=False, batchsize=1, Y_metadata=None):
|
2014-08-27 12:05:13 -07:00
|
|
|
|
2014-05-22 11:39:04 +01:00
|
|
|
self.logger = logging.getLogger(self.__class__.__name__)
|
2014-10-09 10:34:01 +01:00
|
|
|
if X is None:
|
2014-02-24 14:47:43 +00:00
|
|
|
from ..util.initialization import initialize_latent
|
2014-05-21 16:32:06 +01:00
|
|
|
self.logger.info("initializing latent space X with method {}".format(init))
|
2014-03-24 11:22:31 +00:00
|
|
|
X, fracs = initialize_latent(init, input_dim, Y)
|
|
|
|
|
else:
|
|
|
|
|
fracs = np.ones(input_dim)
|
|
|
|
|
|
2013-06-05 14:11:49 +01:00
|
|
|
self.init = init
|
|
|
|
|
|
|
|
|
|
if X_variance is None:
|
2014-07-02 11:04:59 -07:00
|
|
|
self.logger.info("initializing latent space variance ~ uniform(0,.1)")
|
2014-02-24 09:49:29 +00:00
|
|
|
X_variance = np.random.uniform(0,.1,X.shape)
|
2013-06-05 14:11:49 +01:00
|
|
|
|
|
|
|
|
if Z is None:
|
2014-06-27 16:18:41 -07:00
|
|
|
self.logger.info("initializing inducing inputs")
|
2013-06-05 15:29:45 +01:00
|
|
|
Z = np.random.permutation(X.copy())[:num_inducing]
|
2013-06-05 14:11:49 +01:00
|
|
|
assert Z.shape[1] == X.shape[1]
|
|
|
|
|
|
|
|
|
|
if kernel is None:
|
2014-06-27 16:18:41 -07:00
|
|
|
self.logger.info("initializing kernel RBF")
|
2014-08-27 12:05:13 -07:00
|
|
|
kernel = kern.RBF(input_dim, lengthscale=1./fracs, ARD=True) #+ kern.Bias(input_dim) + kern.White(input_dim)
|
2014-02-24 15:44:11 +00:00
|
|
|
|
2014-02-12 17:11:55 +00:00
|
|
|
if likelihood is None:
|
|
|
|
|
likelihood = Gaussian()
|
2014-02-24 15:44:11 +00:00
|
|
|
|
2014-02-24 09:49:29 +00:00
|
|
|
self.variational_prior = NormalPrior()
|
2014-02-24 15:44:11 +00:00
|
|
|
X = NormalPosterior(X, X_variance)
|
|
|
|
|
|
2014-03-24 13:33:16 +00:00
|
|
|
if inference_method is None:
|
2014-10-08 12:03:51 +01:00
|
|
|
if mpi_comm is not None:
|
2014-05-21 12:44:24 +01:00
|
|
|
inference_method = VarDTC_minibatch(mpi_comm=mpi_comm)
|
2014-03-24 13:33:16 +00:00
|
|
|
else:
|
|
|
|
|
from ..inference.latent_function_inference.var_dtc import VarDTC
|
2014-05-21 16:32:06 +01:00
|
|
|
self.logger.debug("creating inference_method var_dtc")
|
2014-10-08 12:03:51 +01:00
|
|
|
inference_method = VarDTC(limit=1 if not missing_data else Y.shape[1])
|
2014-07-02 10:52:47 +01:00
|
|
|
if isinstance(inference_method,VarDTC_minibatch):
|
|
|
|
|
inference_method.mpi_comm = mpi_comm
|
2014-08-27 12:05:13 -07:00
|
|
|
|
2014-10-08 12:03:51 +01:00
|
|
|
super(BayesianGPLVM,self).__init__(X, Y, Z, kernel, likelihood=likelihood,
|
|
|
|
|
name=name, inference_method=inference_method,
|
|
|
|
|
normalizer=normalizer, mpi_comm=mpi_comm,
|
|
|
|
|
variational_prior=self.variational_prior,
|
2015-04-10 17:58:51 +01:00
|
|
|
Y_metadata=Y_metadata
|
2014-11-03 14:19:03 +00:00
|
|
|
)
|
2014-11-05 14:34:03 +00:00
|
|
|
self.link_parameter(self.X, index=0)
|
2014-05-21 12:44:24 +01:00
|
|
|
|
2014-03-21 22:04:25 +00:00
|
|
|
def set_X_gradients(self, X, X_grad):
|
|
|
|
|
"""Set the gradients of the posterior distribution of X in its specific form."""
|
|
|
|
|
X.mean.gradient, X.variance.gradient = X_grad
|
2014-08-27 12:05:13 -07:00
|
|
|
|
2014-05-21 12:44:24 +01:00
|
|
|
def get_X_gradients(self, X):
|
|
|
|
|
"""Get the gradients of the posterior distribution of X in its specific form."""
|
|
|
|
|
return X.mean.gradient, X.variance.gradient
|
2014-03-21 22:04:25 +00:00
|
|
|
|
2014-11-03 14:19:03 +00:00
|
|
|
def parameters_changed(self):
|
|
|
|
|
super(BayesianGPLVM,self).parameters_changed()
|
2014-11-05 16:33:02 +00:00
|
|
|
if isinstance(self.inference_method, VarDTC_minibatch):
|
2015-04-10 15:24:28 +01:00
|
|
|
return
|
2014-10-08 12:03:51 +01:00
|
|
|
|
2014-10-16 12:52:17 +01:00
|
|
|
kl_fctr = 1.
|
2014-11-03 14:19:03 +00:00
|
|
|
self._log_marginal_likelihood -= kl_fctr*self.variational_prior.KL_divergence(self.X)
|
|
|
|
|
|
|
|
|
|
self.X.mean.gradient, self.X.variance.gradient = self.kern.gradients_qX_expectations(
|
|
|
|
|
variational_posterior=self.X,
|
|
|
|
|
Z=self.Z,
|
|
|
|
|
dL_dpsi0=self.grad_dict['dL_dpsi0'],
|
|
|
|
|
dL_dpsi1=self.grad_dict['dL_dpsi1'],
|
|
|
|
|
dL_dpsi2=self.grad_dict['dL_dpsi2'])
|
|
|
|
|
|
|
|
|
|
self.variational_prior.update_gradients_KL(self.X)
|
2015-09-24 13:44:28 +01:00
|
|
|
self._Xgrad = self.X.gradient.copy()
|
2014-10-08 12:03:51 +01:00
|
|
|
|