GPy/GPy/core/gp.py

237 lines
9.1 KiB
Python
Raw Normal View History

2013-06-05 14:11:49 +01:00
# Copyright (c) 2012, GPy authors (see AUTHORS.txt).
# Licensed under the BSD 3-clause license (see LICENSE.txt)
import numpy as np
2014-01-28 13:39:59 +00:00
import sys
2014-01-22 15:06:53 +00:00
import warnings
from .. import kern
2014-01-24 10:26:44 +00:00
from ..util.linalg import dtrtrs
2014-01-22 15:06:53 +00:00
from model import Model
from parameterization import ObservableArray
from .. import likelihoods
2014-01-24 10:24:17 +00:00
from ..likelihoods.gaussian import Gaussian
from ..inference.latent_function_inference import exact_gaussian_inference
2013-06-05 14:11:49 +01:00
2014-01-22 15:06:53 +00:00
class GP(Model):
2013-06-05 14:11:49 +01:00
"""
2014-01-22 15:06:53 +00:00
General purpose Gaussian process model
2013-06-05 14:11:49 +01:00
:param X: input observations
2014-01-22 15:06:53 +00:00
:param Y: output observations
2013-06-05 14:11:49 +01:00
:param kernel: a GPy kernel, defaults to rbf+white
2013-09-20 17:46:23 +01:00
:param likelihood: a GPy likelihood
2013-06-05 14:11:49 +01:00
:rtype: model object
.. Note:: Multiple independent outputs are allowed using columns of Y
2014-01-22 15:06:53 +00:00
2013-06-05 14:11:49 +01:00
"""
def __init__(self, X, Y, kernel, likelihood, inference_method=None, name='gp'):
2014-01-24 10:24:17 +00:00
super(GP, self).__init__(name)
2014-01-22 15:06:53 +00:00
assert X.ndim == 2
self.X = ObservableArray(X)
self.num_data, self.input_dim = self.X.shape
assert Y.ndim == 2
self.Y = ObservableArray(Y)
assert Y.shape[0] == self.num_data
_, self.output_dim = self.Y.shape
assert isinstance(kernel, kern.kern)
self.kern = kernel
assert isinstance(likelihood, likelihoods.Likelihood)
self.likelihood = likelihood
2013-12-10 12:17:59 -08:00
#find a sensible inference method
if inference_method is None:
if isinstance(likelihood, likelihoods.Gaussian):
inference_method = exact_gaussian_inference.ExactGaussianInference()
2014-01-29 17:02:44 +00:00
else:
inference_method = expectation_propagation
print "defaulting to ", inference_method, "for latent function inference"
2014-01-24 10:24:17 +00:00
self.inference_method = inference_method
2013-06-05 14:11:49 +01:00
2014-01-24 15:07:28 +00:00
self.add_parameter(self.kern)
self.add_parameter(self.likelihood)
2014-01-22 15:06:53 +00:00
#self.parameters_changed()
2014-01-24 10:24:17 +00:00
def parameters_changed(self):
self.posterior, self._log_marginal_likelihood, grad_dict = self.inference_method.inference(self.kern, self.X, self.likelihood, self.Y)
self._dL_dK = grad_dict['dL_dK']
2013-06-05 14:11:49 +01:00
def log_likelihood(self):
return self._log_marginal_likelihood
2013-06-05 14:11:49 +01:00
2014-01-22 15:06:53 +00:00
def dL_dtheta_K(self):
return self.kern.dK_dtheta(self.posterior.dL_dK, self.X)
def _raw_predict(self, _Xnew, which_parts='all', full_cov=False, stop=False):
2013-06-05 14:11:49 +01:00
"""
Internal helper function for making predictions, does not account
for normalization or likelihood
2013-12-04 20:12:40 +00:00
full_cov is a boolean which defines whether the full covariance matrix
of the prediction is computed. If full_cov is False (default), only the
diagonal of the covariance is returned.
2013-06-05 14:11:49 +01:00
"""
Kx = self.kern.K(_Xnew, self.X, which_parts=which_parts).T
LiKx, _ = dtrtrs(self.posterior._woodbury_chol, np.asfortranarray(Kx), lower=1)
2013-12-04 20:12:40 +00:00
mu = np.dot(Kx.T, self.posterior._woodbury_vector)
2013-06-05 14:11:49 +01:00
if full_cov:
Kxx = self.kern.K(_Xnew, which_parts=which_parts)
2013-12-04 20:12:40 +00:00
var = Kxx - tdot(LiKx.T)
2013-06-05 14:11:49 +01:00
else:
Kxx = self.kern.Kdiag(_Xnew, which_parts=which_parts)
var = Kxx - np.sum(LiKx*LiKx, 0)
var = var.reshape(-1, 1)
2013-06-05 14:11:49 +01:00
return mu, var
def predict(self, Xnew, which_parts='all', full_cov=False, **likelihood_args):
2013-06-05 14:11:49 +01:00
"""
Predict the function(s) at the new point(s) Xnew.
2013-09-20 17:46:23 +01:00
2013-06-05 14:11:49 +01:00
:param Xnew: The points at which to make a prediction
:type Xnew: np.ndarray, Nnew x self.input_dim
:param which_parts: specifies which outputs kernel(s) to use in prediction
:type which_parts: ('all', list of bools)
2014-01-28 14:45:00 +00:00
:param full_cov: whether to return the full covariance matrix, or just
the diagonal
2013-06-05 14:11:49 +01:00
:type full_cov: bool
2013-09-20 17:46:23 +01:00
:returns: mean: posterior mean, a Numpy array, Nnew x self.input_dim
2014-01-28 14:45:00 +00:00
:returns: var: posterior variance, a Numpy array, Nnew x 1 if
full_cov=False, Nnew x Nnew otherwise
:returns: lower and upper boundaries of the 95% confidence intervals,
Numpy arrays, Nnew x self.input_dim
2013-06-05 14:11:49 +01:00
If full_cov and self.input_dim > 1, the return shape of var is Nnew x Nnew x self.input_dim. If self.input_dim == 1, the return shape is Nnew x Nnew.
This is to allow for different normalizations of the output dimensions.
"""
# normalize X values
mu, var = self._raw_predict(Xnew, full_cov=full_cov, which_parts=which_parts)
# now push through likelihood
2013-07-18 15:36:04 +01:00
mean, var, _025pm, _975pm = self.likelihood.predictive_values(mu, var, full_cov, **likelihood_args)
2013-07-18 18:49:26 +01:00
return mean, var, _025pm, _975pm
2013-06-05 14:11:49 +01:00
2014-01-22 15:06:53 +00:00
def posterior_samples_f(self,X,size=10,which_parts='all',full_cov=True):
"""
Samples the posterior GP at the points X.
:param X: The points at which to take the samples.
:type X: np.ndarray, Nnew x self.input_dim.
2014-01-28 13:39:59 +00:00
:param size: the number of a posteriori samples.
2014-01-22 15:06:53 +00:00
:type size: int.
2014-01-28 13:39:59 +00:00
:param which_parts: which of the kernel functions to use (additively).
2014-01-22 15:06:53 +00:00
:type which_parts: 'all', or list of bools.
:param full_cov: whether to return the full covariance matrix, or just the diagonal.
:type full_cov: bool.
:returns: Ysim: set of simulations, a Numpy array (N x samples).
"""
m, v = self._raw_predict(X, which_parts=which_parts, full_cov=full_cov)
v = v.reshape(m.size,-1) if len(v.shape)==3 else v
if not full_cov:
Ysim = np.random.multivariate_normal(m.flatten(), np.diag(v.flatten()), size).T
else:
Ysim = np.random.multivariate_normal(m.flatten(), v, size).T
return Ysim
def posterior_samples(self,X,size=10,which_parts='all',full_cov=True,noise_model=None):
"""
Samples the posterior GP at the points X.
:param X: the points at which to take the samples.
:type X: np.ndarray, Nnew x self.input_dim.
2014-01-28 13:39:59 +00:00
:param size: the number of a posteriori samples.
2014-01-22 15:06:53 +00:00
:type size: int.
2014-01-28 13:39:59 +00:00
:param which_parts: which of the kernel functions to use (additively).
2014-01-22 15:06:53 +00:00
:type which_parts: 'all', or list of bools.
:param full_cov: whether to return the full covariance matrix, or just the diagonal.
:type full_cov: bool.
:param noise_model: for mixed noise likelihood, the noise model to use in the samples.
:type noise_model: integer.
:returns: Ysim: set of simulations, a Numpy array (N x samples).
"""
Ysim = self.posterior_samples_f(X, size, which_parts=which_parts, full_cov=full_cov)
if isinstance(self.likelihood, Gaussian):
noise_std = np.sqrt(self.likelihood._get_params())
Ysim += np.random.normal(0,noise_std,Ysim.shape)
elif isinstance(self.likelihood, Gaussian_Mixed_Noise):
assert noise_model is not None, "A noise model must be specified."
noise_std = np.sqrt(self.likelihood._get_params()[noise_model])
Ysim += np.random.normal(0,noise_std,Ysim.shape)
else:
Ysim = self.likelihood.noise_model.samples(Ysim)
return Ysim
def plot_f(self, *args, **kwargs):
"""
2014-01-28 14:45:00 +00:00
Plot the GP's view of the world, where the data is normalized and
before applying a likelihood.
This is a convenience function: arguments are passed to
GPy.plotting.matplot_dep.models_plots.plot_f_fit
2014-01-22 15:06:53 +00:00
"""
2014-01-28 13:39:59 +00:00
assert "matplotlib" in sys.modules, "matplotlib package has not been imported."
from ..plotting.matplot_dep import models_plots
models_plots.plot_fit_f(self,*args,**kwargs)
def plot(self, *args):
2014-01-22 15:06:53 +00:00
"""
Plot the posterior of the GP.
2014-01-28 14:45:00 +00:00
- In one dimension, the function is plotted with a shaded region
identifying two standard deviations.
- In two dimsensions, a contour-plot shows the mean predicted
function
- In higher dimensions, use fixed_inputs to plot the GP with some of
the inputs fixed.
2014-01-22 15:06:53 +00:00
Can plot only part of the data and part of the posterior functions
2014-01-28 13:39:59 +00:00
using which_data_rows which_data_ycols and which_parts
2014-01-22 15:06:53 +00:00
2014-01-28 14:45:00 +00:00
This is a convenience function: arguments are passed to
GPy.plotting.matplot_dep.models_plots.plot_fit
2014-01-28 13:39:59 +00:00
"""
assert "matplotlib" in sys.modules, "matplotlib package has not been imported."
from ..plotting.matplot_dep import models_plots
models_plots.plot_fit(self,*args)
2014-01-22 15:06:53 +00:00
def _getstate(self):
2014-01-22 15:06:53 +00:00
"""
2014-01-28 14:45:00 +00:00
Get the current state of the class, here we return everything that is
needed to recompute the model.
2014-01-22 15:06:53 +00:00
"""
2014-01-28 14:45:00 +00:00
return Model._getstate(self) + [self.X,
2014-01-22 15:06:53 +00:00
self.num_data,
self.input_dim,
self.kern,
self.likelihood,
self.output_dim,
self._Xoffset,
self._Xscale,
]
def _setstate(self, state):
2014-01-22 15:06:53 +00:00
self._Xscale = state.pop()
self._Xoffset = state.pop()
self.output_dim = state.pop()
self.likelihood = state.pop()
self.kern = state.pop()
self.input_dim = state.pop()
self.num_data = state.pop()
self.X = state.pop()
Model._setstate(self, state)