mirror of
https://github.com/SheffieldML/GPy.git
synced 2026-05-04 17:22:39 +02:00
226 lines
8.6 KiB
Python
226 lines
8.6 KiB
Python
# Copyright (c) 2012, GPy authors (see AUTHORS.txt).
|
|
# Licensed under the BSD 3-clause license (see LICENSE.txt)
|
|
|
|
|
|
import numpy as np
|
|
from scipy import stats
|
|
import scipy as sp
|
|
import pylab as pb
|
|
from ..util.plot import gpplot
|
|
|
|
class likelihood:
|
|
"""
|
|
Likelihood class for doing Expectation propagation
|
|
|
|
:param Y: observed output (Nx1 numpy.darray)
|
|
..Note:: Y values allowed depend on the likelihood used
|
|
"""
|
|
def __init__(self,Y,location=0,scale=1):
|
|
self.Y = Y
|
|
self.N = self.Y.shape[0]
|
|
self.location = location
|
|
self.scale = scale
|
|
|
|
def plot2D(self,X,X_new,F_new,U=None):
|
|
"""
|
|
Predictive distribution of the fitted GP model for 2-dimensional inputs
|
|
|
|
:param X_new: The points at which to make a prediction
|
|
:param Mean_new: mean values at X_new
|
|
:param Var_new: variance values at X_new
|
|
:param X_u: input points used to train the model
|
|
:param Mean_u: mean values at X_u
|
|
:param Var_new: variance values at X_u
|
|
"""
|
|
N,D = X_new.shape
|
|
assert D == 2, 'Number of dimensions must be 2'
|
|
n = np.sqrt(N)
|
|
x1min = X_new[:,0].min()
|
|
x1max = X_new[:,0].max()
|
|
x2min = X_new[:,1].min()
|
|
x2max = X_new[:,1].max()
|
|
pb.imshow(F_new.reshape(n,n),extent=(x1min,x1max,x2max,x2min),vmin=0,vmax=1)
|
|
pb.colorbar()
|
|
C1 = np.arange(self.N)[self.Y.flatten()==1]
|
|
C2 = np.arange(self.N)[self.Y.flatten()==-1]
|
|
[pb.plot(X[i,0],X[i,1],'ro') for i in C1]
|
|
[pb.plot(X[i,0],X[i,1],'bo') for i in C2]
|
|
pb.xlim(x1min,x1max)
|
|
pb.ylim(x2min,x2max)
|
|
if U is not None:
|
|
[pb.plot(a,b,'wo') for a,b in U]
|
|
|
|
class probit(likelihood):
|
|
"""
|
|
Probit likelihood
|
|
Y is expected to take values in {-1,1}
|
|
-----
|
|
$$
|
|
L(x) = \\Phi (Y_i*f_i)
|
|
$$
|
|
"""
|
|
def __init__(self,Y,location=0,scale=1):
|
|
assert np.sum(np.abs(Y)-1) == 0, "Output values must be either -1 or 1"
|
|
likelihood.__init__(self,Y,location,scale)
|
|
|
|
def moments_match(self,i,tau_i,v_i):
|
|
"""
|
|
Moments match of the marginal approximation in EP algorithm
|
|
|
|
:param i: number of observation (int)
|
|
:param tau_i: precision of the cavity distribution (float)
|
|
:param v_i: mean/variance of the cavity distribution (float)
|
|
"""
|
|
z = self.Y[i]*v_i/np.sqrt(tau_i**2 + tau_i)
|
|
Z_hat = stats.norm.cdf(z)
|
|
phi = stats.norm.pdf(z)
|
|
mu_hat = v_i/tau_i + self.Y[i]*phi/(Z_hat*np.sqrt(tau_i**2 + tau_i))
|
|
sigma2_hat = 1./tau_i - (phi/((tau_i**2+tau_i)*Z_hat))*(z+phi/Z_hat)
|
|
return Z_hat, mu_hat, sigma2_hat
|
|
|
|
def predictive_mean(self,mu,var):
|
|
mu = mu.flatten()
|
|
var = var.flatten()
|
|
return stats.norm.cdf(mu/np.sqrt(1+var))
|
|
|
|
def predictive_var(self,mu,var):
|
|
p=self.predictive_mean(mu,var)
|
|
return p*(1-p)
|
|
|
|
def _log_likelihood_gradients():
|
|
raise NotImplementedError
|
|
|
|
def plot(self,X,mu,var,phi,X_obs,Z=None,samples=0):
|
|
assert X_obs.shape[1] == 1, 'Number of dimensions must be 1'
|
|
phi_var = self.predictive_var(mu,var)
|
|
gpplot(X,phi,phi_var)
|
|
if samples:
|
|
phi_samples = np.vstack([np.random.binomial(1,phi.flatten()) for s in range(samples)])
|
|
pb.plot(X,phi_samples.T,'x', alpha = 0.4, c='#3465a4' )
|
|
pb.plot(X_obs,(self.Y+1)/2,'kx',mew=1.5)
|
|
if Z is not None:
|
|
pb.plot(Z,Z*0+.5,'r|',mew=1.5,markersize=12)
|
|
pb.ylim(-0.2,1.2)
|
|
|
|
class poisson(likelihood):
|
|
"""
|
|
Poisson likelihood
|
|
Y is expected to take values in {0,1,2,...}
|
|
-----
|
|
$$
|
|
L(x) = \exp(\lambda) * \lambda**Y_i / Y_i!
|
|
$$
|
|
"""
|
|
def __init__(self,Y,location=0,scale=1):
|
|
assert len(Y[Y<0]) == 0, "Output cannot have negative values"
|
|
likelihood.__init__(self,Y,location,scale)
|
|
|
|
def moments_match(self,i,tau_i,v_i):
|
|
"""
|
|
Moments match of the marginal approximation in EP algorithm
|
|
|
|
:param i: number of observation (int)
|
|
:param tau_i: precision of the cavity distribution (float)
|
|
:param v_i: mean/variance of the cavity distribution (float)
|
|
"""
|
|
mu = v_i/tau_i
|
|
sigma = np.sqrt(1./tau_i)
|
|
def poisson_norm(f):
|
|
"""
|
|
Product of the likelihood and the cavity distribution
|
|
"""
|
|
pdf_norm_f = stats.norm.pdf(f,loc=mu,scale=sigma)
|
|
rate = np.exp( (f*self.scale)+self.location)
|
|
poisson = stats.poisson.pmf(float(self.Y[i]),rate)
|
|
return pdf_norm_f*poisson
|
|
|
|
def log_pnm(f):
|
|
"""
|
|
Log of poisson_norm
|
|
"""
|
|
return -(-.5*(f-mu)**2/sigma**2 - np.exp( (f*self.scale)+self.location) + ( (f*self.scale)+self.location)*self.Y[i])
|
|
|
|
"""
|
|
Golden Search and Simpson's Rule
|
|
--------------------------------
|
|
Simpson's Rule is used to calculate the moments mumerically, it needs a grid of points as input.
|
|
Golden Search is used to find the mode in the poisson_norm distribution and define around it the grid for Simpson's Rule
|
|
"""
|
|
#TODO golden search & simpson's rule can be defined in the general likelihood class, rather than in each specific case.
|
|
|
|
#Golden search
|
|
golden_A = -1 if self.Y[i] == 0 else np.array([np.log(self.Y[i]),mu]).min() #Lower limit
|
|
golden_B = np.array([np.log(self.Y[i]),mu]).max() #Upper limit
|
|
golden_A = (golden_A - self.location)/self.scale
|
|
golden_B = (golden_B - self.location)/self.scale
|
|
opt = sp.optimize.golden(log_pnm,brack=(golden_A,golden_B)) #Better to work with log_pnm than with poisson_norm
|
|
|
|
# Simpson's approximation
|
|
width = 3./np.log(max(self.Y[i],2))
|
|
A = opt - width #Lower limit
|
|
B = opt + width #Upper limit
|
|
K = 10*int(np.log(max(self.Y[i],150))) #Number of points in the grid, we DON'T want K to be the same number for every case
|
|
h = (B-A)/K # length of the intervals
|
|
grid_x = np.hstack([np.linspace(opt-width,opt,K/2+1)[1:-1], np.linspace(opt,opt+width,K/2+1)]) # grid of points (X axis)
|
|
x = np.hstack([A,B,grid_x[range(1,K,2)],grid_x[range(2,K-1,2)]]) # grid_x rearranged, just to make Simpson's algorithm easier
|
|
zeroth = np.hstack([poisson_norm(A),poisson_norm(B),[4*poisson_norm(f) for f in grid_x[range(1,K,2)]],[2*poisson_norm(f) for f in grid_x[range(2,K-1,2)]]]) # grid of points (Y axis) rearranged like x
|
|
first = zeroth*x
|
|
second = first*x
|
|
Z_hat = sum(zeroth)*h/3 # Zero-th moment
|
|
mu_hat = sum(first)*h/(3*Z_hat) # First moment
|
|
m2 = sum(second)*h/(3*Z_hat) # Second moment
|
|
sigma2_hat = m2 - mu_hat**2 # Second central moment
|
|
return float(Z_hat), float(mu_hat), float(sigma2_hat)
|
|
|
|
def predictive_mean(self,mu,var):
|
|
return np.exp(mu*self.scale + self.location)
|
|
|
|
def predictive_var(self,mu,var):
|
|
return predictive_mean(mu,var)
|
|
|
|
def _log_likelihood_gradients():
|
|
raise NotImplementedError
|
|
|
|
def plot(self,X,mu,var,phi,X_obs,Z=None,samples=0):
|
|
assert X_obs.shape[1] == 1, 'Number of dimensions must be 1'
|
|
gpplot(X,phi,phi.flatten())
|
|
pb.plot(X_obs,self.Y,'kx',mew=1.5)
|
|
if samples:
|
|
phi_samples = np.vstack([np.random.poisson(phi.flatten(),phi.size) for s in range(samples)])
|
|
pb.plot(X,phi_samples.T, alpha = 0.4, c='#3465a4', linewidth = 0.8)
|
|
if Z is not None:
|
|
pb.plot(Z,Z*0+pb.ylim()[0],'k|',mew=1.5,markersize=12)
|
|
|
|
class gaussian(likelihood):
|
|
"""
|
|
Gaussian likelihood
|
|
Y is expected to take values in (-inf,inf)
|
|
"""
|
|
def moments_match(self,i,tau_i,v_i):
|
|
"""
|
|
Moments match of the marginal approximation in EP algorithm
|
|
|
|
:param i: number of observation (int)
|
|
:param tau_i: precision of the cavity distribution (float)
|
|
:param v_i: mean/variance of the cavity distribution (float)
|
|
"""
|
|
mu = v_i/tau_i
|
|
sigma = np.sqrt(1./tau_i)
|
|
s = 1. if self.Y[i] == 0 else 1./self.Y[i]
|
|
sigma2_hat = 1./(1./sigma**2 + 1./s**2)
|
|
mu_hat = sigma2_hat*(mu/sigma**2 + self.Y[i]/s**2)
|
|
Z_hat = 1./np.sqrt(2*np.pi) * 1./np.sqrt(sigma**2+s**2) * np.exp(-.5*(mu-self.Y[i])**2/(sigma**2 + s**2))
|
|
return Z_hat, mu_hat, sigma2_hat
|
|
|
|
def plot1Db(self,X,X_new,F_new,U=None):
|
|
assert X.shape[1] == 1, 'Number of dimensions must be 1'
|
|
gpplot(X_new,F_new,np.zeros(X_new.shape[0]))
|
|
pb.plot(X,self.Y,'kx',mew=1.5)
|
|
if U is not None:
|
|
pb.plot(U,np.ones(U.shape[0])*self.Y.min()*.8,'r|',mew=1.5,markersize=12)
|
|
|
|
def predictive_mean(self,mu,Sigma):
|
|
return mu
|
|
|
|
def _log_likelihood_gradients():
|
|
raise NotImplementedError
|