# Copyright (c) 2012, GPy authors (see AUTHORS.txt). # Licensed under the BSD 3-clause license (see LICENSE.txt) import numpy as np from scipy import stats import scipy as sp import pylab as pb from ..util.plot import gpplot class likelihood: def __init__(self,Y): """ Likelihood class for doing Expectation propagation :param Y: observed output (Nx1 numpy.darray) ..Note:: Y values allowed depend on the likelihood used """ self.Y = Y self.N = self.Y.shape[0] def plot1Da(self,X,mean,var,Z=None,mean_Z=None,var_Z=None): """ Plot the predictive distribution of the GP model for 1-dimensional inputs :param X_new: The points at which to make a prediction :param Mean_new: mean values at X_new :param Var_new: variance values at X_new :param X_u: input (inducing) points used to train the model :param Mean_u: mean values at X_u :param Var_new: variance values at X_u """ assert X.shape[1] == 1, 'Number of dimensions must be 1' gpplot(X,mean,var.flatten()) pb.errorbar(Z.flatten(),mean_Z.flatten(),2*np.sqrt(var_Z.flatten()),fmt='r+') pb.plot(Z,mean_Z,'ro') def plot1Db(self,X_obs,X,phi,Z=None): assert X_obs.shape[1] == 1, 'Number of dimensions must be 1' gpplot(X,phi,np.zeros(X.shape[0])) pb.plot(X_obs,(self.Y+1)/2,'kx',mew=1.5) pb.ylim(-0.2,1.2) if Z is not None: pb.plot(Z,Z*0+.5,'r|',mew=1.5,markersize=12) def plot2D(self,X,X_new,F_new,U=None): """ Predictive distribution of the fitted GP model for 2-dimensional inputs :param X_new: The points at which to make a prediction :param Mean_new: mean values at X_new :param Var_new: variance values at X_new :param X_u: input points used to train the model :param Mean_u: mean values at X_u :param Var_new: variance values at X_u """ N,D = X_new.shape assert D == 2, 'Number of dimensions must be 2' n = np.sqrt(N) x1min = X_new[:,0].min() x1max = X_new[:,0].max() x2min = X_new[:,1].min() x2max = X_new[:,1].max() pb.imshow(F_new.reshape(n,n),extent=(x1min,x1max,x2max,x2min),vmin=0,vmax=1) pb.colorbar() C1 = np.arange(self.N)[self.Y.flatten()==1] C2 = np.arange(self.N)[self.Y.flatten()==-1] [pb.plot(X[i,0],X[i,1],'ro') for i in C1] [pb.plot(X[i,0],X[i,1],'bo') for i in C2] pb.xlim(x1min,x1max) pb.ylim(x2min,x2max) if U is not None: [pb.plot(a,b,'wo') for a,b in U] class probit(likelihood): """ Probit likelihood Y is expected to take values in {-1,1} ----- $$ L(x) = \\Phi (Y_i*f_i) $$ """ def moments_match(self,i,tau_i,v_i): """ Moments match of the marginal approximation in EP algorithm :param i: number of observation (int) :param tau_i: precision of the cavity distribution (float) :param v_i: mean/variance of the cavity distribution (float) """ z = self.Y[i]*v_i/np.sqrt(tau_i**2 + tau_i) Z_hat = stats.norm.cdf(z) phi = stats.norm.pdf(z) mu_hat = v_i/tau_i + self.Y[i]*phi/(Z_hat*np.sqrt(tau_i**2 + tau_i)) sigma2_hat = 1./tau_i - (phi/((tau_i**2+tau_i)*Z_hat))*(z+phi/Z_hat) return Z_hat, mu_hat, sigma2_hat def predictive_mean(self,mu,var): mu = mu.flatten() var = var.flatten() return stats.norm.cdf(mu/np.sqrt(1+var)) def _log_likelihood_gradients(): raise NotImplementedError class poisson(likelihood): """ Poisson likelihood Y is expected to take values in {0,1,2,...} ----- $$ L(x) = \exp(\lambda) * \lambda**Y_i / Y_i! $$ """ def moments_match(self,i,tau_i,v_i): """ Moments match of the marginal approximation in EP algorithm :param i: number of observation (int) :param tau_i: precision of the cavity distribution (float) :param v_i: mean/variance of the cavity distribution (float) """ mu = v_i/tau_i sigma = np.sqrt(1./tau_i) def poisson_norm(f): """ Product of the likelihood and the cavity distribution """ pdf_norm_f = stats.norm.pdf(f,loc=mu,scale=sigma) rate = np.exp( (f*self.scale)+self.location) poisson = stats.poisson.pmf(float(self.Y[i]),rate) return pdf_norm_f*poisson def log_pnm(f): """ Log of poisson_norm """ return -(-.5*(f-mu)**2/sigma**2 - np.exp( (f*self.scale)+self.location) + ( (f*self.scale)+self.location)*self.Y[i]) """ Golden Search and Simpson's Rule -------------------------------- Simpson's Rule is used to calculate the moments mumerically, it needs a grid of points as input. Golden Search is used to find the mode in the poisson_norm distribution and define around it the grid for Simpson's Rule """ #TODO golden search & simpson's rule can be defined in the general likelihood class, rather than in each specific case. #Golden search golden_A = -1 if self.Y[i] == 0 else np.array([np.log(self.Y[i]),mu]).min() #Lower limit golden_B = np.array([np.log(self.Y[i]),mu]).max() #Upper limit golden_A = (golden_A - self.location)/self.scale golden_B = (golden_B - self.location)/self.scale opt = sp.optimize.golden(log_pnm,brack=(golden_A,golden_B)) #Better to work with log_pnm than with poisson_norm # Simpson's approximation width = 3./np.log(max(self.Y[i],2)) A = opt - width #Lower limit B = opt + width #Upper limit K = 10*int(np.log(max(self.Y[i],150))) #Number of points in the grid, we DON'T want K to be the same number for every case h = (B-A)/K # length of the intervals grid_x = np.hstack([np.linspace(opt-width,opt,K/2+1)[1:-1], np.linspace(opt,opt+width,K/2+1)]) # grid of points (X axis) x = np.hstack([A,B,grid_x[range(1,K,2)],grid_x[range(2,K-1,2)]]) # grid_x rearranged, just to make Simpson's algorithm easier zeroth = np.hstack([poisson_norm(A),poisson_norm(B),[4*poisson_norm(f) for f in grid_x[range(1,K,2)]],[2*poisson_norm(f) for f in grid_x[range(2,K-1,2)]]]) # grid of points (Y axis) rearranged like x first = zeroth*x second = first*x Z_hat = sum(zeroth)*h/3 # Zero-th moment mu_hat = sum(first)*h/(3*Z_hat) # First moment m2 = sum(second)*h/(3*Z_hat) # Second moment sigma2_hat = m2 - mu_hat**2 # Second central moment return float(Z_hat), float(mu_hat), float(sigma2_hat) def plot1Db(self,X,X_new,F_new,F2_new=None,U=None): pb.subplot(212) #gpplot(X_new,F_new,np.sqrt(F2_new)) pb.plot(X_new,F_new)#,np.sqrt(F2_new)) #FIXME pb.plot(X,self.Y,'kx',mew=1.5) if U is not None: pb.plot(U,np.ones(U.shape[0])*self.Y.min()*.8,'r|',mew=1.5,markersize=12) def predictive_mean(self,mu,variance): return np.exp(mu*self.scale + self.location) def predictive_variance(self,mu,variance): return mu def _log_likelihood_gradients(): raise NotImplementedError class gaussian(likelihood): """ Gaussian likelihood Y is expected to take values in (-inf,inf) """ def moments_match(self,i,tau_i,v_i): """ Moments match of the marginal approximation in EP algorithm :param i: number of observation (int) :param tau_i: precision of the cavity distribution (float) :param v_i: mean/variance of the cavity distribution (float) """ mu = v_i/tau_i sigma = np.sqrt(1./tau_i) s = 1. if self.Y[i] == 0 else 1./self.Y[i] sigma2_hat = 1./(1./sigma**2 + 1./s**2) mu_hat = sigma2_hat*(mu/sigma**2 + self.Y[i]/s**2) Z_hat = 1./np.sqrt(2*np.pi) * 1./np.sqrt(sigma**2+s**2) * np.exp(-.5*(mu-self.Y[i])**2/(sigma**2 + s**2)) return Z_hat, mu_hat, sigma2_hat def plot1Db(self,X,X_new,F_new,U=None): assert X.shape[1] == 1, 'Number of dimensions must be 1' gpplot(X_new,F_new,np.zeros(X_new.shape[0])) pb.plot(X,self.Y,'kx',mew=1.5) if U is not None: pb.plot(U,np.ones(U.shape[0])*self.Y.min()*.8,'r|',mew=1.5,markersize=12) def predictive_mean(self,mu,Sigma): return mu def _log_likelihood_gradients(): raise NotImplementedError