mirror of
https://github.com/SheffieldML/GPy.git
synced 2026-05-08 19:42:39 +02:00
Merge branch 'master' into SGD
This commit is contained in:
commit
c8f8494887
23 changed files with 959 additions and 114 deletions
3
.gitignore
vendored
3
.gitignore
vendored
|
|
@ -36,3 +36,6 @@ nosetests.xml
|
||||||
|
|
||||||
#vim
|
#vim
|
||||||
*.swp
|
*.swp
|
||||||
|
|
||||||
|
#bfgs optimiser leaves this lying around
|
||||||
|
iterate.dat
|
||||||
|
|
|
||||||
|
|
@ -29,7 +29,8 @@ class model(parameterised):
|
||||||
raise NotImplementedError, "this needs to be implemented to utilise the model class"
|
raise NotImplementedError, "this needs to be implemented to utilise the model class"
|
||||||
|
|
||||||
def set_prior(self,which,what):
|
def set_prior(self,which,what):
|
||||||
"""sets priors on the model parameters.
|
"""
|
||||||
|
Sets priors on the model parameters.
|
||||||
|
|
||||||
Arguments
|
Arguments
|
||||||
---------
|
---------
|
||||||
|
|
@ -79,6 +80,29 @@ class model(parameterised):
|
||||||
for w in which:
|
for w in which:
|
||||||
self.priors[w] = what
|
self.priors[w] = what
|
||||||
|
|
||||||
|
def get(self,name):
|
||||||
|
"""
|
||||||
|
get a model parameter by name
|
||||||
|
"""
|
||||||
|
matches = self.grep_param_names(name)
|
||||||
|
if len(matches):
|
||||||
|
return self.get_param()[matches]
|
||||||
|
else:
|
||||||
|
raise AttributeError, "no parameter matches %s"%name
|
||||||
|
|
||||||
|
def set(self,name,val):
|
||||||
|
"""
|
||||||
|
Set a model parameter by name
|
||||||
|
"""
|
||||||
|
matches = self.grep_param_names(name)
|
||||||
|
if len(matches):
|
||||||
|
x = self.get_param()
|
||||||
|
x[matches] = val
|
||||||
|
self.set_param(x)
|
||||||
|
else:
|
||||||
|
raise AttributeError, "no parameter matches %s"%name
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def log_prior(self):
|
def log_prior(self):
|
||||||
"""evaluate the prior"""
|
"""evaluate the prior"""
|
||||||
|
|
@ -262,7 +286,7 @@ class model(parameterised):
|
||||||
return '\n'.join(s)
|
return '\n'.join(s)
|
||||||
|
|
||||||
|
|
||||||
def checkgrad(self, verbose = True, include_priors=False, step=1e-6, tolerance = 1e-3, *args):
|
def checkgrad(self, verbose=False, include_priors=False, step=1e-6, tolerance = 1e-3, *args):
|
||||||
"""
|
"""
|
||||||
Check the gradient of the model by comparing to a numerical estimate.
|
Check the gradient of the model by comparing to a numerical estimate.
|
||||||
If the overall gradient fails, invividual components are tested.
|
If the overall gradient fails, invividual components are tested.
|
||||||
|
|
@ -284,21 +308,33 @@ class model(parameterised):
|
||||||
numerical_gradient = (f1-f2)/(2*dx)
|
numerical_gradient = (f1-f2)/(2*dx)
|
||||||
ratio = (f1-f2)/(2*np.dot(dx,gradient))
|
ratio = (f1-f2)/(2*np.dot(dx,gradient))
|
||||||
if verbose:
|
if verbose:
|
||||||
#print "gradient = ",gradient
|
print "Gradient ratio = ", ratio, '\n'
|
||||||
#print "numerical gradient = ",numerical_gradient
|
|
||||||
print " Gradient ratio = ", ratio, '\n'
|
|
||||||
sys.stdout.flush()
|
sys.stdout.flush()
|
||||||
|
|
||||||
if not (np.abs(1.-ratio)>tolerance):
|
if (np.abs(1.-ratio)<tolerance) and not np.isnan(ratio):
|
||||||
if verbose:
|
if verbose:
|
||||||
print 'Gradcheck passed'
|
print 'Gradcheck passed'
|
||||||
else:
|
else:
|
||||||
if verbose:
|
if verbose:
|
||||||
print "Global ratio far from unity. Testing individual gradients\n"
|
print "Global check failed. Testing individual gradients\n"
|
||||||
try:
|
|
||||||
names = self.extract_param_names()
|
try:
|
||||||
except NotImplementedError:
|
names = self.extract_param_names()
|
||||||
names = ['Variable %i'%i for i in range(len(x))]
|
except NotImplementedError:
|
||||||
|
names = ['Variable %i'%i for i in range(len(x))]
|
||||||
|
|
||||||
|
# Prepare for pretty-printing
|
||||||
|
header = ['Name', 'Ratio', 'Difference', 'Analytical', 'Numerical']
|
||||||
|
max_names = max([len(names[i]) for i in range(len(names))] + [len(header[0])])
|
||||||
|
float_len = 10
|
||||||
|
cols = [max_names]
|
||||||
|
cols.extend([max(float_len, len(header[i])) for i in range(1, len(header))])
|
||||||
|
cols = np.array(cols) + 5
|
||||||
|
header_string = ["{h:^{col}}".format(h = header[i], col = cols[i]) for i in range(len(cols))]
|
||||||
|
header_string = map(lambda x: '|'.join(x), [header_string])
|
||||||
|
separator = '-'*len(header_string[0])
|
||||||
|
print '\n'.join([header_string[0], separator])
|
||||||
|
|
||||||
for i in range(len(x)):
|
for i in range(len(x)):
|
||||||
xx = x.copy()
|
xx = x.copy()
|
||||||
xx[i] += step
|
xx[i] += step
|
||||||
|
|
@ -314,11 +350,20 @@ class model(parameterised):
|
||||||
numerical_gradient = (f1-f2)/(2*step)
|
numerical_gradient = (f1-f2)/(2*step)
|
||||||
ratio = (f1-f2)/(2*step*gradient)
|
ratio = (f1-f2)/(2*step*gradient)
|
||||||
difference = np.abs((f1-f2)/2/step - gradient)
|
difference = np.abs((f1-f2)/2/step - gradient)
|
||||||
|
|
||||||
if verbose:
|
if verbose:
|
||||||
print "{0:10s} ratio: {1:15f} difference: {2:15f} analytical: {3:15f} numerical: {4:15f}".format(names[i], float(ratio), float(difference), gradient, float(numerical_gradient)),
|
|
||||||
if (np.abs(ratio-1)<tolerance):
|
if (np.abs(ratio-1)<tolerance):
|
||||||
print " "+'\033[92m' + u"\u2713" + '\033[0m' # green chackmark
|
formatted_name = "\033[92m {0} \033[0m".format(names[i])
|
||||||
else:
|
else:
|
||||||
print " "+'\033[91m' + u"\u2717" + '\033[0m' # red crossmark
|
formatted_name = "\033[91m {0} \033[0m".format(names[i])
|
||||||
|
r = '%.6f' % float(ratio)
|
||||||
|
d = '%.6f' % float(difference)
|
||||||
|
g = '%.6f' % gradient
|
||||||
|
ng = '%.6f' % float(numerical_gradient)
|
||||||
|
grad_string = "{0:^{c0}}|{1:^{c1}}|{2:^{c2}}|{3:^{c3}}|{4:^{c4}}".format(formatted_name,r,d,g, ng, c0 = cols[0]+9, c1 = cols[1], c2 = cols[2], c3 = cols[3], c4 = cols[4])
|
||||||
|
print grad_string
|
||||||
|
|
||||||
|
print ''
|
||||||
|
|
||||||
return False
|
return False
|
||||||
return True
|
return True
|
||||||
|
|
|
||||||
|
|
@ -29,7 +29,7 @@ def truncate_pad(string,width,align='m'):
|
||||||
else:
|
else:
|
||||||
raise ValueError
|
raise ValueError
|
||||||
|
|
||||||
class parameterised:
|
class parameterised(object):
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
"""
|
"""
|
||||||
This is the base class for model and kernel. Mostly just handles tieing and constraining of parameters
|
This is the base class for model and kernel. Mostly just handles tieing and constraining of parameters
|
||||||
|
|
@ -215,7 +215,7 @@ class parameterised:
|
||||||
else:
|
else:
|
||||||
self.constrained_fixed_values.append(self.get_param()[self.constrained_fixed_indices[-1]])
|
self.constrained_fixed_values.append(self.get_param()[self.constrained_fixed_indices[-1]])
|
||||||
|
|
||||||
self.constrained_fixed_values.append(value)
|
#self.constrained_fixed_values.append(value)
|
||||||
self.expand_param(self.extract_param())
|
self.expand_param(self.extract_param())
|
||||||
|
|
||||||
def extract_param(self):
|
def extract_param(self):
|
||||||
|
|
|
||||||
|
|
@ -1,7 +1,6 @@
|
||||||
# Copyright (c) 2012, GPy authors (see AUTHORS.txt).
|
# Copyright (c) 2012, GPy authors (see AUTHORS.txt).
|
||||||
# Licensed under the BSD 3-clause license (see LICENSE.txt)
|
# Licensed under the BSD 3-clause license (see LICENSE.txt)
|
||||||
|
|
||||||
|
|
||||||
import pylab as pb
|
import pylab as pb
|
||||||
import numpy as np
|
import numpy as np
|
||||||
import GPy
|
import GPy
|
||||||
|
|
@ -9,18 +8,15 @@ pb.ion()
|
||||||
pb.close('all')
|
pb.close('all')
|
||||||
|
|
||||||
|
|
||||||
######################################
|
|
||||||
## 1 dimensional example
|
|
||||||
|
|
||||||
# sample inputs and outputs
|
# sample inputs and outputs
|
||||||
S = np.ones((20,1))
|
S = np.ones((20,1))
|
||||||
X = np.random.uniform(-3.,3.,(20,1))
|
X = np.random.uniform(-3.,3.,(20,1))
|
||||||
Y = np.sin(X)+np.random.randn(20,1)*0.05
|
Y = np.sin(X)+np.random.randn(20,1)*0.05
|
||||||
|
|
||||||
k = GPy.kern.bias(1) + GPy.kern.white(1)
|
k = GPy.kern.rbf(1) + GPy.kern.white(1)
|
||||||
|
|
||||||
# create simple GP model
|
# create simple GP model
|
||||||
m = GPy.models.uncertain_input_GP_regression(X,Y,S,kernel=k)
|
m = GPy.models.sparse_GP_regression(X,Y,X_uncertainty=S,kernel=k)
|
||||||
|
|
||||||
# contrain all parameters to be positive
|
# contrain all parameters to be positive
|
||||||
m.constrain_positive('(variance|prec)')
|
m.constrain_positive('(variance|prec)')
|
||||||
|
|
|
||||||
32
GPy/examples/uncollapsed_GP_demo.py
Normal file
32
GPy/examples/uncollapsed_GP_demo.py
Normal file
|
|
@ -0,0 +1,32 @@
|
||||||
|
# Copyright (c) 2012, GPy authors (see AUTHORS.txt).
|
||||||
|
# Licensed under the BSD 3-clause license (see LICENSE.txt)
|
||||||
|
|
||||||
|
|
||||||
|
import numpy as np
|
||||||
|
"""
|
||||||
|
Sparse Gaussian Processes regression with an RBF kernel,
|
||||||
|
using the uncollapsed sparse GP (where the distribution of the
|
||||||
|
inducing points is explicitley represented)
|
||||||
|
"""
|
||||||
|
import pylab as pb
|
||||||
|
import numpy as np
|
||||||
|
import GPy
|
||||||
|
np.random.seed(2)
|
||||||
|
pb.ion()
|
||||||
|
N = 500
|
||||||
|
M = 20
|
||||||
|
|
||||||
|
# sample inputs and outputs
|
||||||
|
X = np.random.uniform(-3.,3.,(N,1))
|
||||||
|
Y = np.sin(X)+np.random.randn(N,1)*0.05
|
||||||
|
|
||||||
|
kernel = GPy.kern.rbf(1) + GPy.kern.white(1)
|
||||||
|
|
||||||
|
# create simple GP model
|
||||||
|
m = GPy.models.uncollapsed_sparse_GP(X, Y, kernel=kernel, M=M)#, X_uncertainty=np.zeros_like(X)+0.01)
|
||||||
|
|
||||||
|
# contrain all parameters to be positive
|
||||||
|
m.ensure_default_constraints()
|
||||||
|
m.checkgrad()
|
||||||
|
# optimize and plot
|
||||||
|
m.plot()
|
||||||
|
|
@ -3,7 +3,13 @@
|
||||||
|
|
||||||
|
|
||||||
from scipy import optimize
|
from scipy import optimize
|
||||||
# import rasmussens_minimize as rasm
|
|
||||||
|
try:
|
||||||
|
import rasmussens_minimize as rasm
|
||||||
|
rasm_available = True
|
||||||
|
except ImportError:
|
||||||
|
rasm_available = False
|
||||||
|
|
||||||
import pdb
|
import pdb
|
||||||
import pylab as pb
|
import pylab as pb
|
||||||
import datetime as dt
|
import datetime as dt
|
||||||
|
|
@ -159,47 +165,49 @@ class opt_simplex(Optimizer):
|
||||||
self.trace = None
|
self.trace = None
|
||||||
|
|
||||||
|
|
||||||
# class opt_rasm(Optimizer):
|
class opt_rasm(Optimizer):
|
||||||
# def __init__(self, *args, **kwargs):
|
def __init__(self, *args, **kwargs):
|
||||||
# Optimizer.__init__(self, *args, **kwargs)
|
Optimizer.__init__(self, *args, **kwargs)
|
||||||
# self.opt_name = "Rasmussen's Conjugate Gradient"
|
self.opt_name = "Rasmussen's Conjugate Gradient"
|
||||||
|
|
||||||
# def opt(self):
|
def opt(self, f_fp = None, f = None, fp = None):
|
||||||
# """
|
"""
|
||||||
# Run Rasmussen's Conjugate Gradient optimizer
|
Run Rasmussen's Conjugate Gradient optimizer
|
||||||
# """
|
"""
|
||||||
|
|
||||||
# assert self.f_fp != None, "Rasmussen's minimizer requires f_fp"
|
assert f_fp != None, "Rasmussen's minimizer requires f_fp"
|
||||||
# statuses = ['Converged', 'Line search failed', 'Maximum number of f evaluations reached',
|
statuses = ['Converged', 'Line search failed', 'Maximum number of f evaluations reached',
|
||||||
# 'NaNs in optimization']
|
'NaNs in optimization']
|
||||||
|
|
||||||
# opt_dict = {}
|
opt_dict = {}
|
||||||
# if self.xtol is not None:
|
if self.xtol is not None:
|
||||||
# print "WARNING: minimize doesn't have an xtol arg, so I'm going to ignore it"
|
print "WARNING: minimize doesn't have an xtol arg, so I'm going to ignore it"
|
||||||
# if self.ftol is not None:
|
if self.ftol is not None:
|
||||||
# print "WARNING: minimize doesn't have an ftol arg, so I'm going to ignore it"
|
print "WARNING: minimize doesn't have an ftol arg, so I'm going to ignore it"
|
||||||
# if self.gtol is not None:
|
if self.gtol is not None:
|
||||||
# print "WARNING: minimize doesn't have an gtol arg, so I'm going to ignore it"
|
print "WARNING: minimize doesn't have an gtol arg, so I'm going to ignore it"
|
||||||
|
|
||||||
# opt_result = rasm.minimize(self.x_init, self.f_fp, (), messages = self.messages,
|
opt_result = rasm.minimize(self.x_init, f_fp, (), messages = self.messages,
|
||||||
# maxnumfuneval = self.max_f_eval)
|
maxnumfuneval = self.max_f_eval)
|
||||||
# self.x_opt = opt_result[0]
|
self.x_opt = opt_result[0]
|
||||||
# self.f_opt = opt_result[1][-1]
|
self.f_opt = opt_result[1][-1]
|
||||||
# self.funct_eval = opt_result[2]
|
self.funct_eval = opt_result[2]
|
||||||
# self.status = statuses[opt_result[3]]
|
self.status = statuses[opt_result[3]]
|
||||||
|
|
||||||
# self.trace = opt_result[1]
|
self.trace = opt_result[1]
|
||||||
|
|
||||||
def get_optimizer(f_min):
|
def get_optimizer(f_min):
|
||||||
# import rasmussens_minimize as rasm
|
# import rasmussens_minimize as rasm
|
||||||
from SGD import opt_SGD
|
from SGD import opt_SGD
|
||||||
|
|
||||||
optimizers = {'fmin_tnc': opt_tnc,
|
optimizers = {'fmin_tnc': opt_tnc,
|
||||||
# 'rasmussen': opt_rasm,
|
|
||||||
'simplex': opt_simplex,
|
'simplex': opt_simplex,
|
||||||
'lbfgsb': opt_lbfgsb,
|
'lbfgsb': opt_lbfgsb,
|
||||||
'sgd': opt_SGD}
|
'sgd': opt_SGD}
|
||||||
|
|
||||||
|
if rasm_available:
|
||||||
|
optimizers['rasmussen'] = opt_rasm
|
||||||
|
|
||||||
for opt_name in optimizers.keys():
|
for opt_name in optimizers.keys():
|
||||||
if opt_name.lower().find(f_min.lower()) != -1:
|
if opt_name.lower().find(f_min.lower()) != -1:
|
||||||
return optimizers[opt_name]
|
return optimizers[opt_name]
|
||||||
|
|
|
||||||
|
|
@ -2,5 +2,5 @@
|
||||||
# Licensed under the BSD 3-clause license (see LICENSE.txt)
|
# Licensed under the BSD 3-clause license (see LICENSE.txt)
|
||||||
|
|
||||||
|
|
||||||
from constructors import rbf, Matern32, Matern52, exponential, linear, white, bias, finite_dimensional, rbf_ARD, spline, Brownian, linear_ARD
|
from constructors import rbf, Matern32, Matern52, exponential, linear, white, bias, finite_dimensional, rbf_ARD, spline, Brownian, linear_ARD, rbf_sympy, sympykern
|
||||||
from kern import kern
|
from kern import kern
|
||||||
|
|
|
||||||
|
|
@ -21,6 +21,7 @@ from Brownian import Brownian as Brownianpart
|
||||||
#TODO these s=constructors are not as clean as we'd like. Tidy the code up
|
#TODO these s=constructors are not as clean as we'd like. Tidy the code up
|
||||||
#using meta-classes to make the objects construct properly wthout them.
|
#using meta-classes to make the objects construct properly wthout them.
|
||||||
|
|
||||||
|
|
||||||
def rbf(D,variance=1., lengthscale=1.):
|
def rbf(D,variance=1., lengthscale=1.):
|
||||||
"""
|
"""
|
||||||
Construct an RBF kernel
|
Construct an RBF kernel
|
||||||
|
|
@ -170,3 +171,32 @@ def Brownian(D,variance=1.):
|
||||||
"""
|
"""
|
||||||
part = Brownianpart(D,variance)
|
part = Brownianpart(D,variance)
|
||||||
return kern(D, [part])
|
return kern(D, [part])
|
||||||
|
|
||||||
|
import sympy as sp
|
||||||
|
from sympykern import spkern
|
||||||
|
from sympy.parsing.sympy_parser import parse_expr
|
||||||
|
|
||||||
|
def rbf_sympy(D,ARD=False,variance=1., lengthscale=1.):
|
||||||
|
"""
|
||||||
|
Radial Basis Function covariance.
|
||||||
|
"""
|
||||||
|
X = [sp.var('x%i'%i) for i in range(D)]
|
||||||
|
Z = [sp.var('z%i'%i) for i in range(D)]
|
||||||
|
rbf_variance = sp.var('rbf_variance',positive=True)
|
||||||
|
if ARD:
|
||||||
|
rbf_lengthscales = [sp.var('rbf_lengthscale_%i'%i,positive=True) for i in range(D)]
|
||||||
|
dist_string = ' + '.join(['(x%i-z%i)**2/rbf_lengthscale_%i**2'%(i,i,i) for i in range(D)])
|
||||||
|
dist = parse_expr(dist_string)
|
||||||
|
f = rbf_variance*sp.exp(-dist/2.)
|
||||||
|
else:
|
||||||
|
rbf_lengthscale = sp.var('rbf_lengthscale',positive=True)
|
||||||
|
dist_string = ' + '.join(['(x%i-z%i)**2'%(i,i) for i in range(D)])
|
||||||
|
dist = parse_expr(dist_string)
|
||||||
|
f = rbf_variance*sp.exp(-dist/(2*rbf_lengthscale**2))
|
||||||
|
return kern(D,[spkern(D,f)])
|
||||||
|
|
||||||
|
def sympykern(D,k):
|
||||||
|
"""
|
||||||
|
A kernel from a symbolic sympy representation
|
||||||
|
"""
|
||||||
|
return kern(D,[spkern(D,k)])
|
||||||
|
|
|
||||||
|
|
@ -182,7 +182,7 @@ class kern(parameterised):
|
||||||
X2 = X
|
X2 = X
|
||||||
slices1, slices2 = self._process_slices(slices1,slices2)
|
slices1, slices2 = self._process_slices(slices1,slices2)
|
||||||
target = np.zeros_like(X)
|
target = np.zeros_like(X)
|
||||||
[p.dK_dX(partial[s1,s2],X[s1,i_s],X2[s2,i_s],target[s1,i_s]) for p,i_s,ps,s1,s2 in zip(self.parts,self.input_slices, self.param_slices,slices1,slices2)]
|
[p.dK_dX(partial[s1,s2],X[s1,i_s],X2[s2,i_s],target[s1,i_s]) for p, i_s, s1, s2 in zip(self.parts, self.input_slices, slices1, slices2)]
|
||||||
return target
|
return target
|
||||||
|
|
||||||
def Kdiag(self,X,slices=None):
|
def Kdiag(self,X,slices=None):
|
||||||
|
|
|
||||||
10
GPy/kern/sympy_helpers.cpp
Normal file
10
GPy/kern/sympy_helpers.cpp
Normal file
|
|
@ -0,0 +1,10 @@
|
||||||
|
#include <math.h>
|
||||||
|
double DiracDelta(double x){
|
||||||
|
if((x<0.000001) & (x>-0.000001))//go on, laught at my c++ skills
|
||||||
|
return 1.0;
|
||||||
|
else
|
||||||
|
return 0.0;
|
||||||
|
};
|
||||||
|
double DiracDelta(double x,int foo){
|
||||||
|
return 0.0;
|
||||||
|
};
|
||||||
3
GPy/kern/sympy_helpers.h
Normal file
3
GPy/kern/sympy_helpers.h
Normal file
|
|
@ -0,0 +1,3 @@
|
||||||
|
#include <math.h>
|
||||||
|
double DiracDelta(double x);
|
||||||
|
double DiracDelta(double x, int foo);
|
||||||
258
GPy/kern/sympykern.py
Normal file
258
GPy/kern/sympykern.py
Normal file
|
|
@ -0,0 +1,258 @@
|
||||||
|
import numpy as np
|
||||||
|
import sympy as sp
|
||||||
|
from sympy.utilities.codegen import codegen
|
||||||
|
from sympy.core.cache import clear_cache
|
||||||
|
from scipy import weave
|
||||||
|
import re
|
||||||
|
import os
|
||||||
|
import sys
|
||||||
|
current_dir = os.path.dirname(os.path.abspath(os.path.dirname(__file__)))
|
||||||
|
import tempfile
|
||||||
|
import pdb
|
||||||
|
from kernpart import kernpart
|
||||||
|
|
||||||
|
class spkern(kernpart):
|
||||||
|
"""
|
||||||
|
A kernel object, where all the hard work in done by sympy.
|
||||||
|
|
||||||
|
:param k: the covariance function
|
||||||
|
:type k: a positive definite sympy function of x1, z1, x2, z2...
|
||||||
|
|
||||||
|
To construct a new sympy kernel, you'll need to define:
|
||||||
|
- a kernel function using a sympy object. Ensure that the kernel is of the form k(x,z).
|
||||||
|
- that's it! we'll extract the variables from the function k.
|
||||||
|
|
||||||
|
Note:
|
||||||
|
- to handle multiple inputs, call them x1, z1, etc
|
||||||
|
- to handle multpile correlated outputs, you'll need to define each covariance function and 'cross' variance function. TODO
|
||||||
|
"""
|
||||||
|
def __init__(self,D,k,param=None):
|
||||||
|
self.name='sympykern'
|
||||||
|
self._sp_k = k
|
||||||
|
sp_vars = [e for e in k.atoms() if e.is_Symbol]
|
||||||
|
self._sp_x= sorted([e for e in sp_vars if e.name[0]=='x'],key=lambda x:int(x.name[1:]))
|
||||||
|
self._sp_z= sorted([e for e in sp_vars if e.name[0]=='z'],key=lambda z:int(z.name[1:]))
|
||||||
|
assert all([x.name=='x%i'%i for i,x in enumerate(self._sp_x)])
|
||||||
|
assert all([z.name=='z%i'%i for i,z in enumerate(self._sp_z)])
|
||||||
|
assert len(self._sp_x)==len(self._sp_z)
|
||||||
|
self.D = len(self._sp_x)
|
||||||
|
assert self.D == D
|
||||||
|
self._sp_theta = sorted([e for e in sp_vars if not (e.name[0]=='x' or e.name[0]=='z')],key=lambda e:e.name)
|
||||||
|
self.Nparam = len(self._sp_theta)
|
||||||
|
|
||||||
|
#deal with param
|
||||||
|
if param is None:
|
||||||
|
param = np.ones(self.Nparam)
|
||||||
|
assert param.size==self.Nparam
|
||||||
|
self.set_param(param)
|
||||||
|
|
||||||
|
#Differentiate!
|
||||||
|
self._sp_dk_dtheta = [sp.diff(k,theta).simplify() for theta in self._sp_theta]
|
||||||
|
self._sp_dk_dx = [sp.diff(k,xi).simplify() for xi in self._sp_x]
|
||||||
|
#self._sp_dk_dz = [sp.diff(k,zi) for zi in self._sp_z]
|
||||||
|
|
||||||
|
#self.compute_psi_stats()
|
||||||
|
self._gen_code()
|
||||||
|
|
||||||
|
self.weave_kwargs = {\
|
||||||
|
'support_code':self._function_code,\
|
||||||
|
'include_dirs':[tempfile.gettempdir(), os.path.join(current_dir,'kern/')],\
|
||||||
|
'headers':['"sympy_helpers.h"'],\
|
||||||
|
'sources':[os.path.join(current_dir,"kern/sympy_helpers.cpp")],\
|
||||||
|
#'extra_compile_args':['-ftree-vectorize', '-mssse3', '-ftree-vectorizer-verbose=5'],\
|
||||||
|
'extra_compile_args':[],\
|
||||||
|
'extra_link_args':['-lgomp'],\
|
||||||
|
'verbose':True}
|
||||||
|
|
||||||
|
def __add__(self,other):
|
||||||
|
return spkern(self._sp_k+other._sp_k)
|
||||||
|
|
||||||
|
def compute_psi_stats(self):
|
||||||
|
#define some normal distributions
|
||||||
|
mus = [sp.var('mu%i'%i,real=True) for i in range(self.D)]
|
||||||
|
Ss = [sp.var('S%i'%i,positive=True) for i in range(self.D)]
|
||||||
|
normals = [(2*sp.pi*Si)**(-0.5)*sp.exp(-0.5*(xi-mui)**2/Si) for xi, mui, Si in zip(self._sp_x, mus, Ss)]
|
||||||
|
|
||||||
|
#do some integration!
|
||||||
|
#self._sp_psi0 = ??
|
||||||
|
self._sp_psi1 = self._sp_k
|
||||||
|
for i in range(self.D):
|
||||||
|
print 'perfoming integrals %i of %i'%(i+1,2*self.D)
|
||||||
|
sys.stdout.flush()
|
||||||
|
self._sp_psi1 *= normals[i]
|
||||||
|
self._sp_psi1 = sp.integrate(self._sp_psi1,(self._sp_x[i],-sp.oo,sp.oo))
|
||||||
|
clear_cache()
|
||||||
|
self._sp_psi1 = self._sp_psi1.simplify()
|
||||||
|
|
||||||
|
#and here's psi2 (eek!)
|
||||||
|
zprime = [sp.Symbol('zp%i'%i) for i in range(self.D)]
|
||||||
|
self._sp_psi2 = self._sp_k.copy()*self._sp_k.copy().subs(zip(self._sp_z,zprime))
|
||||||
|
for i in range(self.D):
|
||||||
|
print 'perfoming integrals %i of %i'%(self.D+i+1,2*self.D)
|
||||||
|
sys.stdout.flush()
|
||||||
|
self._sp_psi2 *= normals[i]
|
||||||
|
self._sp_psi2 = sp.integrate(self._sp_psi2,(self._sp_x[i],-sp.oo,sp.oo))
|
||||||
|
clear_cache()
|
||||||
|
self._sp_psi2 = self._sp_psi2.simplify()
|
||||||
|
|
||||||
|
|
||||||
|
def _gen_code(self):
|
||||||
|
#generate c functions from sympy objects
|
||||||
|
(foo_c,self._function_code),(foo_h,self._function_header) = \
|
||||||
|
codegen([('k',self._sp_k)] \
|
||||||
|
+ [('dk_d%s'%x.name,dx) for x,dx in zip(self._sp_x,self._sp_dk_dx)]\
|
||||||
|
#+ [('dk_d%s'%z.name,dz) for z,dz in zip(self._sp_z,self._sp_dk_dz)]\
|
||||||
|
+ [('dk_d%s'%theta.name,dtheta) for theta,dtheta in zip(self._sp_theta,self._sp_dk_dtheta)]\
|
||||||
|
,"C",'foobar',argument_sequence=self._sp_x+self._sp_z+self._sp_theta)
|
||||||
|
#put the header file where we can find it
|
||||||
|
f = file(os.path.join(tempfile.gettempdir(),'foobar.h'),'w')
|
||||||
|
f.write(self._function_header)
|
||||||
|
f.close()
|
||||||
|
|
||||||
|
#get rid of derivatives of DiracDelta
|
||||||
|
self._function_code = re.sub('DiracDelta\(.+?,.+?\)','0.0',self._function_code)
|
||||||
|
|
||||||
|
#Here's some code to do the looping for K
|
||||||
|
arglist = ", ".join(["X[i*D+%s]"%x.name[1:] for x in self._sp_x]\
|
||||||
|
+ ["Z[j*D+%s]"%z.name[1:] for z in self._sp_z]\
|
||||||
|
+ ["param[%i]"%i for i in range(self.Nparam)])
|
||||||
|
|
||||||
|
self._K_code =\
|
||||||
|
"""
|
||||||
|
int i;
|
||||||
|
int j;
|
||||||
|
int N = target_array->dimensions[0];
|
||||||
|
int M = target_array->dimensions[1];
|
||||||
|
int D = X_array->dimensions[1];
|
||||||
|
//#pragma omp parallel for private(j)
|
||||||
|
for (i=0;i<N;i++){
|
||||||
|
for (j=0;j<M;j++){
|
||||||
|
target[i*M+j] = k(%s);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
%s
|
||||||
|
"""%(arglist,"/*"+str(self._sp_k)+"*/") #adding a string representation forces recompile when needed
|
||||||
|
|
||||||
|
diag_arglist = re.sub('Z','X',arglist)
|
||||||
|
diag_arglist = re.sub('j','i',diag_arglist)
|
||||||
|
#Here's some code to do the looping for Kdiag
|
||||||
|
self._Kdiag_code =\
|
||||||
|
"""
|
||||||
|
int i;
|
||||||
|
int N = target_array->dimensions[0];
|
||||||
|
int D = X_array->dimensions[1];
|
||||||
|
//#pragma omp parallel for
|
||||||
|
for (i=0;i<N;i++){
|
||||||
|
target[i] = k(%s);
|
||||||
|
}
|
||||||
|
%s
|
||||||
|
"""%(diag_arglist,"/*"+str(self._sp_k)+"*/") #adding a string representation forces recompile when needed
|
||||||
|
|
||||||
|
#here's some code to compute gradients
|
||||||
|
funclist = '\n'.join([' '*16 + 'target[%i] += partial[i*M+j]*dk_d%s(%s);'%(i,theta.name,arglist) for i,theta in enumerate(self._sp_theta)])
|
||||||
|
self._dK_dtheta_code =\
|
||||||
|
"""
|
||||||
|
int i;
|
||||||
|
int j;
|
||||||
|
int N = partial_array->dimensions[0];
|
||||||
|
int M = partial_array->dimensions[1];
|
||||||
|
int D = X_array->dimensions[1];
|
||||||
|
//#pragma omp parallel for private(j)
|
||||||
|
for (i=0;i<N;i++){
|
||||||
|
for (j=0;j<M;j++){
|
||||||
|
%s
|
||||||
|
}
|
||||||
|
}
|
||||||
|
%s
|
||||||
|
"""%(funclist,"/*"+str(self._sp_k)+"*/") #adding a string representation forces recompile when needed
|
||||||
|
|
||||||
|
#here's some code to compute gradients for Kdiag TODO: thius is yucky.
|
||||||
|
diag_funclist = re.sub('Z','X',funclist,count=0)
|
||||||
|
diag_funclist = re.sub('j','i',diag_funclist)
|
||||||
|
diag_funclist = re.sub('partial\[i\*M\+i\]','partial[i]',diag_funclist)
|
||||||
|
self._dKdiag_dtheta_code =\
|
||||||
|
"""
|
||||||
|
int i;
|
||||||
|
int N = partial_array->dimensions[0];
|
||||||
|
int D = X_array->dimensions[1];
|
||||||
|
for (i=0;i<N;i++){
|
||||||
|
%s
|
||||||
|
}
|
||||||
|
%s
|
||||||
|
"""%(diag_funclist,"/*"+str(self._sp_k)+"*/") #adding a string representation forces recompile when needed
|
||||||
|
|
||||||
|
#Here's some code to do gradients wrt x
|
||||||
|
gradient_funcs = "\n".join(["target[i*D+%i] += partial[i*M+j]*dk_dx%i(%s);"%(q,q,arglist) for q in range(self.D)])
|
||||||
|
self._dK_dX_code = \
|
||||||
|
"""
|
||||||
|
int i;
|
||||||
|
int j;
|
||||||
|
int N = partial_array->dimensions[0];
|
||||||
|
int M = partial_array->dimensions[1];
|
||||||
|
int D = X_array->dimensions[1];
|
||||||
|
//#pragma omp parallel for private(j)
|
||||||
|
for (i=0;i<N; i++){
|
||||||
|
for (j=0; j<M; j++){
|
||||||
|
%s
|
||||||
|
//if(isnan(target[i*D+2])){printf("%%f\\n",dk_dx2(X[i*D+0], X[i*D+1], X[i*D+2], Z[j*D+0], Z[j*D+1], Z[j*D+2], param[0], param[1], param[2], param[3], param[4], param[5]));}
|
||||||
|
//if(isnan(target[i*D+2])){printf("%%f,%%f,%%i,%%i\\n", X[i*D+2], Z[j*D+2],i,j);}
|
||||||
|
|
||||||
|
}
|
||||||
|
}
|
||||||
|
%s
|
||||||
|
"""%(gradient_funcs,"/*"+str(self._sp_k)+"*/") #adding a string representation forces recompile when needed
|
||||||
|
|
||||||
|
#now for gradients of Kdiag wrt X
|
||||||
|
self._dKdiag_dX_code= \
|
||||||
|
"""
|
||||||
|
int i;
|
||||||
|
int j;
|
||||||
|
int N = partial_array->dimensions[0];
|
||||||
|
int M = 0;
|
||||||
|
int D = X_array->dimensions[1];
|
||||||
|
for (i=0;i<N; i++){
|
||||||
|
j = i;
|
||||||
|
%s
|
||||||
|
}
|
||||||
|
%s
|
||||||
|
"""%(gradient_funcs,"/*"+str(self._sp_k)+"*/") #adding a string representation forces recompile when needed
|
||||||
|
|
||||||
|
|
||||||
|
#TODO: insert multiple functions here via string manipulation
|
||||||
|
#TODO: similar functions for psi_stats
|
||||||
|
|
||||||
|
def K(self,X,Z,target):
|
||||||
|
param = self._param
|
||||||
|
weave.inline(self._K_code,arg_names=['target','X','Z','param'],**self.weave_kwargs)
|
||||||
|
|
||||||
|
def Kdiag(self,X,target):
|
||||||
|
param = self._param
|
||||||
|
weave.inline(self._Kdiag_code,arg_names=['target','X','param'],**self.weave_kwargs)
|
||||||
|
|
||||||
|
def dK_dtheta(self,partial,X,Z,target):
|
||||||
|
param = self._param
|
||||||
|
weave.inline(self._dK_dtheta_code,arg_names=['target','X','Z','param','partial'],**self.weave_kwargs)
|
||||||
|
|
||||||
|
def dKdiag_dtheta(self,partial,X,target):
|
||||||
|
param = self._param
|
||||||
|
Z = X
|
||||||
|
weave.inline(self._dKdiag_dtheta_code,arg_names=['target','X','Z','param','partial'],**self.weave_kwargs)
|
||||||
|
|
||||||
|
def dK_dX(self,partial,X,Z,target):
|
||||||
|
param = self._param
|
||||||
|
weave.inline(self._dK_dX_code,arg_names=['target','X','Z','param','partial'],**self.weave_kwargs)
|
||||||
|
|
||||||
|
def dKdiag_dX(self,partial,X,target):
|
||||||
|
param = self._param
|
||||||
|
Z = X
|
||||||
|
weave.inline(self._dKdiag_dX_code,arg_names=['target','X','Z','param','partial'],**self.weave_kwargs)
|
||||||
|
|
||||||
|
def set_param(self,param):
|
||||||
|
#print param.flags['C_CONTIGUOUS']
|
||||||
|
self._param = param.copy()
|
||||||
|
|
||||||
|
def get_param(self):
|
||||||
|
return self._param
|
||||||
|
|
||||||
|
def get_param_names(self):
|
||||||
|
return [x.name for x in self._sp_theta]
|
||||||
|
|
@ -47,7 +47,9 @@ class GP_regression(model):
|
||||||
if normalize_X:
|
if normalize_X:
|
||||||
self._Xmean = X.mean(0)[None,:]
|
self._Xmean = X.mean(0)[None,:]
|
||||||
self._Xstd = X.std(0)[None,:]
|
self._Xstd = X.std(0)[None,:]
|
||||||
self.X = (X.copy()- self._Xmean) / self._Xstd
|
self.X = (X.copy() - self._Xmean) / self._Xstd
|
||||||
|
if hasattr(self,'Z'):
|
||||||
|
self.Z = (self.Z - self._Xmean) / self._Xstd
|
||||||
else:
|
else:
|
||||||
self._Xmean = np.zeros((1,self.X.shape[1]))
|
self._Xmean = np.zeros((1,self.X.shape[1]))
|
||||||
self._Xstd = np.ones((1,self.X.shape[1]))
|
self._Xstd = np.ones((1,self.X.shape[1]))
|
||||||
|
|
@ -83,9 +85,8 @@ class GP_regression(model):
|
||||||
"""
|
"""
|
||||||
Computes the model fit using Youter if it's available
|
Computes the model fit using Youter if it's available
|
||||||
"""
|
"""
|
||||||
|
|
||||||
if self.Youter is None:
|
if self.Youter is None:
|
||||||
return -0.5*np.trace(mdot(self.Y.T,self.Ki,self.Y))
|
return -0.5*np.sum(np.square(np.dot(self.Li,self.Y)))
|
||||||
else:
|
else:
|
||||||
return -0.5*np.sum(np.multiply(self.Ki, self.Youter))
|
return -0.5*np.sum(np.multiply(self.Ki, self.Youter))
|
||||||
|
|
||||||
|
|
@ -105,7 +106,7 @@ class GP_regression(model):
|
||||||
def log_likelihood_gradients(self):
|
def log_likelihood_gradients(self):
|
||||||
return self.kern.dK_dtheta(partial=self.dL_dK(),X=self.X)
|
return self.kern.dK_dtheta(partial=self.dL_dK(),X=self.X)
|
||||||
|
|
||||||
def predict(self,Xnew, slices=None):
|
def predict(self,Xnew, slices=None, full_cov=False):
|
||||||
"""
|
"""
|
||||||
|
|
||||||
Predict the function(s) at the new point(s) Xnew.
|
Predict the function(s) at the new point(s) Xnew.
|
||||||
|
|
@ -116,6 +117,8 @@ class GP_regression(model):
|
||||||
:type Xnew: np.ndarray, Nnew x self.Q
|
:type Xnew: np.ndarray, Nnew x self.Q
|
||||||
:param slices: specifies which outputs kernel(s) the Xnew correspond to (see below)
|
:param slices: specifies which outputs kernel(s) the Xnew correspond to (see below)
|
||||||
:type slices: (None, list of slice objects, list of ints)
|
:type slices: (None, list of slice objects, list of ints)
|
||||||
|
:param full_cov: whether to return the folll covariance matrix, or just the diagonal
|
||||||
|
:type full_cov: bool
|
||||||
:rtype: posterior mean, a Numpy array, Nnew x self.D
|
:rtype: posterior mean, a Numpy array, Nnew x self.D
|
||||||
:rtype: posterior variance, a Numpy array, Nnew x Nnew x (self.D)
|
:rtype: posterior variance, a Numpy array, Nnew x Nnew x (self.D)
|
||||||
|
|
||||||
|
|
@ -125,29 +128,42 @@ class GP_regression(model):
|
||||||
- If a list of slices, the i^th slice specifies which data are affected by the i^th kernel part
|
- If a list of slices, the i^th slice specifies which data are affected by the i^th kernel part
|
||||||
- If a list of booleans, specifying which kernel parts are active
|
- If a list of booleans, specifying which kernel parts are active
|
||||||
|
|
||||||
If self.D > 1, the return shape of var is Nnew x Nnew x self.D. If self.D == 1, the return shape is Nnew x Nnew.
|
If full_cov and self.D > 1, the return shape of var is Nnew x Nnew x self.D. If self.D == 1, the return shape is Nnew x Nnew.
|
||||||
This is to allow for different normalisations of the output dimensions.
|
This is to allow for different normalisations of the output dimensions.
|
||||||
|
|
||||||
|
|
||||||
"""
|
"""
|
||||||
|
|
||||||
#normalise X values
|
#normalise X values
|
||||||
Xnew = (Xnew.copy() - self._Xmean) / self._Xstd
|
Xnew = (Xnew.copy() - self._Xmean) / self._Xstd
|
||||||
mu, var = self._raw_predict(Xnew,slices)
|
mu, var = self._raw_predict(Xnew, slices, full_cov)
|
||||||
|
|
||||||
#un-normalise
|
#un-normalise
|
||||||
mu = mu*self._Ystd + self._Ymean
|
mu = mu*self._Ystd + self._Ymean
|
||||||
if self.D==1:
|
if full_cov:
|
||||||
var *= np.square(self._Ystd)
|
if self.D==1:
|
||||||
|
var *= np.square(self._Ystd)
|
||||||
|
else:
|
||||||
|
var = var[:,:,None] * np.square(self._Ystd)
|
||||||
else:
|
else:
|
||||||
var = var[:,:,None] * np.square(self._Ystd)
|
if self.D==1:
|
||||||
|
var *= np.square(np.squeeze(self._Ystd))
|
||||||
|
else:
|
||||||
|
var = var[:,None] * np.square(self._Ystd)
|
||||||
|
|
||||||
return mu,var
|
return mu,var
|
||||||
|
|
||||||
def _raw_predict(self,_Xnew,slices):
|
def _raw_predict(self,_Xnew,slices, full_cov=False):
|
||||||
"""Internal helper function for making predictions, does not account for normalisation"""
|
"""Internal helper function for making predictions, does not account for normalisation"""
|
||||||
Kx = self.kern.K(self.X,_Xnew, slices1=self.Xslices,slices2=slices)
|
Kx = self.kern.K(self.X,_Xnew, slices1=self.Xslices,slices2=slices)
|
||||||
Kxx = self.kern.K(_Xnew, slices1=slices,slices2=slices)
|
|
||||||
mu = np.dot(np.dot(Kx.T,self.Ki),self.Y)
|
mu = np.dot(np.dot(Kx.T,self.Ki),self.Y)
|
||||||
var = Kxx - np.dot(np.dot(Kx.T,self.Ki),Kx)
|
KiKx = np.dot(self.Ki,Kx)
|
||||||
|
if full_cov:
|
||||||
|
Kxx = self.kern.K(_Xnew, slices1=slices,slices2=slices)
|
||||||
|
var = Kxx - np.dot(KiKx.T,Kx)
|
||||||
|
else:
|
||||||
|
Kxx = self.kern.Kdiag(_Xnew, slices=slices)
|
||||||
|
var = Kxx - np.sum(np.multiply(KiKx,Kx),0)
|
||||||
return mu, var
|
return mu, var
|
||||||
|
|
||||||
def plot(self,samples=0,plot_limits=None,which_data='all',which_functions='all',resolution=None):
|
def plot(self,samples=0,plot_limits=None,which_data='all',which_functions='all',resolution=None):
|
||||||
|
|
|
||||||
|
|
@ -9,3 +9,4 @@ from warped_GP import warpedGP
|
||||||
from GP_EP import GP_EP
|
from GP_EP import GP_EP
|
||||||
from generalized_FITC import generalized_FITC
|
from generalized_FITC import generalized_FITC
|
||||||
from sparse_GPLVM import sparse_GPLVM
|
from sparse_GPLVM import sparse_GPLVM
|
||||||
|
from uncollapsed_sparse_GP import uncollapsed_sparse_GP
|
||||||
|
|
|
||||||
|
|
@ -54,5 +54,6 @@ class sparse_GPLVM(sparse_GP_regression, GPLVM):
|
||||||
|
|
||||||
def plot(self):
|
def plot(self):
|
||||||
GPLVM.plot(self)
|
GPLVM.plot(self)
|
||||||
|
#passing Z without a small amout of jitter will induce the white kernel where we don;t want it!
|
||||||
mu, var = sparse_GP_regression.predict(self, self.Z+np.random.randn(*self.Z.shape)*0.0001)
|
mu, var = sparse_GP_regression.predict(self, self.Z+np.random.randn(*self.Z.shape)*0.0001)
|
||||||
pb.plot(mu[:, 0] , mu[:, 1], 'ko')
|
pb.plot(mu[:, 0] , mu[:, 1], 'ko')
|
||||||
|
|
|
||||||
|
|
@ -1,7 +1,6 @@
|
||||||
# Copyright (c) 2012, GPy authors (see AUTHORS.txt).
|
# Copyright (c) 2012, GPy authors (see AUTHORS.txt).
|
||||||
# Licensed under the BSD 3-clause license (see LICENSE.txt)
|
# Licensed under the BSD 3-clause license (see LICENSE.txt)
|
||||||
|
|
||||||
|
|
||||||
import numpy as np
|
import numpy as np
|
||||||
import pylab as pb
|
import pylab as pb
|
||||||
from ..util.linalg import mdot, jitchol, chol_inv, pdinv
|
from ..util.linalg import mdot, jitchol, chol_inv, pdinv
|
||||||
|
|
@ -45,12 +44,12 @@ class sparse_GP_regression(GP_regression):
|
||||||
else:
|
else:
|
||||||
assert Z.shape[1]==X.shape[1]
|
assert Z.shape[1]==X.shape[1]
|
||||||
self.Z = Z
|
self.Z = Z
|
||||||
self.M = Z.shape[1]
|
self.M = Z.shape[0]
|
||||||
if X_uncertainty is None:
|
if X_uncertainty is None:
|
||||||
self.has_uncertain_inputs=False
|
self.has_uncertain_inputs=False
|
||||||
else:
|
else:
|
||||||
assert X_uncertainty.shape==X.shape
|
assert X_uncertainty.shape==X.shape
|
||||||
self.has_uncertain_inputs=False
|
self.has_uncertain_inputs=True
|
||||||
self.X_uncertainty = X_uncertainty
|
self.X_uncertainty = X_uncertainty
|
||||||
|
|
||||||
GP_regression.__init__(self, X, Y, kernel=kernel, normalize_X=normalize_X, normalize_Y=normalize_Y)
|
GP_regression.__init__(self, X, Y, kernel=kernel, normalize_X=normalize_X, normalize_Y=normalize_Y)
|
||||||
|
|
@ -88,23 +87,22 @@ class sparse_GP_regression(GP_regression):
|
||||||
self.psi1V = np.dot(self.psi1, self.V)
|
self.psi1V = np.dot(self.psi1, self.V)
|
||||||
self.psi1VVpsi1 = np.dot(self.psi1V, self.psi1V.T)
|
self.psi1VVpsi1 = np.dot(self.psi1V, self.psi1V.T)
|
||||||
self.Kmmi, self.Lm, self.Lmi, self.Kmm_logdet = pdinv(self.Kmm)
|
self.Kmmi, self.Lm, self.Lmi, self.Kmm_logdet = pdinv(self.Kmm)
|
||||||
self.A = mdot(self.Lmi, self.psi2, self.Lmi.T)
|
self.A = mdot(self.Lmi, self.beta*self.psi2, self.Lmi.T)
|
||||||
self.B = np.eye(self.M) + self.beta * self.A
|
self.B = np.eye(self.M) + self.A
|
||||||
self.Bi, self.LB, self.LBi, self.B_logdet = pdinv(self.B)
|
self.Bi, self.LB, self.LBi, self.B_logdet = pdinv(self.B)
|
||||||
self.LLambdai = np.dot(self.LBi, self.Lmi)
|
self.LLambdai = np.dot(self.LBi, self.Lmi)
|
||||||
self.trace_K = self.psi0 - np.trace(self.A)
|
self.trace_K = self.psi0 - np.trace(self.A)/self.beta
|
||||||
self.LBL_inv = mdot(self.Lmi.T, self.Bi, self.Lmi)
|
self.LBL_inv = mdot(self.Lmi.T, self.Bi, self.Lmi)
|
||||||
self.C = mdot(self.LLambdai, self.psi1V)
|
self.C = mdot(self.LLambdai, self.psi1V)
|
||||||
self.G = mdot(self.LBL_inv, self.psi1VVpsi1, self.LBL_inv.T)
|
self.G = mdot(self.LBL_inv, self.psi1VVpsi1, self.LBL_inv.T)
|
||||||
|
|
||||||
# Compute dL_dpsi
|
# Compute dL_dpsi
|
||||||
self.dL_dpsi0 = - 0.5 * self.D * self.beta * np.ones(self.N)
|
self.dL_dpsi0 = - 0.5 * self.D * self.beta * np.ones(self.N)
|
||||||
dC_dpsi1 = (self.LLambdai.T[:,:, None, None] * self.V)
|
self.dL_dpsi1 = mdot(self.LLambdai.T,self.C,self.V.T)
|
||||||
self.dL_dpsi1 = (dC_dpsi1*self.C[None,:,None,:]).sum(1).sum(-1)
|
|
||||||
self.dL_dpsi2 = - 0.5 * self.beta * (self.D*(self.LBL_inv - self.Kmmi) + self.G)
|
self.dL_dpsi2 = - 0.5 * self.beta * (self.D*(self.LBL_inv - self.Kmmi) + self.G)
|
||||||
|
|
||||||
# Compute dL_dKmm
|
# Compute dL_dKmm
|
||||||
self.dL_dKmm = -0.5 * self.beta * self.D * mdot(self.Lmi.T, self.A, self.Lmi) # dB
|
self.dL_dKmm = -0.5 * self.D * mdot(self.Lmi.T, self.A, self.Lmi) # dB
|
||||||
self.dL_dKmm += -0.5 * self.D * (- self.LBL_inv - 2.*self.beta*mdot(self.LBL_inv, self.psi2, self.Kmmi) + self.Kmmi) # dC
|
self.dL_dKmm += -0.5 * self.D * (- self.LBL_inv - 2.*self.beta*mdot(self.LBL_inv, self.psi2, self.Kmmi) + self.Kmmi) # dC
|
||||||
self.dL_dKmm += np.dot(np.dot(self.G,self.beta*self.psi2) - np.dot(self.LBL_inv, self.psi1VVpsi1), self.Kmmi) + 0.5*self.G # dE
|
self.dL_dKmm += np.dot(np.dot(self.G,self.beta*self.psi2) - np.dot(self.LBL_inv, self.psi1VVpsi1), self.Kmmi) + 0.5*self.G # dE
|
||||||
|
|
||||||
|
|
@ -128,15 +126,14 @@ class sparse_GP_regression(GP_regression):
|
||||||
def dL_dbeta(self):
|
def dL_dbeta(self):
|
||||||
"""
|
"""
|
||||||
Compute the gradient of the log likelihood wrt beta.
|
Compute the gradient of the log likelihood wrt beta.
|
||||||
TODO: suport heteroscedatic noise
|
|
||||||
"""
|
"""
|
||||||
|
#TODO: suport heteroscedatic noise
|
||||||
dA_dbeta = 0.5 * self.N*self.D/self.beta
|
dA_dbeta = 0.5 * self.N*self.D/self.beta
|
||||||
dB_dbeta = - 0.5 * self.D * self.trace_K
|
dB_dbeta = - 0.5 * self.D * self.trace_K
|
||||||
dC_dbeta = - 0.5 * self.D * np.sum(self.Bi*self.A)
|
dC_dbeta = - 0.5 * self.D * np.sum(self.Bi*self.A)/self.beta
|
||||||
dD_dbeta = - 0.5 * self.trYYT
|
dD_dbeta = - 0.5 * self.trYYT
|
||||||
tmp = mdot(self.LBi.T, self.LLambdai, self.psi1V)
|
tmp = mdot(self.LBi.T, self.LLambdai, self.psi1V)
|
||||||
dE_dbeta = np.sum(np.square(self.C))/self.beta - 0.5 * np.sum(self.A * np.dot(tmp, tmp.T))
|
dE_dbeta = (np.sum(np.square(self.C)) - 0.5 * np.sum(self.A * np.dot(tmp, tmp.T)))/self.beta
|
||||||
|
|
||||||
return np.squeeze(dA_dbeta + dB_dbeta + dC_dbeta + dD_dbeta + dE_dbeta)
|
return np.squeeze(dA_dbeta + dB_dbeta + dC_dbeta + dD_dbeta + dE_dbeta)
|
||||||
|
|
||||||
|
|
@ -174,14 +171,19 @@ class sparse_GP_regression(GP_regression):
|
||||||
def log_likelihood_gradients(self):
|
def log_likelihood_gradients(self):
|
||||||
return np.hstack([self.dL_dZ().flatten(), self.dL_dbeta(), self.dL_dtheta()])
|
return np.hstack([self.dL_dZ().flatten(), self.dL_dbeta(), self.dL_dtheta()])
|
||||||
|
|
||||||
def _raw_predict(self, Xnew, slices):
|
def _raw_predict(self, Xnew, slices, full_cov=False):
|
||||||
"""Internal helper function for making predictions, does not account for normalisation"""
|
"""Internal helper function for making predictions, does not account for normalisation"""
|
||||||
|
|
||||||
Kx = self.kern.K(self.Z, Xnew)
|
Kx = self.kern.K(self.Z, Xnew)
|
||||||
Kxx = self.kern.K(Xnew)
|
|
||||||
|
|
||||||
mu = mdot(Kx.T, self.LBL_inv, self.psi1V)
|
mu = mdot(Kx.T, self.LBL_inv, self.psi1V)
|
||||||
var = Kxx - mdot(Kx.T, (self.Kmmi - self.LBL_inv), Kx) + np.eye(Xnew.shape[0])/self.beta # TODO: This beta doesn't belong here in the EP case.
|
|
||||||
|
if full_cov:
|
||||||
|
Kxx = self.kern.K(Xnew)
|
||||||
|
var = Kxx - mdot(Kx.T, (self.Kmmi - self.LBL_inv), Kx) + np.eye(Xnew.shape[0])/self.beta # TODO: This beta doesn't belong here in the EP case.
|
||||||
|
else:
|
||||||
|
Kxx = self.kern.Kdiag(Xnew)
|
||||||
|
var = Kxx - np.sum(Kx*np.dot(self.Kmmi - self.LBL_inv, Kx),0) + 1./self.beta # TODO: This beta doesn't belong here in the EP case.
|
||||||
|
|
||||||
return mu,var
|
return mu,var
|
||||||
|
|
||||||
def plot(self, *args, **kwargs):
|
def plot(self, *args, **kwargs):
|
||||||
|
|
|
||||||
|
|
@ -32,37 +32,41 @@ class uncollapsed_sparse_GP(sparse_GP_regression):
|
||||||
:type normalize_(X|Y): bool
|
:type normalize_(X|Y): bool
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, X, Y, q_u=None, *args, **kwargs)
|
def __init__(self, X, Y, q_u=None, M=10, *args, **kwargs):
|
||||||
D = Y.shape[1]
|
self.D = Y.shape[1]
|
||||||
if q_u is None:
|
if q_u is None:
|
||||||
if Z is None:
|
if 'Z' in kwargs.keys():
|
||||||
M = Z.shape[0]
|
print kwargs['Z']
|
||||||
|
self.M = kwargs['Z'].shape[0]
|
||||||
|
print self.M
|
||||||
else:
|
else:
|
||||||
M=M
|
self.M = M
|
||||||
q_u = np.hstack((np.ones(M*D)),np.eye(M).flatten())
|
q_u = np.hstack((np.random.randn(self.M*self.D),-0.5*np.eye(self.M).flatten()))
|
||||||
self.set_vb_param(q_u)
|
self.set_vb_param(q_u)
|
||||||
sparse_GP_regression.__init__(self, X, Y, *args, **kwargs)
|
sparse_GP_regression.__init__(self, X, Y, M=self.M,*args, **kwargs)
|
||||||
|
|
||||||
def _computations(self):
|
def _computations(self):
|
||||||
self.V = self.beta*self.Y
|
self.V = self.beta*self.Y
|
||||||
|
self.VmT = np.dot(self.V,self.q_u_expectation[0].T)
|
||||||
self.psi1V = np.dot(self.psi1, self.V)
|
self.psi1V = np.dot(self.psi1, self.V)
|
||||||
self.psi1VVpsi1 = np.dot(self.psi1V, self.psi1V.T)
|
self.psi1VVpsi1 = np.dot(self.psi1V, self.psi1V.T)
|
||||||
self.Lm = jitchol(self.Kmm)
|
self.Kmmi, self.Lm, self.Lmi, self.Kmm_logdet = pdinv(self.Kmm)
|
||||||
self.Lmi = chol_inv(self.Lm)
|
self.A = mdot(self.Lmi, self.beta*self.psi2, self.Lmi.T)
|
||||||
self.Kmmi = np.dot(self.Lmi.T, self.Lmi)
|
self.B = np.eye(self.M) + self.A
|
||||||
self.A = mdot(self.Lmi, self.psi2, self.Lmi.T)
|
self.Lambda = mdot(self.Lmi.T,self.B,self.Lmi)
|
||||||
self.B = np.eye(self.M) + self.beta * self.A
|
self.trace_K = self.psi0 - np.trace(self.A)/self.beta
|
||||||
self.Lambda = mdot(self.Lmi.T,self.B,sel.Lmi)
|
self.projected_mean = mdot(self.psi1.T,self.Kmmi,self.q_u_expectation[0])
|
||||||
|
|
||||||
# Compute dL_dpsi
|
# Compute dL_dpsi
|
||||||
self.dL_dpsi0 = - 0.5 * self.D * self.beta * np.ones(self.N)
|
self.dL_dpsi0 = - 0.5 * self.D * self.beta * np.ones(self.N)
|
||||||
self.dL_dpsi1 =
|
self.dL_dpsi1 = np.dot(self.VmT,self.Kmmi).T # This is the correct term for E I think...
|
||||||
self.dL_dpsi2 =
|
self.dL_dpsi2 = 0.5 * self.beta * self.D * (self.Kmmi - mdot(self.Kmmi,self.q_u_expectation[1],self.Kmmi))
|
||||||
|
|
||||||
# Compute dL_dKmm
|
# Compute dL_dKmm
|
||||||
self.dL_dKmm =
|
tmp = self.beta*mdot(self.psi2,self.Kmmi,self.q_u_expectation[1]) -np.dot(self.q_u_expectation[0],self.psi1V.T)
|
||||||
self.dL_dKmm +=
|
tmp += tmp.T
|
||||||
self.dL_dKmm +=
|
tmp += self.D*(-self.beta*self.psi2 - self.Kmm + self.q_u_expectation[1])
|
||||||
|
self.dL_dKmm = 0.5*mdot(self.Kmmi,tmp,self.Kmmi)
|
||||||
|
|
||||||
def log_likelihood(self):
|
def log_likelihood(self):
|
||||||
"""
|
"""
|
||||||
|
|
@ -70,10 +74,10 @@ class uncollapsed_sparse_GP(sparse_GP_regression):
|
||||||
"""
|
"""
|
||||||
A = -0.5*self.N*self.D*(np.log(2.*np.pi) - np.log(self.beta))
|
A = -0.5*self.N*self.D*(np.log(2.*np.pi) - np.log(self.beta))
|
||||||
B = -0.5*self.beta*self.D*self.trace_K
|
B = -0.5*self.beta*self.D*self.trace_K
|
||||||
C = -self.D *(self.Kmm_hld +0.5*np.sum(self.Lambda * self.mmT_S) + self.M/2.)
|
C = -0.5*self.D *(self.Kmm_logdet-self.q_u_logdet + np.sum(self.Lambda * self.q_u_expectation[1]) - self.M)
|
||||||
E = -0.5*self.beta*self.trYYT
|
D = -0.5*self.beta*self.trYYT
|
||||||
F = np.sum(np.dot(self.V.T,self.projected_mean))
|
E = np.sum(np.dot(self.V.T,self.projected_mean))
|
||||||
return A+B+C+D+E+F
|
return A+B+C+D+E
|
||||||
|
|
||||||
def dL_dbeta(self):
|
def dL_dbeta(self):
|
||||||
"""
|
"""
|
||||||
|
|
@ -82,24 +86,33 @@ class uncollapsed_sparse_GP(sparse_GP_regression):
|
||||||
"""
|
"""
|
||||||
dA_dbeta = 0.5 * self.N*self.D/self.beta
|
dA_dbeta = 0.5 * self.N*self.D/self.beta
|
||||||
dB_dbeta = - 0.5 * self.D * self.trace_K
|
dB_dbeta = - 0.5 * self.D * self.trace_K
|
||||||
dC_dbeta = - 0.5 * self.D * #TODO
|
dC_dbeta = - 0.5 * self.D * np.sum(self.q_u_expectation[1]*mdot(self.Kmmi,self.psi2,self.Kmmi))
|
||||||
dD_dbeta = - 0.5 * self.trYYT
|
dD_dbeta = - 0.5 * self.trYYT
|
||||||
|
dE_dbeta = np.sum(np.dot(self.Y.T,self.projected_mean))
|
||||||
|
|
||||||
return np.squeeze(dA_dbeta + dB_dbeta + dC_dbeta + dD_dbeta + dE_dbeta)
|
return np.squeeze(dA_dbeta + dB_dbeta + dC_dbeta + dD_dbeta + dE_dbeta)
|
||||||
|
|
||||||
def _raw_predict(self, Xnew, slices):
|
def _raw_predict(self, Xnew, slices,full_cov=False):
|
||||||
"""Internal helper function for making predictions, does not account for normalisation"""
|
"""Internal helper function for making predictions, does not account for normalisation"""
|
||||||
|
Kx = self.kern.K(Xnew,self.Z)
|
||||||
|
mu = mdot(Kx,self.Kmmi,self.q_u_expectation[0])
|
||||||
|
|
||||||
#TODO
|
tmp = self.Kmmi- mdot(self.Kmmi,self.q_u_cov,self.Kmmi)
|
||||||
|
if full_cov:
|
||||||
|
Kxx = self.kern.K(Xnew)
|
||||||
|
var = Kxx - mdot(Kx,tmp,Kx.T) + np.eye(Xnew.shape[0])/self.beta
|
||||||
|
else:
|
||||||
|
Kxx = self.kern.Kdiag(Xnew)
|
||||||
|
var = Kxx - np.sum(Kx*np.dot(Kx,tmp),1) + 1./self.beta
|
||||||
return mu,var
|
return mu,var
|
||||||
|
|
||||||
|
|
||||||
def set_vb_param(self,vb_param):
|
def set_vb_param(self,vb_param):
|
||||||
"""set the distribution q(u) from the canonical parameters"""
|
"""set the distribution q(u) from the canonical parameters"""
|
||||||
self.q_u_prec = -2.*vb_param[self.M*self.D:].reshape(self.M,self.M)
|
self.q_u_prec = -2.*vb_param[self.M*self.D:].reshape(self.M,self.M)
|
||||||
self.q_u_prec_L = jitchol(self.q_u_prec)
|
self.q_u_cov, q_u_Li, q_u_L, tmp = pdinv(self.q_u_prec)
|
||||||
self.q_u_cov_L = chol_inv(self.q_u_prec_L)
|
self.q_u_logdet = -tmp
|
||||||
self.q_u_cov = np.dot(self.q_u_cov_L,self.q_u_cov_L.T)
|
self.q_u_mean = np.dot(self.q_u_cov,vb_param[:self.M*self.D].reshape(self.M,self.D))
|
||||||
self.q_u_mean = -2.*np.dot(self.q_u_cov,vb_param[:self.M*self.D].reshape(self.M,self.D))
|
|
||||||
|
|
||||||
self.q_u_expectation = (self.q_u_mean, np.dot(self.q_u_mean,self.q_u_mean.T)+self.q_u_cov)
|
self.q_u_expectation = (self.q_u_mean, np.dot(self.q_u_mean,self.q_u_mean.T)+self.q_u_cov)
|
||||||
|
|
||||||
|
|
@ -119,11 +132,21 @@ class uncollapsed_sparse_GP(sparse_GP_regression):
|
||||||
|
|
||||||
Note that the natural gradient in either is given by the gradient in the other (See Hensman et al 2012 Fast Variational inference in the conjugate exponential Family)
|
Note that the natural gradient in either is given by the gradient in the other (See Hensman et al 2012 Fast Variational inference in the conjugate exponential Family)
|
||||||
"""
|
"""
|
||||||
foobar #TODO
|
dL_dmmT_S = -0.5*self.Lambda-self.q_u_canonical[1]
|
||||||
|
#dL_dm = np.dot(self.Kmmi,self.psi1V) - np.dot(self.Lambda,self.q_u_mean)
|
||||||
|
dL_dm = np.dot(self.Kmmi,self.psi1V) - self.q_u_canonical[0]
|
||||||
|
|
||||||
|
#dL_dSim =
|
||||||
|
#dL_dmhSi =
|
||||||
|
|
||||||
|
return np.hstack((dL_dm.flatten(),dL_dmmT_S.flatten())) # natgrad only, grad TODO
|
||||||
|
|
||||||
|
|
||||||
def plot(self, *args, **kwargs):
|
def plot(self, *args, **kwargs):
|
||||||
"""
|
"""
|
||||||
add the distribution q(u) to the plot from sparse_GP_regression
|
add the distribution q(u) to the plot from sparse_GP_regression
|
||||||
"""
|
"""
|
||||||
sparse_GP_regression.plot(self,*args,**kwargs)
|
sparse_GP_regression.plot(self,*args,**kwargs)
|
||||||
#TODO: plot the q(u) dist.
|
if self.Q==1:
|
||||||
|
pb.errorbar(self.Z[:,0],self.q_u_expectation[0][:,0],yerr=2.*np.sqrt(np.diag(self.q_u_cov)),fmt=None,ecolor='b')
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -22,7 +22,7 @@ class warpedGP(GP_regression):
|
||||||
if warping_function == None:
|
if warping_function == None:
|
||||||
self.warping_function = TanhWarpingFunction(warping_terms)
|
self.warping_function = TanhWarpingFunction(warping_terms)
|
||||||
# self.warping_params = np.random.randn(self.warping_function.n_terms, 3)
|
# self.warping_params = np.random.randn(self.warping_function.n_terms, 3)
|
||||||
self.warping_params = np.ones((self.warping_function.n_terms, 3))*1.0 # TODO better init
|
self.warping_params = np.ones((self.warping_function.n_terms, 3))*0.0 # TODO better init
|
||||||
self.warp_params_shape = (self.warping_function.n_terms, 3) # todo get this from the subclass
|
self.warp_params_shape = (self.warping_function.n_terms, 3) # todo get this from the subclass
|
||||||
|
|
||||||
self.Z = Y.copy()
|
self.Z = Y.copy()
|
||||||
|
|
|
||||||
|
|
@ -140,6 +140,7 @@ class GradientTests(unittest.TestCase):
|
||||||
self.assertTrue(m.checkgrad())
|
self.assertTrue(m.checkgrad())
|
||||||
|
|
||||||
def test_GP_EP(self):
|
def test_GP_EP(self):
|
||||||
|
return # Disabled TODO
|
||||||
N = 20
|
N = 20
|
||||||
X = np.hstack([np.random.rand(N/2)+1,np.random.rand(N/2)-1])[:,None]
|
X = np.hstack([np.random.rand(N/2)+1,np.random.rand(N/2)-1])[:,None]
|
||||||
k = GPy.kern.rbf(1) + GPy.kern.white(1)
|
k = GPy.kern.rbf(1) + GPy.kern.white(1)
|
||||||
|
|
|
||||||
|
|
@ -1,5 +1,4 @@
|
||||||
import os
|
import os
|
||||||
import posix
|
|
||||||
import pylab as pb
|
import pylab as pb
|
||||||
import numpy as np
|
import numpy as np
|
||||||
import GPy
|
import GPy
|
||||||
|
|
|
||||||
153
doc/Makefile
Normal file
153
doc/Makefile
Normal file
|
|
@ -0,0 +1,153 @@
|
||||||
|
# Makefile for Sphinx documentation
|
||||||
|
#
|
||||||
|
|
||||||
|
# You can set these variables from the command line.
|
||||||
|
SPHINXOPTS =
|
||||||
|
SPHINXBUILD = sphinx-build
|
||||||
|
PAPER =
|
||||||
|
BUILDDIR = _build
|
||||||
|
|
||||||
|
# Internal variables.
|
||||||
|
PAPEROPT_a4 = -D latex_paper_size=a4
|
||||||
|
PAPEROPT_letter = -D latex_paper_size=letter
|
||||||
|
ALLSPHINXOPTS = -d $(BUILDDIR)/doctrees $(PAPEROPT_$(PAPER)) $(SPHINXOPTS) .
|
||||||
|
# the i18n builder cannot share the environment and doctrees with the others
|
||||||
|
I18NSPHINXOPTS = $(PAPEROPT_$(PAPER)) $(SPHINXOPTS) .
|
||||||
|
|
||||||
|
.PHONY: help clean html dirhtml singlehtml pickle json htmlhelp qthelp devhelp epub latex latexpdf text man changes linkcheck doctest gettext
|
||||||
|
|
||||||
|
help:
|
||||||
|
@echo "Please use \`make <target>' where <target> is one of"
|
||||||
|
@echo " html to make standalone HTML files"
|
||||||
|
@echo " dirhtml to make HTML files named index.html in directories"
|
||||||
|
@echo " singlehtml to make a single large HTML file"
|
||||||
|
@echo " pickle to make pickle files"
|
||||||
|
@echo " json to make JSON files"
|
||||||
|
@echo " htmlhelp to make HTML files and a HTML help project"
|
||||||
|
@echo " qthelp to make HTML files and a qthelp project"
|
||||||
|
@echo " devhelp to make HTML files and a Devhelp project"
|
||||||
|
@echo " epub to make an epub"
|
||||||
|
@echo " latex to make LaTeX files, you can set PAPER=a4 or PAPER=letter"
|
||||||
|
@echo " latexpdf to make LaTeX files and run them through pdflatex"
|
||||||
|
@echo " text to make text files"
|
||||||
|
@echo " man to make manual pages"
|
||||||
|
@echo " texinfo to make Texinfo files"
|
||||||
|
@echo " info to make Texinfo files and run them through makeinfo"
|
||||||
|
@echo " gettext to make PO message catalogs"
|
||||||
|
@echo " changes to make an overview of all changed/added/deprecated items"
|
||||||
|
@echo " linkcheck to check all external links for integrity"
|
||||||
|
@echo " doctest to run all doctests embedded in the documentation (if enabled)"
|
||||||
|
|
||||||
|
clean:
|
||||||
|
-rm -rf $(BUILDDIR)/*
|
||||||
|
|
||||||
|
html:
|
||||||
|
$(SPHINXBUILD) -b html $(ALLSPHINXOPTS) $(BUILDDIR)/html
|
||||||
|
@echo
|
||||||
|
@echo "Build finished. The HTML pages are in $(BUILDDIR)/html."
|
||||||
|
|
||||||
|
dirhtml:
|
||||||
|
$(SPHINXBUILD) -b dirhtml $(ALLSPHINXOPTS) $(BUILDDIR)/dirhtml
|
||||||
|
@echo
|
||||||
|
@echo "Build finished. The HTML pages are in $(BUILDDIR)/dirhtml."
|
||||||
|
|
||||||
|
singlehtml:
|
||||||
|
$(SPHINXBUILD) -b singlehtml $(ALLSPHINXOPTS) $(BUILDDIR)/singlehtml
|
||||||
|
@echo
|
||||||
|
@echo "Build finished. The HTML page is in $(BUILDDIR)/singlehtml."
|
||||||
|
|
||||||
|
pickle:
|
||||||
|
$(SPHINXBUILD) -b pickle $(ALLSPHINXOPTS) $(BUILDDIR)/pickle
|
||||||
|
@echo
|
||||||
|
@echo "Build finished; now you can process the pickle files."
|
||||||
|
|
||||||
|
json:
|
||||||
|
$(SPHINXBUILD) -b json $(ALLSPHINXOPTS) $(BUILDDIR)/json
|
||||||
|
@echo
|
||||||
|
@echo "Build finished; now you can process the JSON files."
|
||||||
|
|
||||||
|
htmlhelp:
|
||||||
|
$(SPHINXBUILD) -b htmlhelp $(ALLSPHINXOPTS) $(BUILDDIR)/htmlhelp
|
||||||
|
@echo
|
||||||
|
@echo "Build finished; now you can run HTML Help Workshop with the" \
|
||||||
|
".hhp project file in $(BUILDDIR)/htmlhelp."
|
||||||
|
|
||||||
|
qthelp:
|
||||||
|
$(SPHINXBUILD) -b qthelp $(ALLSPHINXOPTS) $(BUILDDIR)/qthelp
|
||||||
|
@echo
|
||||||
|
@echo "Build finished; now you can run "qcollectiongenerator" with the" \
|
||||||
|
".qhcp project file in $(BUILDDIR)/qthelp, like this:"
|
||||||
|
@echo "# qcollectiongenerator $(BUILDDIR)/qthelp/GPy.qhcp"
|
||||||
|
@echo "To view the help file:"
|
||||||
|
@echo "# assistant -collectionFile $(BUILDDIR)/qthelp/GPy.qhc"
|
||||||
|
|
||||||
|
devhelp:
|
||||||
|
$(SPHINXBUILD) -b devhelp $(ALLSPHINXOPTS) $(BUILDDIR)/devhelp
|
||||||
|
@echo
|
||||||
|
@echo "Build finished."
|
||||||
|
@echo "To view the help file:"
|
||||||
|
@echo "# mkdir -p $$HOME/.local/share/devhelp/GPy"
|
||||||
|
@echo "# ln -s $(BUILDDIR)/devhelp $$HOME/.local/share/devhelp/GPy"
|
||||||
|
@echo "# devhelp"
|
||||||
|
|
||||||
|
epub:
|
||||||
|
$(SPHINXBUILD) -b epub $(ALLSPHINXOPTS) $(BUILDDIR)/epub
|
||||||
|
@echo
|
||||||
|
@echo "Build finished. The epub file is in $(BUILDDIR)/epub."
|
||||||
|
|
||||||
|
latex:
|
||||||
|
$(SPHINXBUILD) -b latex $(ALLSPHINXOPTS) $(BUILDDIR)/latex
|
||||||
|
@echo
|
||||||
|
@echo "Build finished; the LaTeX files are in $(BUILDDIR)/latex."
|
||||||
|
@echo "Run \`make' in that directory to run these through (pdf)latex" \
|
||||||
|
"(use \`make latexpdf' here to do that automatically)."
|
||||||
|
|
||||||
|
latexpdf:
|
||||||
|
$(SPHINXBUILD) -b latex $(ALLSPHINXOPTS) $(BUILDDIR)/latex
|
||||||
|
@echo "Running LaTeX files through pdflatex..."
|
||||||
|
$(MAKE) -C $(BUILDDIR)/latex all-pdf
|
||||||
|
@echo "pdflatex finished; the PDF files are in $(BUILDDIR)/latex."
|
||||||
|
|
||||||
|
text:
|
||||||
|
$(SPHINXBUILD) -b text $(ALLSPHINXOPTS) $(BUILDDIR)/text
|
||||||
|
@echo
|
||||||
|
@echo "Build finished. The text files are in $(BUILDDIR)/text."
|
||||||
|
|
||||||
|
man:
|
||||||
|
$(SPHINXBUILD) -b man $(ALLSPHINXOPTS) $(BUILDDIR)/man
|
||||||
|
@echo
|
||||||
|
@echo "Build finished. The manual pages are in $(BUILDDIR)/man."
|
||||||
|
|
||||||
|
texinfo:
|
||||||
|
$(SPHINXBUILD) -b texinfo $(ALLSPHINXOPTS) $(BUILDDIR)/texinfo
|
||||||
|
@echo
|
||||||
|
@echo "Build finished. The Texinfo files are in $(BUILDDIR)/texinfo."
|
||||||
|
@echo "Run \`make' in that directory to run these through makeinfo" \
|
||||||
|
"(use \`make info' here to do that automatically)."
|
||||||
|
|
||||||
|
info:
|
||||||
|
$(SPHINXBUILD) -b texinfo $(ALLSPHINXOPTS) $(BUILDDIR)/texinfo
|
||||||
|
@echo "Running Texinfo files through makeinfo..."
|
||||||
|
make -C $(BUILDDIR)/texinfo info
|
||||||
|
@echo "makeinfo finished; the Info files are in $(BUILDDIR)/texinfo."
|
||||||
|
|
||||||
|
gettext:
|
||||||
|
$(SPHINXBUILD) -b gettext $(I18NSPHINXOPTS) $(BUILDDIR)/locale
|
||||||
|
@echo
|
||||||
|
@echo "Build finished. The message catalogs are in $(BUILDDIR)/locale."
|
||||||
|
|
||||||
|
changes:
|
||||||
|
$(SPHINXBUILD) -b changes $(ALLSPHINXOPTS) $(BUILDDIR)/changes
|
||||||
|
@echo
|
||||||
|
@echo "The overview file is in $(BUILDDIR)/changes."
|
||||||
|
|
||||||
|
linkcheck:
|
||||||
|
$(SPHINXBUILD) -b linkcheck $(ALLSPHINXOPTS) $(BUILDDIR)/linkcheck
|
||||||
|
@echo
|
||||||
|
@echo "Link check complete; look for any errors in the above output " \
|
||||||
|
"or in $(BUILDDIR)/linkcheck/output.txt."
|
||||||
|
|
||||||
|
doctest:
|
||||||
|
$(SPHINXBUILD) -b doctest $(ALLSPHINXOPTS) $(BUILDDIR)/doctest
|
||||||
|
@echo "Testing of doctests in the sources finished, look at the " \
|
||||||
|
"results in $(BUILDDIR)/doctest/output.txt."
|
||||||
242
doc/conf.py
Normal file
242
doc/conf.py
Normal file
|
|
@ -0,0 +1,242 @@
|
||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
#
|
||||||
|
# GPy documentation build configuration file, created by
|
||||||
|
# sphinx-quickstart on Wed Jan 9 15:21:20 2013.
|
||||||
|
#
|
||||||
|
# This file is execfile()d with the current directory set to its containing dir.
|
||||||
|
#
|
||||||
|
# Note that not all possible configuration values are present in this
|
||||||
|
# autogenerated file.
|
||||||
|
#
|
||||||
|
# All configuration values have a default; values that are commented out
|
||||||
|
# serve to show the default.
|
||||||
|
|
||||||
|
import sys, os
|
||||||
|
|
||||||
|
# If extensions (or modules to document with autodoc) are in another directory,
|
||||||
|
# add these directories to sys.path here. If the directory is relative to the
|
||||||
|
# documentation root, use os.path.abspath to make it absolute, like shown here.
|
||||||
|
#sys.path.insert(0, os.path.abspath('.'))
|
||||||
|
|
||||||
|
# -- General configuration -----------------------------------------------------
|
||||||
|
|
||||||
|
# If your documentation needs a minimal Sphinx version, state it here.
|
||||||
|
#needs_sphinx = '1.0'
|
||||||
|
|
||||||
|
# Add any Sphinx extension module names here, as strings. They can be extensions
|
||||||
|
# coming with Sphinx (named 'sphinx.ext.*') or your custom ones.
|
||||||
|
extensions = ['sphinx.ext.autodoc', 'sphinx.ext.todo', 'sphinx.ext.pngmath', 'sphinx.ext.mathjax', 'sphinx.ext.viewcode']
|
||||||
|
|
||||||
|
# Add any paths that contain templates here, relative to this directory.
|
||||||
|
templates_path = ['_templates']
|
||||||
|
|
||||||
|
# The suffix of source filenames.
|
||||||
|
source_suffix = '.rst'
|
||||||
|
|
||||||
|
# The encoding of source files.
|
||||||
|
#source_encoding = 'utf-8-sig'
|
||||||
|
|
||||||
|
# The master toctree document.
|
||||||
|
master_doc = 'index'
|
||||||
|
|
||||||
|
# General information about the project.
|
||||||
|
project = u'GPy'
|
||||||
|
copyright = u'2013, The GPy authors'
|
||||||
|
|
||||||
|
# The version info for the project you're documenting, acts as replacement for
|
||||||
|
# |version| and |release|, also used in various other places throughout the
|
||||||
|
# built documents.
|
||||||
|
#
|
||||||
|
# The short X.Y version.
|
||||||
|
version = '0.00001'
|
||||||
|
# The full version, including alpha/beta/rc tags.
|
||||||
|
release = '0.00001'
|
||||||
|
|
||||||
|
# The language for content autogenerated by Sphinx. Refer to documentation
|
||||||
|
# for a list of supported languages.
|
||||||
|
#language = None
|
||||||
|
|
||||||
|
# There are two options for replacing |today|: either, you set today to some
|
||||||
|
# non-false value, then it is used:
|
||||||
|
#today = ''
|
||||||
|
# Else, today_fmt is used as the format for a strftime call.
|
||||||
|
#today_fmt = '%B %d, %Y'
|
||||||
|
|
||||||
|
# List of patterns, relative to source directory, that match files and
|
||||||
|
# directories to ignore when looking for source files.
|
||||||
|
exclude_patterns = ['_build']
|
||||||
|
|
||||||
|
# The reST default role (used for this markup: `text`) to use for all documents.
|
||||||
|
#default_role = None
|
||||||
|
|
||||||
|
# If true, '()' will be appended to :func: etc. cross-reference text.
|
||||||
|
#add_function_parentheses = True
|
||||||
|
|
||||||
|
# If true, the current module name will be prepended to all description
|
||||||
|
# unit titles (such as .. function::).
|
||||||
|
#add_module_names = True
|
||||||
|
|
||||||
|
# If true, sectionauthor and moduleauthor directives will be shown in the
|
||||||
|
# output. They are ignored by default.
|
||||||
|
#show_authors = False
|
||||||
|
|
||||||
|
# The name of the Pygments (syntax highlighting) style to use.
|
||||||
|
pygments_style = 'sphinx'
|
||||||
|
|
||||||
|
# A list of ignored prefixes for module index sorting.
|
||||||
|
#modindex_common_prefix = []
|
||||||
|
|
||||||
|
|
||||||
|
# -- Options for HTML output ---------------------------------------------------
|
||||||
|
|
||||||
|
# The theme to use for HTML and HTML Help pages. See the documentation for
|
||||||
|
# a list of builtin themes.
|
||||||
|
html_theme = 'default'
|
||||||
|
|
||||||
|
# Theme options are theme-specific and customize the look and feel of a theme
|
||||||
|
# further. For a list of options available for each theme, see the
|
||||||
|
# documentation.
|
||||||
|
#html_theme_options = {}
|
||||||
|
|
||||||
|
# Add any paths that contain custom themes here, relative to this directory.
|
||||||
|
#html_theme_path = []
|
||||||
|
|
||||||
|
# The name for this set of Sphinx documents. If None, it defaults to
|
||||||
|
# "<project> v<release> documentation".
|
||||||
|
#html_title = None
|
||||||
|
|
||||||
|
# A shorter title for the navigation bar. Default is the same as html_title.
|
||||||
|
#html_short_title = None
|
||||||
|
|
||||||
|
# The name of an image file (relative to this directory) to place at the top
|
||||||
|
# of the sidebar.
|
||||||
|
#html_logo = None
|
||||||
|
|
||||||
|
# The name of an image file (within the static path) to use as favicon of the
|
||||||
|
# docs. This file should be a Windows icon file (.ico) being 16x16 or 32x32
|
||||||
|
# pixels large.
|
||||||
|
#html_favicon = None
|
||||||
|
|
||||||
|
# Add any paths that contain custom static files (such as style sheets) here,
|
||||||
|
# relative to this directory. They are copied after the builtin static files,
|
||||||
|
# so a file named "default.css" will overwrite the builtin "default.css".
|
||||||
|
html_static_path = ['_static']
|
||||||
|
|
||||||
|
# If not '', a 'Last updated on:' timestamp is inserted at every page bottom,
|
||||||
|
# using the given strftime format.
|
||||||
|
#html_last_updated_fmt = '%b %d, %Y'
|
||||||
|
|
||||||
|
# If true, SmartyPants will be used to convert quotes and dashes to
|
||||||
|
# typographically correct entities.
|
||||||
|
#html_use_smartypants = True
|
||||||
|
|
||||||
|
# Custom sidebar templates, maps document names to template names.
|
||||||
|
#html_sidebars = {}
|
||||||
|
|
||||||
|
# Additional templates that should be rendered to pages, maps page names to
|
||||||
|
# template names.
|
||||||
|
#html_additional_pages = {}
|
||||||
|
|
||||||
|
# If false, no module index is generated.
|
||||||
|
#html_domain_indices = True
|
||||||
|
|
||||||
|
# If false, no index is generated.
|
||||||
|
#html_use_index = True
|
||||||
|
|
||||||
|
# If true, the index is split into individual pages for each letter.
|
||||||
|
#html_split_index = False
|
||||||
|
|
||||||
|
# If true, links to the reST sources are added to the pages.
|
||||||
|
#html_show_sourcelink = True
|
||||||
|
|
||||||
|
# If true, "Created using Sphinx" is shown in the HTML footer. Default is True.
|
||||||
|
#html_show_sphinx = True
|
||||||
|
|
||||||
|
# If true, "(C) Copyright ..." is shown in the HTML footer. Default is True.
|
||||||
|
#html_show_copyright = True
|
||||||
|
|
||||||
|
# If true, an OpenSearch description file will be output, and all pages will
|
||||||
|
# contain a <link> tag referring to it. The value of this option must be the
|
||||||
|
# base URL from which the finished HTML is served.
|
||||||
|
#html_use_opensearch = ''
|
||||||
|
|
||||||
|
# This is the file name suffix for HTML files (e.g. ".xhtml").
|
||||||
|
#html_file_suffix = None
|
||||||
|
|
||||||
|
# Output file base name for HTML help builder.
|
||||||
|
htmlhelp_basename = 'GPydoc'
|
||||||
|
|
||||||
|
|
||||||
|
# -- Options for LaTeX output --------------------------------------------------
|
||||||
|
|
||||||
|
latex_elements = {
|
||||||
|
# The paper size ('letterpaper' or 'a4paper').
|
||||||
|
#'papersize': 'letterpaper',
|
||||||
|
|
||||||
|
# The font size ('10pt', '11pt' or '12pt').
|
||||||
|
#'pointsize': '10pt',
|
||||||
|
|
||||||
|
# Additional stuff for the LaTeX preamble.
|
||||||
|
#'preamble': '',
|
||||||
|
}
|
||||||
|
|
||||||
|
# Grouping the document tree into LaTeX files. List of tuples
|
||||||
|
# (source start file, target name, title, author, documentclass [howto/manual]).
|
||||||
|
latex_documents = [
|
||||||
|
('index', 'GPy.tex', u'GPy Documentation',
|
||||||
|
u'The GPy authors', 'manual'),
|
||||||
|
]
|
||||||
|
|
||||||
|
# The name of an image file (relative to this directory) to place at the top of
|
||||||
|
# the title page.
|
||||||
|
#latex_logo = None
|
||||||
|
|
||||||
|
# For "manual" documents, if this is true, then toplevel headings are parts,
|
||||||
|
# not chapters.
|
||||||
|
#latex_use_parts = False
|
||||||
|
|
||||||
|
# If true, show page references after internal links.
|
||||||
|
#latex_show_pagerefs = False
|
||||||
|
|
||||||
|
# If true, show URL addresses after external links.
|
||||||
|
#latex_show_urls = False
|
||||||
|
|
||||||
|
# Documents to append as an appendix to all manuals.
|
||||||
|
#latex_appendices = []
|
||||||
|
|
||||||
|
# If false, no module index is generated.
|
||||||
|
#latex_domain_indices = True
|
||||||
|
|
||||||
|
|
||||||
|
# -- Options for manual page output --------------------------------------------
|
||||||
|
|
||||||
|
# One entry per manual page. List of tuples
|
||||||
|
# (source start file, name, description, authors, manual section).
|
||||||
|
man_pages = [
|
||||||
|
('index', 'gpy', u'GPy Documentation',
|
||||||
|
[u'The GPy authors'], 1)
|
||||||
|
]
|
||||||
|
|
||||||
|
# If true, show URL addresses after external links.
|
||||||
|
#man_show_urls = False
|
||||||
|
|
||||||
|
|
||||||
|
# -- Options for Texinfo output ------------------------------------------------
|
||||||
|
|
||||||
|
# Grouping the document tree into Texinfo files. List of tuples
|
||||||
|
# (source start file, target name, title, author,
|
||||||
|
# dir menu entry, description, category)
|
||||||
|
texinfo_documents = [
|
||||||
|
('index', 'GPy', u'GPy Documentation',
|
||||||
|
u'The GPy authors', 'GPy', 'One line description of project.',
|
||||||
|
'Miscellaneous'),
|
||||||
|
]
|
||||||
|
|
||||||
|
# Documents to append as an appendix to all manuals.
|
||||||
|
#texinfo_appendices = []
|
||||||
|
|
||||||
|
# If false, no module index is generated.
|
||||||
|
#texinfo_domain_indices = True
|
||||||
|
|
||||||
|
# How to display URL addresses: 'footnote', 'no', or 'inline'.
|
||||||
|
#texinfo_show_urls = 'footnote'
|
||||||
22
doc/index.rst
Normal file
22
doc/index.rst
Normal file
|
|
@ -0,0 +1,22 @@
|
||||||
|
.. GPy documentation master file, created by
|
||||||
|
sphinx-quickstart on Wed Jan 9 15:21:20 2013.
|
||||||
|
You can adapt this file completely to your liking, but it should at least
|
||||||
|
contain the root `toctree` directive.
|
||||||
|
|
||||||
|
Welcome to GPy's documentation!
|
||||||
|
===============================
|
||||||
|
|
||||||
|
Contents:
|
||||||
|
|
||||||
|
.. toctree::
|
||||||
|
:maxdepth: 2
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
Indices and tables
|
||||||
|
==================
|
||||||
|
|
||||||
|
* :ref:`genindex`
|
||||||
|
* :ref:`modindex`
|
||||||
|
* :ref:`search`
|
||||||
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue