mirror of
https://github.com/SheffieldML/GPy.git
synced 2026-04-27 13:56:23 +02:00
Merge branch 'devel' of github.com:SheffieldML/GPy into devel
This commit is contained in:
commit
369cc0ba2b
25 changed files with 737 additions and 277 deletions
|
|
@ -12,6 +12,10 @@ from .. import likelihoods
|
||||||
from ..likelihoods.gaussian import Gaussian
|
from ..likelihoods.gaussian import Gaussian
|
||||||
from ..inference.latent_function_inference import exact_gaussian_inference, expectation_propagation, LatentFunctionInference
|
from ..inference.latent_function_inference import exact_gaussian_inference, expectation_propagation, LatentFunctionInference
|
||||||
from parameterization.variational import VariationalPosterior
|
from parameterization.variational import VariationalPosterior
|
||||||
|
from scipy.sparse.base import issparse
|
||||||
|
|
||||||
|
import logging
|
||||||
|
logger = logging.getLogger("GP")
|
||||||
|
|
||||||
class GP(Model):
|
class GP(Model):
|
||||||
"""
|
"""
|
||||||
|
|
@ -34,12 +38,14 @@ class GP(Model):
|
||||||
assert X.ndim == 2
|
assert X.ndim == 2
|
||||||
if isinstance(X, (ObsAr, VariationalPosterior)):
|
if isinstance(X, (ObsAr, VariationalPosterior)):
|
||||||
self.X = X.copy()
|
self.X = X.copy()
|
||||||
else: self.X = ObsAr(X.copy())
|
else: self.X = ObsAr(X)
|
||||||
|
|
||||||
self.num_data, self.input_dim = self.X.shape
|
self.num_data, self.input_dim = self.X.shape
|
||||||
|
|
||||||
assert Y.ndim == 2
|
assert Y.ndim == 2
|
||||||
self.Y = ObsAr(Y.copy())
|
logger.info("initializing Y")
|
||||||
|
if issparse(Y): self.Y = Y
|
||||||
|
else: self.Y = ObsAr(Y)
|
||||||
assert Y.shape[0] == self.num_data
|
assert Y.shape[0] == self.num_data
|
||||||
_, self.output_dim = self.Y.shape
|
_, self.output_dim = self.Y.shape
|
||||||
|
|
||||||
|
|
@ -54,6 +60,7 @@ class GP(Model):
|
||||||
self.likelihood = likelihood
|
self.likelihood = likelihood
|
||||||
|
|
||||||
#find a sensible inference method
|
#find a sensible inference method
|
||||||
|
logger.info("initializing inference method")
|
||||||
if inference_method is None:
|
if inference_method is None:
|
||||||
if isinstance(likelihood, likelihoods.Gaussian) or isinstance(likelihood, likelihoods.MixedNoise):
|
if isinstance(likelihood, likelihoods.Gaussian) or isinstance(likelihood, likelihoods.MixedNoise):
|
||||||
inference_method = exact_gaussian_inference.ExactGaussianInference()
|
inference_method = exact_gaussian_inference.ExactGaussianInference()
|
||||||
|
|
@ -62,6 +69,7 @@ class GP(Model):
|
||||||
print "defaulting to ", inference_method, "for latent function inference"
|
print "defaulting to ", inference_method, "for latent function inference"
|
||||||
self.inference_method = inference_method
|
self.inference_method = inference_method
|
||||||
|
|
||||||
|
logger.info("adding kernel and likelihood as parameters")
|
||||||
self.add_parameter(self.kern)
|
self.add_parameter(self.kern)
|
||||||
self.add_parameter(self.likelihood)
|
self.add_parameter(self.likelihood)
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -225,14 +225,18 @@ class Model(Parameterized):
|
||||||
if self.size == 0:
|
if self.size == 0:
|
||||||
raise RuntimeError, "Model without parameters cannot be optimized"
|
raise RuntimeError, "Model without parameters cannot be optimized"
|
||||||
|
|
||||||
if optimizer is None:
|
|
||||||
optimizer = self.preferred_optimizer
|
|
||||||
|
|
||||||
if start == None:
|
if start == None:
|
||||||
start = self.optimizer_array
|
start = self.optimizer_array
|
||||||
|
|
||||||
optimizer = optimization.get_optimizer(optimizer)
|
if optimizer is None:
|
||||||
opt = optimizer(start, model=self, **kwargs)
|
optimizer = self.preferred_optimizer
|
||||||
|
|
||||||
|
if isinstance(optimizer, optimization.Optimizer):
|
||||||
|
opt = optimizer
|
||||||
|
opt.model = self
|
||||||
|
else:
|
||||||
|
optimizer = optimization.get_optimizer(optimizer)
|
||||||
|
opt = optimizer(start, model=self, **kwargs)
|
||||||
|
|
||||||
opt.run(f_fp=self._objective_grads, f=self._objective, fp=self._grads)
|
opt.run(f_fp=self._objective_grads, f=self._objective, fp=self._grads)
|
||||||
|
|
||||||
|
|
@ -249,7 +253,7 @@ class Model(Parameterized):
|
||||||
def _checkgrad(self, target_param=None, verbose=False, step=1e-6, tolerance=1e-3):
|
def _checkgrad(self, target_param=None, verbose=False, step=1e-6, tolerance=1e-3):
|
||||||
"""
|
"""
|
||||||
Check the gradient of the ,odel by comparing to a numerical
|
Check the gradient of the ,odel by comparing to a numerical
|
||||||
estimate. If the verbose flag is passed, invividual
|
estimate. If the verbose flag is passed, individual
|
||||||
components are tested (and printed)
|
components are tested (and printed)
|
||||||
|
|
||||||
:param verbose: If True, print a "full" checking of each parameter
|
:param verbose: If True, print a "full" checking of each parameter
|
||||||
|
|
|
||||||
|
|
@ -751,8 +751,6 @@ class OptimizationHandlable(Indexable):
|
||||||
Transform the gradients by multiplying the gradient factor for each
|
Transform the gradients by multiplying the gradient factor for each
|
||||||
constraint to it.
|
constraint to it.
|
||||||
"""
|
"""
|
||||||
if self.has_parent():
|
|
||||||
return g
|
|
||||||
[np.put(g, i, g[i] * c.gradfactor(self.param_array[i])) for c, i in self.constraints.iteritems() if c != __fixed__]
|
[np.put(g, i, g[i] * c.gradfactor(self.param_array[i])) for c, i in self.constraints.iteritems() if c != __fixed__]
|
||||||
if self._has_fixes(): return g[self._fixes_]
|
if self._has_fixes(): return g[self._fixes_]
|
||||||
return g
|
return g
|
||||||
|
|
@ -793,7 +791,7 @@ class OptimizationHandlable(Indexable):
|
||||||
#===========================================================================
|
#===========================================================================
|
||||||
# Randomizeable
|
# Randomizeable
|
||||||
#===========================================================================
|
#===========================================================================
|
||||||
def randomize(self, rand_gen=np.random.normal, loc=0, scale=1, *args, **kwargs):
|
def randomize(self, rand_gen=np.random.normal, *args, **kwargs):
|
||||||
"""
|
"""
|
||||||
Randomize the model.
|
Randomize the model.
|
||||||
Make this draw from the prior if one exists, else draw from given random generator
|
Make this draw from the prior if one exists, else draw from given random generator
|
||||||
|
|
@ -804,7 +802,7 @@ class OptimizationHandlable(Indexable):
|
||||||
:param args, kwargs: will be passed through to random number generator
|
:param args, kwargs: will be passed through to random number generator
|
||||||
"""
|
"""
|
||||||
# first take care of all parameters (from N(0,1))
|
# first take care of all parameters (from N(0,1))
|
||||||
x = rand_gen(loc=loc, scale=scale, size=self._size_transformed(), *args, **kwargs)
|
x = rand_gen(size=self._size_transformed(), *args, **kwargs)
|
||||||
# now draw from prior where possible
|
# now draw from prior where possible
|
||||||
[np.put(x, ind, p.rvs(ind.size)) for p, ind in self.priors.iteritems() if not p is None]
|
[np.put(x, ind, p.rvs(ind.size)) for p, ind in self.priors.iteritems() if not p is None]
|
||||||
self.optimizer_array = x # makes sure all of the tied parameters get the same init (since there's only one prior object...)
|
self.optimizer_array = x # makes sure all of the tied parameters get the same init (since there's only one prior object...)
|
||||||
|
|
@ -835,6 +833,11 @@ class OptimizationHandlable(Indexable):
|
||||||
1.) connect param_array of children to self.param_array
|
1.) connect param_array of children to self.param_array
|
||||||
2.) tell all children to propagate further
|
2.) tell all children to propagate further
|
||||||
"""
|
"""
|
||||||
|
if self.param_array.size != self.size:
|
||||||
|
self._param_array_ = np.empty(self.size, dtype=np.float64)
|
||||||
|
if self.gradient.size != self.size:
|
||||||
|
self._gradient_array_ = np.empty(self.size, dtype=np.float64)
|
||||||
|
|
||||||
pi_old_size = 0
|
pi_old_size = 0
|
||||||
for pi in self.parameters:
|
for pi in self.parameters:
|
||||||
pislice = slice(pi_old_size, pi_old_size + pi.size)
|
pislice = slice(pi_old_size, pi_old_size + pi.size)
|
||||||
|
|
@ -848,6 +851,9 @@ class OptimizationHandlable(Indexable):
|
||||||
pi._propagate_param_grad(parray[pislice], garray[pislice])
|
pi._propagate_param_grad(parray[pislice], garray[pislice])
|
||||||
pi_old_size += pi.size
|
pi_old_size += pi.size
|
||||||
|
|
||||||
|
def _connect_parameters(self):
|
||||||
|
pass
|
||||||
|
|
||||||
class Parameterizable(OptimizationHandlable):
|
class Parameterizable(OptimizationHandlable):
|
||||||
"""
|
"""
|
||||||
A parameterisable class.
|
A parameterisable class.
|
||||||
|
|
@ -874,6 +880,9 @@ class Parameterizable(OptimizationHandlable):
|
||||||
"""
|
"""
|
||||||
Array representing the parameters of this class.
|
Array representing the parameters of this class.
|
||||||
There is only one copy of all parameters in memory, two during optimization.
|
There is only one copy of all parameters in memory, two during optimization.
|
||||||
|
|
||||||
|
!WARNING!: setting the parameter array MUST always be done in memory:
|
||||||
|
m.param_array[:] = m_copy.param_array
|
||||||
"""
|
"""
|
||||||
if self.__dict__.get('_param_array_', None) is None:
|
if self.__dict__.get('_param_array_', None) is None:
|
||||||
self._param_array_ = np.empty(self.size, dtype=np.float64)
|
self._param_array_ = np.empty(self.size, dtype=np.float64)
|
||||||
|
|
@ -986,6 +995,11 @@ class Parameterizable(OptimizationHandlable):
|
||||||
# notification system
|
# notification system
|
||||||
#===========================================================================
|
#===========================================================================
|
||||||
def _parameters_changed_notification(self, me, which=None):
|
def _parameters_changed_notification(self, me, which=None):
|
||||||
|
"""
|
||||||
|
In parameterizable we just need to make sure, that the next call to optimizer_array
|
||||||
|
will update the optimizer_array to the latest parameters
|
||||||
|
"""
|
||||||
|
self._optimizer_copy_transformed = False # tells the optimizer array to update on next request
|
||||||
self.parameters_changed()
|
self.parameters_changed()
|
||||||
def _pass_through_notify_observers(self, me, which=None):
|
def _pass_through_notify_observers(self, me, which=None):
|
||||||
self.notify_observers(which=which)
|
self.notify_observers(which=which)
|
||||||
|
|
@ -1017,4 +1031,3 @@ class Parameterizable(OptimizationHandlable):
|
||||||
updates get passed through. See :py:function:``GPy.core.param.Observable.add_observer``
|
updates get passed through. See :py:function:``GPy.core.param.Observable.add_observer``
|
||||||
"""
|
"""
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -8,11 +8,23 @@ from re import compile, _pattern_type
|
||||||
from param import ParamConcatenation
|
from param import ParamConcatenation
|
||||||
from parameter_core import HierarchyError, Parameterizable, adjust_name_for_printing
|
from parameter_core import HierarchyError, Parameterizable, adjust_name_for_printing
|
||||||
|
|
||||||
|
import logging
|
||||||
|
logger = logging.getLogger("parameters changed meta")
|
||||||
|
|
||||||
class ParametersChangedMeta(type):
|
class ParametersChangedMeta(type):
|
||||||
def __call__(self, *args, **kw):
|
def __call__(self, *args, **kw):
|
||||||
instance = super(ParametersChangedMeta, self).__call__(*args, **kw)
|
self._in_init_ = True
|
||||||
instance.parameters_changed()
|
#import ipdb;ipdb.set_trace()
|
||||||
return instance
|
self = super(ParametersChangedMeta, self).__call__(*args, **kw)
|
||||||
|
logger.debug("finished init")
|
||||||
|
self._in_init_ = False
|
||||||
|
logger.debug("connecting parameters")
|
||||||
|
self._highest_parent_._connect_parameters()
|
||||||
|
self._highest_parent_._notify_parent_change()
|
||||||
|
self._highest_parent_._connect_fixes()
|
||||||
|
logger.debug("calling parameters changed")
|
||||||
|
self.parameters_changed()
|
||||||
|
return self
|
||||||
|
|
||||||
class Parameterized(Parameterizable):
|
class Parameterized(Parameterizable):
|
||||||
"""
|
"""
|
||||||
|
|
@ -64,14 +76,12 @@ class Parameterized(Parameterizable):
|
||||||
#===========================================================================
|
#===========================================================================
|
||||||
def __init__(self, name=None, parameters=[], *a, **kw):
|
def __init__(self, name=None, parameters=[], *a, **kw):
|
||||||
super(Parameterized, self).__init__(name=name, *a, **kw)
|
super(Parameterized, self).__init__(name=name, *a, **kw)
|
||||||
self._in_init_ = True
|
|
||||||
self.size = sum(p.size for p in self.parameters)
|
self.size = sum(p.size for p in self.parameters)
|
||||||
self.add_observer(self, self._parameters_changed_notification, -100)
|
self.add_observer(self, self._parameters_changed_notification, -100)
|
||||||
if not self._has_fixes():
|
if not self._has_fixes():
|
||||||
self._fixes_ = None
|
self._fixes_ = None
|
||||||
self._param_slices_ = []
|
self._param_slices_ = []
|
||||||
self._connect_parameters()
|
#self._connect_parameters()
|
||||||
del self._in_init_
|
|
||||||
self.add_parameters(*parameters)
|
self.add_parameters(*parameters)
|
||||||
|
|
||||||
def build_pydot(self, G=None):
|
def build_pydot(self, G=None):
|
||||||
|
|
@ -125,6 +135,9 @@ class Parameterized(Parameterizable):
|
||||||
param._parent_.remove_parameter(param)
|
param._parent_.remove_parameter(param)
|
||||||
# make sure the size is set
|
# make sure the size is set
|
||||||
if index is None:
|
if index is None:
|
||||||
|
start = sum(p.size for p in self.parameters)
|
||||||
|
self.constraints.shift_right(start, param.size)
|
||||||
|
self.priors.shift_right(start, param.size)
|
||||||
self.constraints.update(param.constraints, self.size)
|
self.constraints.update(param.constraints, self.size)
|
||||||
self.priors.update(param.priors, self.size)
|
self.priors.update(param.priors, self.size)
|
||||||
self.parameters.append(param)
|
self.parameters.append(param)
|
||||||
|
|
@ -143,14 +156,16 @@ class Parameterized(Parameterizable):
|
||||||
parent.size += param.size
|
parent.size += param.size
|
||||||
parent = parent._parent_
|
parent = parent._parent_
|
||||||
|
|
||||||
self._connect_parameters()
|
if not self._in_init_:
|
||||||
|
self._connect_parameters()
|
||||||
|
self._notify_parent_change()
|
||||||
|
|
||||||
self._highest_parent_._connect_parameters(ignore_added_names=_ignore_added_names)
|
self._highest_parent_._connect_parameters(ignore_added_names=_ignore_added_names)
|
||||||
self._highest_parent_._notify_parent_change()
|
self._highest_parent_._notify_parent_change()
|
||||||
self._highest_parent_._connect_fixes()
|
self._highest_parent_._connect_fixes()
|
||||||
|
|
||||||
else:
|
else:
|
||||||
raise HierarchyError, """Parameter exists already and no copy made"""
|
raise HierarchyError, """Parameter exists already, try making a copy"""
|
||||||
|
|
||||||
|
|
||||||
def add_parameters(self, *parameters):
|
def add_parameters(self, *parameters):
|
||||||
|
|
@ -198,26 +213,28 @@ class Parameterized(Parameterizable):
|
||||||
# no parameters for this class
|
# no parameters for this class
|
||||||
return
|
return
|
||||||
if self.param_array.size != self.size:
|
if self.param_array.size != self.size:
|
||||||
self.param_array = np.empty(self.size, dtype=np.float64)
|
self._param_array_ = np.empty(self.size, dtype=np.float64)
|
||||||
if self.gradient.size != self.size:
|
if self.gradient.size != self.size:
|
||||||
self._gradient_array_ = np.empty(self.size, dtype=np.float64)
|
self._gradient_array_ = np.empty(self.size, dtype=np.float64)
|
||||||
|
|
||||||
old_size = 0
|
old_size = 0
|
||||||
self._param_slices_ = []
|
self._param_slices_ = []
|
||||||
for i, p in enumerate(self.parameters):
|
for i, p in enumerate(self.parameters):
|
||||||
|
if not p.param_array.flags['C_CONTIGUOUS']:
|
||||||
|
raise ValueError, "This should not happen! Please write an email to the developers with the code, which reproduces this error. All parameter arrays must be C_CONTIGUOUS"
|
||||||
|
|
||||||
p._parent_ = self
|
p._parent_ = self
|
||||||
p._parent_index_ = i
|
p._parent_index_ = i
|
||||||
|
|
||||||
pslice = slice(old_size, old_size + p.size)
|
pslice = slice(old_size, old_size + p.size)
|
||||||
|
|
||||||
# first connect all children
|
# first connect all children
|
||||||
p._propagate_param_grad(self.param_array[pslice], self.gradient_full[pslice])
|
p._propagate_param_grad(self.param_array[pslice], self.gradient_full[pslice])
|
||||||
|
|
||||||
# then connect children to self
|
# then connect children to self
|
||||||
self.param_array[pslice] = p.param_array.flat # , requirements=['C', 'W']).ravel(order='C')
|
self.param_array[pslice] = p.param_array.flat # , requirements=['C', 'W']).ravel(order='C')
|
||||||
self.gradient_full[pslice] = p.gradient_full.flat # , requirements=['C', 'W']).ravel(order='C')
|
self.gradient_full[pslice] = p.gradient_full.flat # , requirements=['C', 'W']).ravel(order='C')
|
||||||
|
|
||||||
if not p.param_array.flags['C_CONTIGUOUS']:
|
|
||||||
raise ValueError, "This should not happen! Please write an email to the developers with the code, which reproduces this error. All parameter arrays must be C_CONTIGUOUS"
|
|
||||||
|
|
||||||
p.param_array.data = self.param_array[pslice].data
|
p.param_array.data = self.param_array[pslice].data
|
||||||
p.gradient_full.data = self.gradient_full[pslice].data
|
p.gradient_full.data = self.gradient_full[pslice].data
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -8,6 +8,9 @@ from ..inference.latent_function_inference import var_dtc
|
||||||
from .. import likelihoods
|
from .. import likelihoods
|
||||||
from parameterization.variational import VariationalPosterior
|
from parameterization.variational import VariationalPosterior
|
||||||
|
|
||||||
|
import logging
|
||||||
|
logger = logging.getLogger("sparse gp")
|
||||||
|
|
||||||
class SparseGP(GP):
|
class SparseGP(GP):
|
||||||
"""
|
"""
|
||||||
A general purpose Sparse GP model
|
A general purpose Sparse GP model
|
||||||
|
|
@ -46,7 +49,7 @@ class SparseGP(GP):
|
||||||
self.num_inducing = Z.shape[0]
|
self.num_inducing = Z.shape[0]
|
||||||
|
|
||||||
GP.__init__(self, X, Y, kernel, likelihood, inference_method=inference_method, name=name, Y_metadata=Y_metadata)
|
GP.__init__(self, X, Y, kernel, likelihood, inference_method=inference_method, name=name, Y_metadata=Y_metadata)
|
||||||
|
logger.info("Adding Z as parameter")
|
||||||
self.add_parameter(self.Z, index=0)
|
self.add_parameter(self.Z, index=0)
|
||||||
|
|
||||||
def has_uncertain_inputs(self):
|
def has_uncertain_inputs(self):
|
||||||
|
|
|
||||||
|
|
@ -296,15 +296,16 @@ def bgplvm_simulation_missing_data(optimize=True, verbose=1,
|
||||||
from GPy.models import BayesianGPLVM
|
from GPy.models import BayesianGPLVM
|
||||||
from GPy.inference.latent_function_inference.var_dtc import VarDTCMissingData
|
from GPy.inference.latent_function_inference.var_dtc import VarDTCMissingData
|
||||||
|
|
||||||
D1, D2, D3, N, num_inducing, Q = 13, 5, 8, 45, 7, 9
|
D1, D2, D3, N, num_inducing, Q = 13, 5, 8, 400, 3, 4
|
||||||
_, _, Ylist = _simulate_sincos(D1, D2, D3, N, num_inducing, Q, plot_sim)
|
_, _, Ylist = _simulate_sincos(D1, D2, D3, N, num_inducing, Q, plot_sim)
|
||||||
Y = Ylist[0]
|
Y = Ylist[0]
|
||||||
k = kern.Linear(Q, ARD=True)# + kern.white(Q, _np.exp(-2)) # + kern.bias(Q)
|
k = kern.Linear(Q, ARD=True)# + kern.white(Q, _np.exp(-2)) # + kern.bias(Q)
|
||||||
|
|
||||||
inan = _np.random.binomial(1, .6, size=Y.shape).astype(bool)
|
inan = _np.random.binomial(1, .8, size=Y.shape).astype(bool) # 80% missing data
|
||||||
Y[inan] = _np.nan
|
Ymissing = Y.copy()
|
||||||
|
Ymissing[inan] = _np.nan
|
||||||
|
|
||||||
m = BayesianGPLVM(Y.copy(), Q, init="random", num_inducing=num_inducing,
|
m = BayesianGPLVM(Ymissing, Q, init="random", num_inducing=num_inducing,
|
||||||
inference_method=VarDTCMissingData(inan=inan), kernel=k)
|
inference_method=VarDTCMissingData(inan=inan), kernel=k)
|
||||||
|
|
||||||
m.X.variance[:] = _np.random.uniform(0,.01,m.X.shape)
|
m.X.variance[:] = _np.random.uniform(0,.01,m.X.shape)
|
||||||
|
|
@ -414,7 +415,7 @@ def olivetti_faces(optimize=True, verbose=True, plot=True):
|
||||||
if optimize: m.optimize('bfgs', messages=verbose, max_iters=1000)
|
if optimize: m.optimize('bfgs', messages=verbose, max_iters=1000)
|
||||||
if plot:
|
if plot:
|
||||||
ax = m.plot_latent(which_indices=(0, 1))
|
ax = m.plot_latent(which_indices=(0, 1))
|
||||||
y = m.likelihood.Y[0, :]
|
y = m.Y[0, :]
|
||||||
data_show = GPy.plotting.matplot_dep.visualize.image_show(y[None, :], dimensions=(112, 92), transpose=False, invert=False, scale=False)
|
data_show = GPy.plotting.matplot_dep.visualize.image_show(y[None, :], dimensions=(112, 92), transpose=False, invert=False, scale=False)
|
||||||
lvm = GPy.plotting.matplot_dep.visualize.lvm(m.X.mean[0, :].copy(), m, data_show, ax)
|
lvm = GPy.plotting.matplot_dep.visualize.lvm(m.X.mean[0, :].copy(), m, data_show, ax)
|
||||||
raw_input('Press enter to finish')
|
raw_input('Press enter to finish')
|
||||||
|
|
|
||||||
|
|
@ -9,6 +9,8 @@ import numpy as np
|
||||||
from ...util.misc import param_to_array
|
from ...util.misc import param_to_array
|
||||||
from . import LatentFunctionInference
|
from . import LatentFunctionInference
|
||||||
log_2_pi = np.log(2*np.pi)
|
log_2_pi = np.log(2*np.pi)
|
||||||
|
import logging, itertools
|
||||||
|
logger = logging.getLogger('vardtc')
|
||||||
|
|
||||||
class VarDTC(LatentFunctionInference):
|
class VarDTC(LatentFunctionInference):
|
||||||
"""
|
"""
|
||||||
|
|
@ -196,7 +198,8 @@ class VarDTCMissingData(LatentFunctionInference):
|
||||||
def __init__(self, limit=1, inan=None):
|
def __init__(self, limit=1, inan=None):
|
||||||
from ...util.caching import Cacher
|
from ...util.caching import Cacher
|
||||||
self._Y = Cacher(self._subarray_computations, limit)
|
self._Y = Cacher(self._subarray_computations, limit)
|
||||||
self._inan = inan
|
if inan is not None: self._inan = ~inan
|
||||||
|
else: self._inan = None
|
||||||
pass
|
pass
|
||||||
|
|
||||||
def set_limit(self, limit):
|
def set_limit(self, limit):
|
||||||
|
|
@ -217,21 +220,35 @@ class VarDTCMissingData(LatentFunctionInference):
|
||||||
if self._inan is None:
|
if self._inan is None:
|
||||||
inan = np.isnan(Y)
|
inan = np.isnan(Y)
|
||||||
has_none = inan.any()
|
has_none = inan.any()
|
||||||
|
self._inan = ~inan
|
||||||
else:
|
else:
|
||||||
inan = self._inan
|
inan = self._inan
|
||||||
has_none = True
|
has_none = True
|
||||||
if has_none:
|
if has_none:
|
||||||
from ...util.subarray_and_sorting import common_subarrays
|
#print "caching missing data slices, this can take several minutes depending on the number of unique dimensions of the data..."
|
||||||
self._subarray_indices = []
|
#csa = common_subarrays(inan, 1)
|
||||||
for v,ind in common_subarrays(inan, 1).iteritems():
|
size = Y.shape[1]
|
||||||
if not np.all(v):
|
#logger.info('preparing subarrays {:3.3%}'.format((i+1.)/size))
|
||||||
v = ~np.array(v, dtype=bool)
|
Ys = []
|
||||||
ind = np.array(ind, dtype=int)
|
next_ten = [0.]
|
||||||
if ind.size == Y.shape[1]:
|
count = itertools.count()
|
||||||
ind = slice(None)
|
for v, y in itertools.izip(inan.T, Y.T[:,:,None]):
|
||||||
self._subarray_indices.append([v,ind])
|
i = count.next()
|
||||||
Ys = [Y[v, :][:, ind] for v, ind in self._subarray_indices]
|
if ((i+1.)/size) >= next_ten[0]:
|
||||||
traces = [(y**2).sum() for y in Ys]
|
logger.info('preparing subarrays {:>6.1%}'.format((i+1.)/size))
|
||||||
|
next_ten[0] += .1
|
||||||
|
Ys.append(y[v,:])
|
||||||
|
|
||||||
|
next_ten = [0.]
|
||||||
|
count = itertools.count()
|
||||||
|
def trace(y):
|
||||||
|
i = count.next()
|
||||||
|
if ((i+1.)/size) >= next_ten[0]:
|
||||||
|
logger.info('preparing traces {:>6.1%}'.format((i+1.)/size))
|
||||||
|
next_ten[0] += .1
|
||||||
|
y = y[inan[:,i],i:i+1]
|
||||||
|
return np.einsum('ij,ij->', y,y)
|
||||||
|
traces = [trace(Y) for _ in xrange(size)]
|
||||||
return Ys, traces
|
return Ys, traces
|
||||||
else:
|
else:
|
||||||
self._subarray_indices = [[slice(None),slice(None)]]
|
self._subarray_indices = [[slice(None),slice(None)]]
|
||||||
|
|
@ -253,7 +270,6 @@ class VarDTCMissingData(LatentFunctionInference):
|
||||||
beta_all = 1./np.fmax(likelihood.gaussian_variance(Y_metadata), 1e-6)
|
beta_all = 1./np.fmax(likelihood.gaussian_variance(Y_metadata), 1e-6)
|
||||||
het_noise = beta_all.size != 1
|
het_noise = beta_all.size != 1
|
||||||
|
|
||||||
import itertools
|
|
||||||
num_inducing = Z.shape[0]
|
num_inducing = Z.shape[0]
|
||||||
|
|
||||||
dL_dpsi0_all = np.zeros(Y.shape[0])
|
dL_dpsi0_all = np.zeros(Y.shape[0])
|
||||||
|
|
@ -273,22 +289,24 @@ class VarDTCMissingData(LatentFunctionInference):
|
||||||
Lm = jitchol(Kmm)
|
Lm = jitchol(Kmm)
|
||||||
if uncertain_inputs: LmInv = dtrtri(Lm)
|
if uncertain_inputs: LmInv = dtrtri(Lm)
|
||||||
|
|
||||||
VVT_factor_all = np.empty(Y.shape)
|
#VVT_factor_all = np.empty(Y.shape)
|
||||||
full_VVT_factor = VVT_factor_all.shape[1] == Y.shape[1]
|
#full_VVT_factor = VVT_factor_all.shape[1] == Y.shape[1]
|
||||||
if not full_VVT_factor:
|
#if not full_VVT_factor:
|
||||||
psi1V = np.dot(Y.T*beta_all, psi1_all).T
|
# psi1V = np.dot(Y.T*beta_all, psi1_all).T
|
||||||
|
|
||||||
for y, trYYT, [v, ind] in itertools.izip(Ys, traces, self._subarray_indices):
|
#logger.info('computing dimension-wise likelihood and derivatives')
|
||||||
if het_noise: beta = beta_all[ind]
|
#size = len(Ys)
|
||||||
|
size = Y.shape[1]
|
||||||
|
next_ten = 0
|
||||||
|
for i, [y, v, trYYT] in enumerate(itertools.izip(Ys, self._inan.T, traces)):
|
||||||
|
if ((i+1.)/size) >= next_ten:
|
||||||
|
logger.info('inference {:> 6.1%}'.format((i+1.)/size))
|
||||||
|
next_ten += .1
|
||||||
|
if het_noise: beta = beta_all[i]
|
||||||
else: beta = beta_all
|
else: beta = beta_all
|
||||||
|
|
||||||
VVT_factor = (beta*y)
|
VVT_factor = (y*beta)
|
||||||
try:
|
output_dim = 1#len(ind)
|
||||||
VVT_factor_all[v, ind].flat = VVT_factor.flat
|
|
||||||
except ValueError:
|
|
||||||
mult = np.ravel_multi_index((v.nonzero()[0][:,None],ind[None,:]), VVT_factor_all.shape)
|
|
||||||
VVT_factor_all.flat[mult] = VVT_factor
|
|
||||||
output_dim = y.shape[1]
|
|
||||||
|
|
||||||
psi0 = psi0_all[v]
|
psi0 = psi0_all[v]
|
||||||
psi1 = psi1_all[v, :]
|
psi1 = psi1_all[v, :]
|
||||||
|
|
@ -347,19 +365,20 @@ class VarDTCMissingData(LatentFunctionInference):
|
||||||
psi0, psi1, beta,
|
psi0, psi1, beta,
|
||||||
data_fit, num_data, output_dim, trYYT, Y)
|
data_fit, num_data, output_dim, trYYT, Y)
|
||||||
|
|
||||||
if full_VVT_factor: woodbury_vector[:, ind] = Cpsi1Vf
|
#if full_VVT_factor:
|
||||||
else:
|
woodbury_vector[:, i:i+1] = Cpsi1Vf
|
||||||
print 'foobar'
|
#else:
|
||||||
tmp, _ = dtrtrs(Lm, psi1V, lower=1, trans=0)
|
# print 'foobar'
|
||||||
tmp, _ = dpotrs(LB, tmp, lower=1)
|
# tmp, _ = dtrtrs(Lm, psi1V, lower=1, trans=0)
|
||||||
woodbury_vector[:, ind] = dtrtrs(Lm, tmp, lower=1, trans=1)[0]
|
# tmp, _ = dpotrs(LB, tmp, lower=1)
|
||||||
|
# woodbury_vector[:, ind] = dtrtrs(Lm, tmp, lower=1, trans=1)[0]
|
||||||
|
|
||||||
#import ipdb;ipdb.set_trace()
|
#import ipdb;ipdb.set_trace()
|
||||||
Bi, _ = dpotri(LB, lower=1)
|
Bi, _ = dpotri(LB, lower=1)
|
||||||
symmetrify(Bi)
|
symmetrify(Bi)
|
||||||
Bi = -dpotri(LB, lower=1)[0]
|
Bi = -dpotri(LB, lower=1)[0]
|
||||||
diag.add(Bi, 1)
|
diag.add(Bi, 1)
|
||||||
woodbury_inv_all[:, :, ind] = backsub_both_sides(Lm, Bi)[:,:,None]
|
woodbury_inv_all[:, :, i:i+1] = backsub_both_sides(Lm, Bi)[:,:,None]
|
||||||
|
|
||||||
dL_dthetaL = likelihood.exact_inference_gradients(dL_dR)
|
dL_dthetaL = likelihood.exact_inference_gradients(dL_dR)
|
||||||
|
|
||||||
|
|
@ -376,23 +395,6 @@ class VarDTCMissingData(LatentFunctionInference):
|
||||||
'dL_dKnm':dL_dpsi1_all,
|
'dL_dKnm':dL_dpsi1_all,
|
||||||
'dL_dthetaL':dL_dthetaL}
|
'dL_dthetaL':dL_dthetaL}
|
||||||
|
|
||||||
#get sufficient things for posterior prediction
|
|
||||||
#TODO: do we really want to do this in the loop?
|
|
||||||
#if not full_VVT_factor:
|
|
||||||
# print 'foobar'
|
|
||||||
# psi1V = np.dot(Y.T*beta_all, psi1_all).T
|
|
||||||
# tmp, _ = dtrtrs(Lm, psi1V, lower=1, trans=0)
|
|
||||||
# tmp, _ = dpotrs(LB_all, tmp, lower=1)
|
|
||||||
# woodbury_vector, _ = dtrtrs(Lm, tmp, lower=1, trans=1)
|
|
||||||
#import ipdb;ipdb.set_trace()
|
|
||||||
#Bi, _ = dpotri(LB_all, lower=1)
|
|
||||||
#symmetrify(Bi)
|
|
||||||
#Bi = -dpotri(LB_all, lower=1)[0]
|
|
||||||
#from ...util import diag
|
|
||||||
#diag.add(Bi, 1)
|
|
||||||
|
|
||||||
#woodbury_inv = backsub_both_sides(Lm, Bi)
|
|
||||||
|
|
||||||
post = Posterior(woodbury_inv=woodbury_inv_all, woodbury_vector=woodbury_vector, K=Kmm, mean=None, cov=None, K_chol=Lm)
|
post = Posterior(woodbury_inv=woodbury_inv_all, woodbury_vector=woodbury_vector, K=Kmm, mean=None, cov=None, K_chol=Lm)
|
||||||
|
|
||||||
return post, log_marginal, grad_dict
|
return post, log_marginal, grad_dict
|
||||||
|
|
|
||||||
|
|
@ -112,12 +112,12 @@ class VarDTC_minibatch(LatentFunctionInference):
|
||||||
if het_noise:
|
if het_noise:
|
||||||
psi2_full += beta_slice*psi2
|
psi2_full += beta_slice*psi2
|
||||||
else:
|
else:
|
||||||
psi2_full += psi2
|
psi2_full += psi2.sum(0)
|
||||||
else:
|
else:
|
||||||
if het_noise:
|
if het_noise:
|
||||||
psi2_full += beta_slice*np.outer(psi1,psi1)
|
psi2_full += beta_slice*np.outer(psi1,psi1)
|
||||||
else:
|
else:
|
||||||
psi2_full += np.outer(psi1,psi1)
|
psi2_full += np.einsum('nm,jk->mk',psi1,psi1)
|
||||||
|
|
||||||
if not het_noise:
|
if not het_noise:
|
||||||
psi0_full *= beta
|
psi0_full *= beta
|
||||||
|
|
@ -128,7 +128,7 @@ class VarDTC_minibatch(LatentFunctionInference):
|
||||||
#======================================================================
|
#======================================================================
|
||||||
# Compute Common Components
|
# Compute Common Components
|
||||||
#======================================================================
|
#======================================================================
|
||||||
|
self.psi1Y = psi1Y_full
|
||||||
Kmm = kern.K(Z).copy()
|
Kmm = kern.K(Z).copy()
|
||||||
diag.add(Kmm, self.const_jitter)
|
diag.add(Kmm, self.const_jitter)
|
||||||
Lm = jitchol(Kmm)
|
Lm = jitchol(Kmm)
|
||||||
|
|
@ -159,7 +159,10 @@ class VarDTC_minibatch(LatentFunctionInference):
|
||||||
logL_R = -np.log(beta).sum()
|
logL_R = -np.log(beta).sum()
|
||||||
else:
|
else:
|
||||||
logL_R = -num_data*np.log(beta)
|
logL_R = -num_data*np.log(beta)
|
||||||
logL = -(output_dim*(num_data*log_2_pi+logL_R+psi0_full-np.trace(LmInvPsi2LmInvT))+YRY_full-bbt)/2.-output_dim*(-np.log(np.diag(Lm)).sum()+np.log(np.diag(LL)).sum())
|
logL = (
|
||||||
|
-(output_dim*(num_data*log_2_pi+logL_R+psi0_full-np.trace(LmInvPsi2LmInvT))+YRY_full-bbt)/2.
|
||||||
|
-output_dim*(-np.log(np.diag(Lm)).sum()+np.log(np.diag(LL)).sum())
|
||||||
|
)
|
||||||
|
|
||||||
#======================================================================
|
#======================================================================
|
||||||
# Compute dL_dKmm
|
# Compute dL_dKmm
|
||||||
|
|
@ -256,14 +259,14 @@ class VarDTC_minibatch(LatentFunctionInference):
|
||||||
|
|
||||||
if het_noise:
|
if het_noise:
|
||||||
if uncertain_inputs:
|
if uncertain_inputs:
|
||||||
psiR = np.einsum('mo,nmo->n',dL_dpsi2R,psi2)
|
psiR = np.einsum('mo,nmo->',dL_dpsi2R,psi2)
|
||||||
else:
|
else:
|
||||||
psiR = np.einsum('nm,no,mo->n',psi1,psi1,dL_dpsi2R)
|
psiR = np.einsum('nm,no,mo->',psi1,psi1,dL_dpsi2R)
|
||||||
|
|
||||||
dL_dthetaL = ((np.square(betaY)).sum(axis=-1) + np.square(beta)*(output_dim*psi0)-output_dim*beta)/2. - np.square(beta)*psiR- (betaY*np.dot(betapsi1,v)).sum(axis=-1)
|
dL_dthetaL = ((np.square(betaY)).sum(axis=-1) + np.square(beta)*(output_dim*psi0)-output_dim*beta)/2. - np.square(beta)*psiR- (betaY*np.dot(betapsi1,v)).sum(axis=-1)
|
||||||
else:
|
else:
|
||||||
if uncertain_inputs:
|
if uncertain_inputs:
|
||||||
psiR = np.einsum('mo,mo->',dL_dpsi2R,psi2)
|
psiR = np.einsum('mo,nmo->',dL_dpsi2R,psi2)
|
||||||
else:
|
else:
|
||||||
psiR = np.einsum('nm,no,mo->',psi1,psi1,dL_dpsi2R)
|
psiR = np.einsum('nm,no,mo->',psi1,psi1,dL_dpsi2R)
|
||||||
|
|
||||||
|
|
@ -305,30 +308,44 @@ def update_gradients(model):
|
||||||
if isinstance(model.X, VariationalPosterior):
|
if isinstance(model.X, VariationalPosterior):
|
||||||
X_slice = model.X[n_range[0]:n_range[1]]
|
X_slice = model.X[n_range[0]:n_range[1]]
|
||||||
|
|
||||||
|
dL_dpsi1 = grad_dict['dL_dpsi1']#[None, :]
|
||||||
|
dL_dpsi2 = grad_dict['dL_dpsi2'][None, :, :]
|
||||||
#gradients w.r.t. kernel
|
#gradients w.r.t. kernel
|
||||||
model.kern.update_gradients_expectations(variational_posterior=X_slice, Z=model.Z, dL_dpsi0=grad_dict['dL_dpsi0'], dL_dpsi1=grad_dict['dL_dpsi1'], dL_dpsi2=grad_dict['dL_dpsi2'])
|
model.kern.update_gradients_expectations(variational_posterior=X_slice,Z=model.Z,dL_dpsi0=grad_dict['dL_dpsi0'],dL_dpsi1=dL_dpsi1,dL_dpsi2=dL_dpsi2)
|
||||||
kern_grad += model.kern.gradient
|
kern_grad += model.kern.gradient
|
||||||
|
|
||||||
#gradients w.r.t. Z
|
#gradients w.r.t. Z
|
||||||
model.Z.gradient += model.kern.gradients_Z_expectations(
|
model.Z.gradient += model.kern.gradients_Z_expectations(
|
||||||
dL_dpsi0=grad_dict['dL_dpsi0'], dL_dpsi1=grad_dict['dL_dpsi1'], dL_dpsi2=grad_dict['dL_dpsi2'], Z=model.Z, variational_posterior=X_slice)
|
dL_dpsi0=grad_dict['dL_dpsi0'],
|
||||||
|
dL_dpsi1=dL_dpsi1,
|
||||||
|
dL_dpsi2=dL_dpsi2,
|
||||||
|
Z=model.Z, variational_posterior=X_slice)
|
||||||
|
|
||||||
#gradients w.r.t. posterior parameters of X
|
#gradients w.r.t. posterior parameters of X
|
||||||
X_grad = model.kern.gradients_qX_expectations(variational_posterior=X_slice, Z=model.Z, dL_dpsi0=grad_dict['dL_dpsi0'], dL_dpsi1=grad_dict['dL_dpsi1'], dL_dpsi2=grad_dict['dL_dpsi2'])
|
X_grad = model.kern.gradients_qX_expectations(
|
||||||
model.set_X_gradients(X_slice, X_grad)
|
variational_posterior=X_slice,
|
||||||
|
Z=model.Z,
|
||||||
|
dL_dpsi0=grad_dict['dL_dpsi0'],
|
||||||
|
dL_dpsi1=dL_dpsi1,
|
||||||
|
dL_dpsi2=dL_dpsi2)
|
||||||
|
|
||||||
|
model.X.mean[n_range[0]:n_range[1]].gradient = X_grad[0]
|
||||||
|
model.X.variance[n_range[0]:n_range[1]].gradient = X_grad[1]
|
||||||
|
|
||||||
if het_noise:
|
if het_noise:
|
||||||
dL_dthetaL[n_range[0]:n_range[1]] = grad_dict['dL_dthetaL']
|
dL_dthetaL[n_range[0]:n_range[1]] = grad_dict['dL_dthetaL']
|
||||||
else:
|
else:
|
||||||
dL_dthetaL += grad_dict['dL_dthetaL']
|
dL_dthetaL += grad_dict['dL_dthetaL']
|
||||||
|
#import ipdb;ipdb.set_trace()
|
||||||
|
model.grad_dict = grad_dict
|
||||||
|
if isinstance(model.X, VariationalPosterior):
|
||||||
|
# Update Log-likelihood
|
||||||
|
model._log_marginal_likelihood -= model.variational_prior.KL_divergence(model.X)
|
||||||
|
# update for the KL divergence
|
||||||
|
model.variational_prior.update_gradients_KL(model.X)
|
||||||
|
|
||||||
# Set the gradients w.r.t. kernel
|
# Set the gradients w.r.t. kernel
|
||||||
model.kern.gradient = kern_grad
|
model.kern.gradient = kern_grad
|
||||||
|
|
||||||
# Update Log-likelihood
|
|
||||||
model._log_marginal_likelihood -= model.variational_prior.KL_divergence(model.X)
|
|
||||||
# update for the KL divergence
|
|
||||||
model.variational_prior.update_gradients_KL(model.X)
|
|
||||||
|
|
||||||
# dL_dthetaL
|
# dL_dthetaL
|
||||||
model.likelihood.update_gradients(dL_dthetaL)
|
model.likelihood.update_gradients(dL_dthetaL)
|
||||||
|
|
|
||||||
|
|
@ -56,13 +56,13 @@ def SCG(f, gradf, x, optargs=(), maxiters=500, max_f_eval=np.inf, display=True,
|
||||||
if gtol is None:
|
if gtol is None:
|
||||||
gtol = 1e-5
|
gtol = 1e-5
|
||||||
|
|
||||||
sigma0 = 1.0e-8
|
sigma0 = 1.0e-7
|
||||||
fold = f(x, *optargs) # Initial function value.
|
fold = f(x, *optargs) # Initial function value.
|
||||||
function_eval = 1
|
function_eval = 1
|
||||||
fnow = fold
|
fnow = fold
|
||||||
gradnew = gradf(x, *optargs) # Initial gradient.
|
gradnew = gradf(x, *optargs) # Initial gradient.
|
||||||
if any(np.isnan(gradnew)):
|
#if any(np.isnan(gradnew)):
|
||||||
raise UnexpectedInfOrNan, "Gradient contribution resulted in a NaN value"
|
# raise UnexpectedInfOrNan, "Gradient contribution resulted in a NaN value"
|
||||||
current_grad = np.dot(gradnew, gradnew)
|
current_grad = np.dot(gradnew, gradnew)
|
||||||
gradold = gradnew.copy()
|
gradold = gradnew.copy()
|
||||||
d = -gradnew # Initial search direction.
|
d = -gradnew # Initial search direction.
|
||||||
|
|
@ -168,13 +168,13 @@ def SCG(f, gradf, x, optargs=(), maxiters=500, max_f_eval=np.inf, display=True,
|
||||||
if Delta < 0.25:
|
if Delta < 0.25:
|
||||||
beta = min(4.0 * beta, betamax)
|
beta = min(4.0 * beta, betamax)
|
||||||
if Delta > 0.75:
|
if Delta > 0.75:
|
||||||
beta = max(0.5 * beta, betamin)
|
beta = max(0.25 * beta, betamin)
|
||||||
|
|
||||||
# Update search direction using Polak-Ribiere formula, or re-start
|
# Update search direction using Polak-Ribiere formula, or re-start
|
||||||
# in direction of negative gradient after nparams steps.
|
# in direction of negative gradient after nparams steps.
|
||||||
if nsuccess == x.size:
|
if nsuccess == x.size:
|
||||||
d = -gradnew
|
d = -gradnew
|
||||||
# beta = 1. # TODO: betareset!!
|
beta = 1. # This is not in the original paper
|
||||||
nsuccess = 0
|
nsuccess = 0
|
||||||
elif success:
|
elif success:
|
||||||
Gamma = np.dot(gradold - gradnew, gradnew) / (mu)
|
Gamma = np.dot(gradold - gradnew, gradnew) / (mu)
|
||||||
|
|
|
||||||
|
|
@ -1,2 +1,2 @@
|
||||||
# This is the local configuration file for GPy
|
# This is the local installation configuration file for GPy
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -20,6 +20,8 @@ def index_to_slices(index):
|
||||||
returns
|
returns
|
||||||
>>> [[slice(0,2,None),slice(4,5,None)],[slice(2,4,None),slice(8,10,None)],[slice(5,8,None)]]
|
>>> [[slice(0,2,None),slice(4,5,None)],[slice(2,4,None),slice(8,10,None)],[slice(5,8,None)]]
|
||||||
"""
|
"""
|
||||||
|
if len(index)==0:
|
||||||
|
return[]
|
||||||
|
|
||||||
#contruct the return structure
|
#contruct the return structure
|
||||||
ind = np.asarray(index,dtype=np.int)
|
ind = np.asarray(index,dtype=np.int)
|
||||||
|
|
|
||||||
|
|
@ -101,6 +101,7 @@ class PeriodicExponential(Periodic):
|
||||||
Flower = np.array(self._cos(self.basis_alpha,self.basis_omega,self.basis_phi)(self.lower))[:,None]
|
Flower = np.array(self._cos(self.basis_alpha,self.basis_omega,self.basis_phi)(self.lower))[:,None]
|
||||||
return(self.lengthscale/(2*self.variance) * Gint + 1./self.variance*np.dot(Flower,Flower.T))
|
return(self.lengthscale/(2*self.variance) * Gint + 1./self.variance*np.dot(Flower,Flower.T))
|
||||||
|
|
||||||
|
@silence_errors
|
||||||
def update_gradients_full(self, dL_dK, X, X2=None):
|
def update_gradients_full(self, dL_dK, X, X2=None):
|
||||||
"""derivative of the covariance matrix with respect to the parameters (shape is N x num_inducing x num_params)"""
|
"""derivative of the covariance matrix with respect to the parameters (shape is N x num_inducing x num_params)"""
|
||||||
if X2 is None: X2 = X
|
if X2 is None: X2 = X
|
||||||
|
|
@ -213,7 +214,7 @@ class PeriodicMatern32(Periodic):
|
||||||
return(self.lengthscale**3/(12*np.sqrt(3)*self.variance) * Gint + 1./self.variance*np.dot(Flower,Flower.T) + self.lengthscale**2/(3.*self.variance)*np.dot(F1lower,F1lower.T))
|
return(self.lengthscale**3/(12*np.sqrt(3)*self.variance) * Gint + 1./self.variance*np.dot(Flower,Flower.T) + self.lengthscale**2/(3.*self.variance)*np.dot(F1lower,F1lower.T))
|
||||||
|
|
||||||
|
|
||||||
#@silence_errors
|
@silence_errors
|
||||||
def update_gradients_full(self,dL_dK,X,X2):
|
def update_gradients_full(self,dL_dK,X,X2):
|
||||||
"""derivative of the covariance matrix with respect to the parameters (shape is num_data x num_inducing x num_params)"""
|
"""derivative of the covariance matrix with respect to the parameters (shape is num_data x num_inducing x num_params)"""
|
||||||
if X2 is None: X2 = X
|
if X2 is None: X2 = X
|
||||||
|
|
|
||||||
|
|
@ -20,6 +20,9 @@ class DiffGenomeKern(Kern):
|
||||||
assert X2==None
|
assert X2==None
|
||||||
K = self.kern.K(X,X2)
|
K = self.kern.K(X,X2)
|
||||||
|
|
||||||
|
if self.idx_p<=0 or self.idx_p>X.shape[0]/2:
|
||||||
|
return K
|
||||||
|
|
||||||
slices = index_to_slices(X[:,self.index_dim])
|
slices = index_to_slices(X[:,self.index_dim])
|
||||||
idx_start = slices[1][0].start
|
idx_start = slices[1][0].start
|
||||||
idx_end = idx_start+self.idx_p
|
idx_end = idx_start+self.idx_p
|
||||||
|
|
@ -33,6 +36,9 @@ class DiffGenomeKern(Kern):
|
||||||
def Kdiag(self,X):
|
def Kdiag(self,X):
|
||||||
Kdiag = self.kern.Kdiag(X)
|
Kdiag = self.kern.Kdiag(X)
|
||||||
|
|
||||||
|
if self.idx_p<=0 or self.idx_p>X.shape[0]/2:
|
||||||
|
return Kdiag
|
||||||
|
|
||||||
slices = index_to_slices(X[:,self.index_dim])
|
slices = index_to_slices(X[:,self.index_dim])
|
||||||
idx_start = slices[1][0].start
|
idx_start = slices[1][0].start
|
||||||
idx_end = idx_start+self.idx_p
|
idx_end = idx_start+self.idx_p
|
||||||
|
|
@ -42,6 +48,10 @@ class DiffGenomeKern(Kern):
|
||||||
|
|
||||||
def update_gradients_full(self,dL_dK,X,X2=None):
|
def update_gradients_full(self,dL_dK,X,X2=None):
|
||||||
assert X2==None
|
assert X2==None
|
||||||
|
if self.idx_p<=0 or self.idx_p>X.shape[0]/2:
|
||||||
|
self.kern.update_gradients_full(dL_dK, X)
|
||||||
|
return
|
||||||
|
|
||||||
slices = index_to_slices(X[:,self.index_dim])
|
slices = index_to_slices(X[:,self.index_dim])
|
||||||
idx_start = slices[1][0].start
|
idx_start = slices[1][0].start
|
||||||
idx_end = idx_start+self.idx_p
|
idx_end = idx_start+self.idx_p
|
||||||
|
|
|
||||||
|
|
@ -37,19 +37,21 @@ class BayesianGPLVM(SparseGP):
|
||||||
self.init = init
|
self.init = init
|
||||||
|
|
||||||
if X_variance is None:
|
if X_variance is None:
|
||||||
|
self.logger.info("initializing latent space variance ~ uniform(0,.1)")
|
||||||
X_variance = np.random.uniform(0,.1,X.shape)
|
X_variance = np.random.uniform(0,.1,X.shape)
|
||||||
|
|
||||||
if Z is None:
|
if Z is None:
|
||||||
|
self.logger.info("initializing inducing inputs")
|
||||||
Z = np.random.permutation(X.copy())[:num_inducing]
|
Z = np.random.permutation(X.copy())[:num_inducing]
|
||||||
assert Z.shape[1] == X.shape[1]
|
assert Z.shape[1] == X.shape[1]
|
||||||
|
|
||||||
if kernel is None:
|
if kernel is None:
|
||||||
|
self.logger.info("initializing kernel RBF")
|
||||||
kernel = kern.RBF(input_dim, lengthscale=1./fracs, ARD=True) # + kern.white(input_dim)
|
kernel = kern.RBF(input_dim, lengthscale=1./fracs, ARD=True) # + kern.white(input_dim)
|
||||||
|
|
||||||
if likelihood is None:
|
if likelihood is None:
|
||||||
likelihood = Gaussian()
|
likelihood = Gaussian()
|
||||||
|
|
||||||
|
|
||||||
self.variational_prior = NormalPrior()
|
self.variational_prior = NormalPrior()
|
||||||
X = NormalPosterior(X, X_variance)
|
X = NormalPosterior(X, X_variance)
|
||||||
|
|
||||||
|
|
@ -65,6 +67,7 @@ class BayesianGPLVM(SparseGP):
|
||||||
inference_method = VarDTC()
|
inference_method = VarDTC()
|
||||||
|
|
||||||
SparseGP.__init__(self, X, Y, Z, kernel, likelihood, inference_method, name, **kwargs)
|
SparseGP.__init__(self, X, Y, Z, kernel, likelihood, inference_method, name, **kwargs)
|
||||||
|
self.logger.info("Adding X as parameter")
|
||||||
self.add_parameter(self.X, index=0)
|
self.add_parameter(self.X, index=0)
|
||||||
|
|
||||||
def set_X_gradients(self, X, X_grad):
|
def set_X_gradients(self, X, X_grad):
|
||||||
|
|
|
||||||
|
|
@ -8,7 +8,7 @@ from base_plots import gpplot, x_frame1D, x_frame2D
|
||||||
from ...util.misc import param_to_array
|
from ...util.misc import param_to_array
|
||||||
from ...models.gp_coregionalized_regression import GPCoregionalizedRegression
|
from ...models.gp_coregionalized_regression import GPCoregionalizedRegression
|
||||||
from ...models.sparse_gp_coregionalized_regression import SparseGPCoregionalizedRegression
|
from ...models.sparse_gp_coregionalized_regression import SparseGPCoregionalizedRegression
|
||||||
|
from scipy import sparse
|
||||||
|
|
||||||
def plot_fit(model, plot_limits=None, which_data_rows='all',
|
def plot_fit(model, plot_limits=None, which_data_rows='all',
|
||||||
which_data_ycols='all', fixed_inputs=[],
|
which_data_ycols='all', fixed_inputs=[],
|
||||||
|
|
@ -61,11 +61,14 @@ def plot_fit(model, plot_limits=None, which_data_rows='all',
|
||||||
|
|
||||||
if hasattr(model, 'has_uncertain_inputs') and model.has_uncertain_inputs():
|
if hasattr(model, 'has_uncertain_inputs') and model.has_uncertain_inputs():
|
||||||
X = model.X.mean
|
X = model.X.mean
|
||||||
X_variance = param_to_array(model.X.variance)
|
X_variance = model.X.variance
|
||||||
else:
|
else:
|
||||||
X = model.X
|
X = model.X
|
||||||
X, Y = param_to_array(X, model.Y)
|
#X, Y = param_to_array(X, model.Y)
|
||||||
if hasattr(model, 'Z'): Z = param_to_array(model.Z)
|
Y = model.Y
|
||||||
|
if sparse.issparse(Y): Y = Y.todense().view(np.ndarray)
|
||||||
|
|
||||||
|
if hasattr(model, 'Z'): Z = model.Z
|
||||||
|
|
||||||
#work out what the inputs are for plotting (1D or 2D)
|
#work out what the inputs are for plotting (1D or 2D)
|
||||||
fixed_dims = np.array([i for i,v in fixed_inputs])
|
fixed_dims = np.array([i for i,v in fixed_inputs])
|
||||||
|
|
|
||||||
|
|
@ -8,6 +8,7 @@ import GPy
|
||||||
import numpy as np
|
import numpy as np
|
||||||
from GPy.core.parameterization.parameter_core import HierarchyError
|
from GPy.core.parameterization.parameter_core import HierarchyError
|
||||||
from GPy.core.parameterization.observable_array import ObsAr
|
from GPy.core.parameterization.observable_array import ObsAr
|
||||||
|
from GPy.core.parameterization.transformations import NegativeLogexp
|
||||||
|
|
||||||
class ArrayCoreTest(unittest.TestCase):
|
class ArrayCoreTest(unittest.TestCase):
|
||||||
def setUp(self):
|
def setUp(self):
|
||||||
|
|
@ -38,10 +39,25 @@ class ParameterizedTest(unittest.TestCase):
|
||||||
self.test1.kern = self.rbf+self.white
|
self.test1.kern = self.rbf+self.white
|
||||||
self.test1.add_parameter(self.test1.kern)
|
self.test1.add_parameter(self.test1.kern)
|
||||||
self.test1.add_parameter(self.param, 0)
|
self.test1.add_parameter(self.param, 0)
|
||||||
|
# print self.test1:
|
||||||
|
#=============================================================================
|
||||||
|
# test_model. | Value | Constraint | Prior | Tied to
|
||||||
|
# param | (25L, 2L) | {0.0,1.0} | |
|
||||||
|
# add.rbf.variance | 1.0 | 0.0,1.0 +ve | |
|
||||||
|
# add.rbf.lengthscale | 1.0 | 0.0,1.0 +ve | |
|
||||||
|
# add.white.variance | 1.0 | 0.0,1.0 +ve | |
|
||||||
|
#=============================================================================
|
||||||
|
|
||||||
x = np.linspace(-2,6,4)[:,None]
|
x = np.linspace(-2,6,4)[:,None]
|
||||||
y = np.sin(x)
|
y = np.sin(x)
|
||||||
self.testmodel = GPy.models.GPRegression(x,y)
|
self.testmodel = GPy.models.GPRegression(x,y)
|
||||||
|
# print self.testmodel:
|
||||||
|
#=============================================================================
|
||||||
|
# GP_regression. | Value | Constraint | Prior | Tied to
|
||||||
|
# rbf.variance | 1.0 | +ve | |
|
||||||
|
# rbf.lengthscale | 1.0 | +ve | |
|
||||||
|
# Gaussian_noise.variance | 1.0 | +ve | |
|
||||||
|
#=============================================================================
|
||||||
|
|
||||||
def test_add_parameter(self):
|
def test_add_parameter(self):
|
||||||
self.assertEquals(self.rbf._parent_index_, 0)
|
self.assertEquals(self.rbf._parent_index_, 0)
|
||||||
|
|
@ -142,7 +158,12 @@ class ParameterizedTest(unittest.TestCase):
|
||||||
self.testmodel.randomize()
|
self.testmodel.randomize()
|
||||||
self.assertEqual(val, self.testmodel.kern.lengthscale)
|
self.assertEqual(val, self.testmodel.kern.lengthscale)
|
||||||
|
|
||||||
|
def test_add_parameter_in_hierarchy(self):
|
||||||
|
from GPy.core import Param
|
||||||
|
self.test1.kern.rbf.add_parameter(Param("NEW", np.random.rand(2), NegativeLogexp()), 1)
|
||||||
|
self.assertListEqual(self.test1.constraints[NegativeLogexp()].tolist(), range(self.param.size+1, self.param.size+1 + 2))
|
||||||
|
self.assertListEqual(self.test1.constraints[GPy.transformations.Logistic(0,1)].tolist(), range(self.param.size))
|
||||||
|
self.assertListEqual(self.test1.constraints[GPy.transformations.Logexp(0,1)].tolist(), np.r_[50, 53:55].tolist())
|
||||||
|
|
||||||
def test_regular_expression_misc(self):
|
def test_regular_expression_misc(self):
|
||||||
self.testmodel.kern.lengthscale.fix()
|
self.testmodel.kern.lengthscale.fix()
|
||||||
|
|
|
||||||
|
|
@ -18,7 +18,6 @@ class Cacher(object):
|
||||||
self.operation = operation
|
self.operation = operation
|
||||||
self.order = collections.deque()
|
self.order = collections.deque()
|
||||||
self.cached_inputs = {} # point from cache_ids to a list of [ind_ids], which where used in cache cache_id
|
self.cached_inputs = {} # point from cache_ids to a list of [ind_ids], which where used in cache cache_id
|
||||||
self.logger = logging.getLogger("cache")
|
|
||||||
|
|
||||||
#=======================================================================
|
#=======================================================================
|
||||||
# point from each ind_id to [ref(obj), cache_ids]
|
# point from each ind_id to [ref(obj), cache_ids]
|
||||||
|
|
@ -36,23 +35,18 @@ class Cacher(object):
|
||||||
|
|
||||||
def combine_inputs(self, args, kw):
|
def combine_inputs(self, args, kw):
|
||||||
"Combines the args and kw in a unique way, such that ordering of kwargs does not lead to recompute"
|
"Combines the args and kw in a unique way, such that ordering of kwargs does not lead to recompute"
|
||||||
self.logger.debug("combining args and kw")
|
|
||||||
return args + tuple(c[1] for c in sorted(kw.items(), key=lambda x: x[0]))
|
return args + tuple(c[1] for c in sorted(kw.items(), key=lambda x: x[0]))
|
||||||
|
|
||||||
def prepare_cache_id(self, combined_args_kw, ignore_args):
|
def prepare_cache_id(self, combined_args_kw, ignore_args):
|
||||||
"get the cacheid (conc. string of argument self.ids in order) ignoring ignore_args"
|
"get the cacheid (conc. string of argument self.ids in order) ignoring ignore_args"
|
||||||
cache_id = "".join(self.id(a) for i, a in enumerate(combined_args_kw) if i not in ignore_args)
|
cache_id = "".join(self.id(a) for i, a in enumerate(combined_args_kw) if i not in ignore_args)
|
||||||
self.logger.debug("cache_id={} was created".format(cache_id))
|
|
||||||
return cache_id
|
return cache_id
|
||||||
|
|
||||||
def ensure_cache_length(self, cache_id):
|
def ensure_cache_length(self, cache_id):
|
||||||
"Ensures the cache is within its limits and has one place free"
|
"Ensures the cache is within its limits and has one place free"
|
||||||
self.logger.debug("cache length gets ensured")
|
|
||||||
if len(self.order) == self.limit:
|
if len(self.order) == self.limit:
|
||||||
self.logger.debug("cache limit of l={} was reached".format(self.limit))
|
|
||||||
# we have reached the limit, so lets release one element
|
# we have reached the limit, so lets release one element
|
||||||
cache_id = self.order.popleft()
|
cache_id = self.order.popleft()
|
||||||
self.logger.debug("cach_id '{}' gets removed".format(cache_id))
|
|
||||||
combined_args_kw = self.cached_inputs[cache_id]
|
combined_args_kw = self.cached_inputs[cache_id]
|
||||||
for ind in combined_args_kw:
|
for ind in combined_args_kw:
|
||||||
if ind is not None:
|
if ind is not None:
|
||||||
|
|
@ -66,7 +60,6 @@ class Cacher(object):
|
||||||
else:
|
else:
|
||||||
cache_ids.remove(cache_id)
|
cache_ids.remove(cache_id)
|
||||||
self.cached_input_ids[ind_id] = [ref, cache_ids]
|
self.cached_input_ids[ind_id] = [ref, cache_ids]
|
||||||
self.logger.debug("removing caches")
|
|
||||||
del self.cached_outputs[cache_id]
|
del self.cached_outputs[cache_id]
|
||||||
del self.inputs_changed[cache_id]
|
del self.inputs_changed[cache_id]
|
||||||
del self.cached_inputs[cache_id]
|
del self.cached_inputs[cache_id]
|
||||||
|
|
@ -81,10 +74,8 @@ class Cacher(object):
|
||||||
if a is not None:
|
if a is not None:
|
||||||
ind_id = self.id(a)
|
ind_id = self.id(a)
|
||||||
v = self.cached_input_ids.get(ind_id, [weakref.ref(a), []])
|
v = self.cached_input_ids.get(ind_id, [weakref.ref(a), []])
|
||||||
self.logger.debug("cache_id '{}' gets stored".format(cache_id))
|
|
||||||
v[1].append(cache_id)
|
v[1].append(cache_id)
|
||||||
if len(v[1]) == 1:
|
if len(v[1]) == 1:
|
||||||
self.logger.debug("adding observer to object {}".format(repr(a)))
|
|
||||||
a.add_observer(self, self.on_cache_changed)
|
a.add_observer(self, self.on_cache_changed)
|
||||||
self.cached_input_ids[ind_id] = v
|
self.cached_input_ids[ind_id] = v
|
||||||
|
|
||||||
|
|
@ -108,28 +99,21 @@ class Cacher(object):
|
||||||
cache_id = self.prepare_cache_id(inputs, self.ignore_args)
|
cache_id = self.prepare_cache_id(inputs, self.ignore_args)
|
||||||
# 2: if anything is not cachable, we will just return the operation, without caching
|
# 2: if anything is not cachable, we will just return the operation, without caching
|
||||||
if reduce(lambda a, b: a or (not (isinstance(b, Observable) or b is None)), inputs, False):
|
if reduce(lambda a, b: a or (not (isinstance(b, Observable) or b is None)), inputs, False):
|
||||||
self.logger.info("some inputs are not observable: returning without caching")
|
|
||||||
self.logger.debug(str(map(lambda x: isinstance(x, Observable) or x is None, inputs)))
|
|
||||||
self.logger.debug(str(map(repr, inputs)))
|
|
||||||
return self.operation(*args, **kw)
|
return self.operation(*args, **kw)
|
||||||
# 3&4: check whether this cache_id has been cached, then has it changed?
|
# 3&4: check whether this cache_id has been cached, then has it changed?
|
||||||
try:
|
try:
|
||||||
if(self.inputs_changed[cache_id]):
|
if(self.inputs_changed[cache_id]):
|
||||||
self.logger.debug("{} already seen, but inputs changed. refreshing cacher".format(cache_id))
|
|
||||||
# 4: This happens, when elements have changed for this cache self.id
|
# 4: This happens, when elements have changed for this cache self.id
|
||||||
self.inputs_changed[cache_id] = False
|
self.inputs_changed[cache_id] = False
|
||||||
self.cached_outputs[cache_id] = self.operation(*args, **kw)
|
self.cached_outputs[cache_id] = self.operation(*args, **kw)
|
||||||
except KeyError:
|
except KeyError:
|
||||||
self.logger.info("{} never seen, creating cache entry".format(cache_id))
|
|
||||||
# 3: This is when we never saw this chache_id:
|
# 3: This is when we never saw this chache_id:
|
||||||
self.ensure_cache_length(cache_id)
|
self.ensure_cache_length(cache_id)
|
||||||
self.add_to_cache(cache_id, inputs, self.operation(*args, **kw))
|
self.add_to_cache(cache_id, inputs, self.operation(*args, **kw))
|
||||||
except:
|
except:
|
||||||
self.logger.error("an error occurred while trying to run caching for {}, resetting".format(cache_id))
|
|
||||||
self.reset()
|
self.reset()
|
||||||
raise
|
raise
|
||||||
# 5: We have seen this cache_id and it is cached:
|
# 5: We have seen this cache_id and it is cached:
|
||||||
self.logger.info("returning cache {}".format(cache_id))
|
|
||||||
return self.cached_outputs[cache_id]
|
return self.cached_outputs[cache_id]
|
||||||
|
|
||||||
def on_cache_changed(self, direct, which=None):
|
def on_cache_changed(self, direct, which=None):
|
||||||
|
|
@ -143,7 +127,6 @@ class Cacher(object):
|
||||||
ind_id = self.id(what)
|
ind_id = self.id(what)
|
||||||
_, cache_ids = self.cached_input_ids.get(ind_id, [None, []])
|
_, cache_ids = self.cached_input_ids.get(ind_id, [None, []])
|
||||||
for cache_id in cache_ids:
|
for cache_id in cache_ids:
|
||||||
self.logger.info("callback from {} changed inputs from {}".format(ind_id, self.inputs_changed[cache_id]))
|
|
||||||
self.inputs_changed[cache_id] = True
|
self.inputs_changed[cache_id] = True
|
||||||
|
|
||||||
def reset(self):
|
def reset(self):
|
||||||
|
|
|
||||||
|
|
@ -385,7 +385,7 @@ def spellman_yeast(data_set='spellman_yeast'):
|
||||||
Y = read_csv(filename, header=0, index_col=0, sep='\t')
|
Y = read_csv(filename, header=0, index_col=0, sep='\t')
|
||||||
return data_details_return({'Y': Y}, data_set)
|
return data_details_return({'Y': Y}, data_set)
|
||||||
|
|
||||||
def spellman_yeast_cdc(data_set='spellman_yeast'):
|
def spellman_yeast_cdc15(data_set='spellman_yeast'):
|
||||||
if not data_available(data_set):
|
if not data_available(data_set):
|
||||||
download_data(data_set)
|
download_data(data_set)
|
||||||
from pandas import read_csv
|
from pandas import read_csv
|
||||||
|
|
@ -405,11 +405,12 @@ def lee_yeast_ChIP(data_set='lee_yeast_ChIP'):
|
||||||
import zipfile
|
import zipfile
|
||||||
dir_path = os.path.join(data_path, data_set)
|
dir_path = os.path.join(data_path, data_set)
|
||||||
filename = os.path.join(dir_path, 'binding_by_gene.tsv')
|
filename = os.path.join(dir_path, 'binding_by_gene.tsv')
|
||||||
X = read_csv(filename, header=1, index_col=0, sep='\t')
|
S = read_csv(filename, header=1, index_col=0, sep='\t')
|
||||||
transcription_factors = [col for col in X.columns if col[:7] != 'Unnamed']
|
transcription_factors = [col for col in S.columns if col[:7] != 'Unnamed']
|
||||||
annotations = X[['Unnamed: 1', 'Unnamed: 2', 'Unnamed: 3']]
|
annotations = S[['Unnamed: 1', 'Unnamed: 2', 'Unnamed: 3']]
|
||||||
X = X[transcription_factors]
|
S = S[transcription_factors]
|
||||||
return data_details_return({'annotations' : annotations, 'X' : X, 'transcription_factors': transcription_factors}, data_set)
|
return data_details_return({'annotations' : annotations, 'Y' : S, 'transcription_factors': transcription_factors}, data_set)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def fruitfly_tomancak(data_set='fruitfly_tomancak', gene_number=None):
|
def fruitfly_tomancak(data_set='fruitfly_tomancak', gene_number=None):
|
||||||
|
|
@ -971,7 +972,8 @@ def olivetti_faces(data_set='olivetti_faces'):
|
||||||
for subject in range(40):
|
for subject in range(40):
|
||||||
for image in range(10):
|
for image in range(10):
|
||||||
image_path = os.path.join(path, 'orl_faces', 's'+str(subject+1), str(image+1) + '.pgm')
|
image_path = os.path.join(path, 'orl_faces', 's'+str(subject+1), str(image+1) + '.pgm')
|
||||||
Y.append(GPy.util.netpbmfile.imread(image_path).flatten())
|
from GPy.util import netpbmfile
|
||||||
|
Y.append(netpbmfile.imread(image_path).flatten())
|
||||||
lbls.append(subject)
|
lbls.append(subject)
|
||||||
Y = np.asarray(Y)
|
Y = np.asarray(Y)
|
||||||
lbls = np.asarray(lbls)[:, None]
|
lbls = np.asarray(lbls)[:, None]
|
||||||
|
|
|
||||||
|
|
@ -18,6 +18,6 @@ def initialize_latent(init, input_dim, Y):
|
||||||
var = Xr.var(0)
|
var = Xr.var(0)
|
||||||
|
|
||||||
Xr -= Xr.mean(0)
|
Xr -= Xr.mean(0)
|
||||||
Xr /= Xr.var(0)
|
Xr /= Xr.std(0)
|
||||||
|
|
||||||
return Xr, var/var.max()
|
return Xr, var/var.max()
|
||||||
|
|
|
||||||
|
|
@ -16,13 +16,17 @@ import warnings
|
||||||
import os
|
import os
|
||||||
from config import *
|
from config import *
|
||||||
|
|
||||||
if np.all(np.float64((scipy.__version__).split('.')[:2]) >= np.array([0, 12])):
|
_scipyversion = np.float64((scipy.__version__).split('.')[:2])
|
||||||
|
_fix_dpotri_scipy_bug = True
|
||||||
|
if np.all(_scipyversion >= np.array([0, 14])):
|
||||||
|
from scipy.linalg import lapack
|
||||||
|
_fix_dpotri_scipy_bug = False
|
||||||
|
elif np.all(_scipyversion >= np.array([0, 12])):
|
||||||
#import scipy.linalg.lapack.clapack as lapack
|
#import scipy.linalg.lapack.clapack as lapack
|
||||||
from scipy.linalg import lapack
|
from scipy.linalg import lapack
|
||||||
else:
|
else:
|
||||||
from scipy.linalg.lapack import flapack as lapack
|
from scipy.linalg.lapack import flapack as lapack
|
||||||
|
|
||||||
|
|
||||||
if config.getboolean('anaconda', 'installed') and config.getboolean('anaconda', 'MKL'):
|
if config.getboolean('anaconda', 'installed') and config.getboolean('anaconda', 'MKL'):
|
||||||
try:
|
try:
|
||||||
anaconda_path = str(config.get('anaconda', 'location'))
|
anaconda_path = str(config.get('anaconda', 'location'))
|
||||||
|
|
@ -30,6 +34,7 @@ if config.getboolean('anaconda', 'installed') and config.getboolean('anaconda',
|
||||||
dsyrk = mkl_rt.dsyrk
|
dsyrk = mkl_rt.dsyrk
|
||||||
dsyr = mkl_rt.dsyr
|
dsyr = mkl_rt.dsyr
|
||||||
_blas_available = True
|
_blas_available = True
|
||||||
|
print 'anaconda installed and mkl is loaded'
|
||||||
except:
|
except:
|
||||||
_blas_available = False
|
_blas_available = False
|
||||||
else:
|
else:
|
||||||
|
|
@ -142,15 +147,22 @@ def dpotri(A, lower=1):
|
||||||
"""
|
"""
|
||||||
Wrapper for lapack dpotri function
|
Wrapper for lapack dpotri function
|
||||||
|
|
||||||
|
DPOTRI - compute the inverse of a real symmetric positive
|
||||||
|
definite matrix A using the Cholesky factorization A =
|
||||||
|
U**T*U or A = L*L**T computed by DPOTRF
|
||||||
|
|
||||||
:param A: Matrix A
|
:param A: Matrix A
|
||||||
:param lower: is matrix lower (true) or upper (false)
|
:param lower: is matrix lower (true) or upper (false)
|
||||||
:returns: A inverse
|
:returns: A inverse
|
||||||
|
|
||||||
"""
|
"""
|
||||||
assert lower==1, "scipy linalg behaviour is very weird. please use lower, fortran ordered arrays"
|
if _fix_dpotri_scipy_bug:
|
||||||
|
assert lower==1, "scipy linalg behaviour is very weird. please use lower, fortran ordered arrays"
|
||||||
|
lower = 0
|
||||||
|
|
||||||
A = force_F_ordered(A)
|
A = force_F_ordered(A)
|
||||||
R, info = lapack.dpotri(A, lower=0)
|
R, info = lapack.dpotri(A, lower=lower) #needs to be zero here, seems to be a scipy bug
|
||||||
|
|
||||||
symmetrify(R)
|
symmetrify(R)
|
||||||
return R, info
|
return R, info
|
||||||
|
|
||||||
|
|
@ -217,7 +229,7 @@ def pdinv(A, *args):
|
||||||
L = jitchol(A, *args)
|
L = jitchol(A, *args)
|
||||||
logdet = 2.*np.sum(np.log(np.diag(L)))
|
logdet = 2.*np.sum(np.log(np.diag(L)))
|
||||||
Li = dtrtri(L)
|
Li = dtrtri(L)
|
||||||
Ai, _ = lapack.dpotri(L)
|
Ai, _ = dpotri(L, lower=1)
|
||||||
# Ai = np.tril(Ai) + np.tril(Ai,-1).T
|
# Ai = np.tril(Ai) + np.tril(Ai,-1).T
|
||||||
symmetrify(Ai)
|
symmetrify(Ai)
|
||||||
|
|
||||||
|
|
|
||||||
331
GPy/util/netpbmfile.py
Normal file
331
GPy/util/netpbmfile.py
Normal file
|
|
@ -0,0 +1,331 @@
|
||||||
|
#!/usr/bin/env python
|
||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
# netpbmfile.py
|
||||||
|
|
||||||
|
# Copyright (c) 2011-2013, Christoph Gohlke
|
||||||
|
# Copyright (c) 2011-2013, The Regents of the University of California
|
||||||
|
# Produced at the Laboratory for Fluorescence Dynamics.
|
||||||
|
# All rights reserved.
|
||||||
|
#
|
||||||
|
# Redistribution and use in source and binary forms, with or without
|
||||||
|
# modification, are permitted provided that the following conditions are met:
|
||||||
|
#
|
||||||
|
# * Redistributions of source code must retain the above copyright
|
||||||
|
# notice, this list of conditions and the following disclaimer.
|
||||||
|
# * Redistributions in binary form must reproduce the above copyright
|
||||||
|
# notice, this list of conditions and the following disclaimer in the
|
||||||
|
# documentation and/or other materials provided with the distribution.
|
||||||
|
# * Neither the name of the copyright holders nor the names of any
|
||||||
|
# contributors may be used to endorse or promote products derived
|
||||||
|
# from this software without specific prior written permission.
|
||||||
|
#
|
||||||
|
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
||||||
|
# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||||
|
# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
||||||
|
# ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
|
||||||
|
# LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
||||||
|
# CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
||||||
|
# SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
||||||
|
# INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
||||||
|
# CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
||||||
|
# ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
||||||
|
# POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
|
||||||
|
"""Read and write image data from respectively to Netpbm files.
|
||||||
|
|
||||||
|
This implementation follows the Netpbm format specifications at
|
||||||
|
http://netpbm.sourceforge.net/doc/. No gamma correction is performed.
|
||||||
|
|
||||||
|
The following image formats are supported: PBM (bi-level), PGM (grayscale),
|
||||||
|
PPM (color), PAM (arbitrary), XV thumbnail (RGB332, read-only).
|
||||||
|
|
||||||
|
:Author:
|
||||||
|
`Christoph Gohlke <http://www.lfd.uci.edu/~gohlke/>`_
|
||||||
|
|
||||||
|
:Organization:
|
||||||
|
Laboratory for Fluorescence Dynamics, University of California, Irvine
|
||||||
|
|
||||||
|
:Version: 2013.01.18
|
||||||
|
|
||||||
|
Requirements
|
||||||
|
------------
|
||||||
|
* `CPython 2.7, 3.2 or 3.3 <http://www.python.org>`_
|
||||||
|
* `Numpy 1.7 <http://www.numpy.org>`_
|
||||||
|
* `Matplotlib 1.2 <http://www.matplotlib.org>`_ (optional for plotting)
|
||||||
|
|
||||||
|
Examples
|
||||||
|
--------
|
||||||
|
>>> im1 = numpy.array([[0, 1],[65534, 65535]], dtype=numpy.uint16)
|
||||||
|
>>> imsave('_tmp.pgm', im1)
|
||||||
|
>>> im2 = imread('_tmp.pgm')
|
||||||
|
>>> assert numpy.all(im1 == im2)
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from __future__ import division, print_function
|
||||||
|
|
||||||
|
import sys
|
||||||
|
import re
|
||||||
|
import math
|
||||||
|
from copy import deepcopy
|
||||||
|
|
||||||
|
import numpy
|
||||||
|
|
||||||
|
__version__ = '2013.01.18'
|
||||||
|
__docformat__ = 'restructuredtext en'
|
||||||
|
__all__ = ['imread', 'imsave', 'NetpbmFile']
|
||||||
|
|
||||||
|
|
||||||
|
def imread(filename, *args, **kwargs):
|
||||||
|
"""Return image data from Netpbm file as numpy array.
|
||||||
|
|
||||||
|
`args` and `kwargs` are arguments to NetpbmFile.asarray().
|
||||||
|
|
||||||
|
Examples
|
||||||
|
--------
|
||||||
|
>>> image = imread('_tmp.pgm')
|
||||||
|
|
||||||
|
"""
|
||||||
|
try:
|
||||||
|
netpbm = NetpbmFile(filename)
|
||||||
|
image = netpbm.asarray()
|
||||||
|
finally:
|
||||||
|
netpbm.close()
|
||||||
|
return image
|
||||||
|
|
||||||
|
|
||||||
|
def imsave(filename, data, maxval=None, pam=False):
|
||||||
|
"""Write image data to Netpbm file.
|
||||||
|
|
||||||
|
Examples
|
||||||
|
--------
|
||||||
|
>>> image = numpy.array([[0, 1],[65534, 65535]], dtype=numpy.uint16)
|
||||||
|
>>> imsave('_tmp.pgm', image)
|
||||||
|
|
||||||
|
"""
|
||||||
|
try:
|
||||||
|
netpbm = NetpbmFile(data, maxval=maxval)
|
||||||
|
netpbm.write(filename, pam=pam)
|
||||||
|
finally:
|
||||||
|
netpbm.close()
|
||||||
|
|
||||||
|
|
||||||
|
class NetpbmFile(object):
|
||||||
|
"""Read and write Netpbm PAM, PBM, PGM, PPM, files."""
|
||||||
|
|
||||||
|
_types = {b'P1': b'BLACKANDWHITE', b'P2': b'GRAYSCALE', b'P3': b'RGB',
|
||||||
|
b'P4': b'BLACKANDWHITE', b'P5': b'GRAYSCALE', b'P6': b'RGB',
|
||||||
|
b'P7 332': b'RGB', b'P7': b'RGB_ALPHA'}
|
||||||
|
|
||||||
|
def __init__(self, arg=None, **kwargs):
|
||||||
|
"""Initialize instance from filename, open file, or numpy array."""
|
||||||
|
for attr in ('header', 'magicnum', 'width', 'height', 'maxval',
|
||||||
|
'depth', 'tupltypes', '_filename', '_fh', '_data'):
|
||||||
|
setattr(self, attr, None)
|
||||||
|
if arg is None:
|
||||||
|
self._fromdata([], **kwargs)
|
||||||
|
elif isinstance(arg, basestring):
|
||||||
|
self._fh = open(arg, 'rb')
|
||||||
|
self._filename = arg
|
||||||
|
self._fromfile(self._fh, **kwargs)
|
||||||
|
elif hasattr(arg, 'seek'):
|
||||||
|
self._fromfile(arg, **kwargs)
|
||||||
|
self._fh = arg
|
||||||
|
else:
|
||||||
|
self._fromdata(arg, **kwargs)
|
||||||
|
|
||||||
|
def asarray(self, copy=True, cache=False, **kwargs):
|
||||||
|
"""Return image data from file as numpy array."""
|
||||||
|
data = self._data
|
||||||
|
if data is None:
|
||||||
|
data = self._read_data(self._fh, **kwargs)
|
||||||
|
if cache:
|
||||||
|
self._data = data
|
||||||
|
else:
|
||||||
|
return data
|
||||||
|
return deepcopy(data) if copy else data
|
||||||
|
|
||||||
|
def write(self, arg, **kwargs):
|
||||||
|
"""Write instance to file."""
|
||||||
|
if hasattr(arg, 'seek'):
|
||||||
|
self._tofile(arg, **kwargs)
|
||||||
|
else:
|
||||||
|
with open(arg, 'wb') as fid:
|
||||||
|
self._tofile(fid, **kwargs)
|
||||||
|
|
||||||
|
def close(self):
|
||||||
|
"""Close open file. Future asarray calls might fail."""
|
||||||
|
if self._filename and self._fh:
|
||||||
|
self._fh.close()
|
||||||
|
self._fh = None
|
||||||
|
|
||||||
|
def __del__(self):
|
||||||
|
self.close()
|
||||||
|
|
||||||
|
def _fromfile(self, fh):
|
||||||
|
"""Initialize instance from open file."""
|
||||||
|
fh.seek(0)
|
||||||
|
data = fh.read(4096)
|
||||||
|
if (len(data) < 7) or not (b'0' < data[1:2] < b'8'):
|
||||||
|
raise ValueError("Not a Netpbm file:\n%s" % data[:32])
|
||||||
|
try:
|
||||||
|
self._read_pam_header(data)
|
||||||
|
except Exception:
|
||||||
|
try:
|
||||||
|
self._read_pnm_header(data)
|
||||||
|
except Exception:
|
||||||
|
raise ValueError("Not a Netpbm file:\n%s" % data[:32])
|
||||||
|
|
||||||
|
def _read_pam_header(self, data):
|
||||||
|
"""Read PAM header and initialize instance."""
|
||||||
|
regroups = re.search(
|
||||||
|
b"(^P7[\n\r]+(?:(?:[\n\r]+)|(?:#.*)|"
|
||||||
|
b"(HEIGHT\s+\d+)|(WIDTH\s+\d+)|(DEPTH\s+\d+)|(MAXVAL\s+\d+)|"
|
||||||
|
b"(?:TUPLTYPE\s+\w+))*ENDHDR\n)", data).groups()
|
||||||
|
self.header = regroups[0]
|
||||||
|
self.magicnum = b'P7'
|
||||||
|
for group in regroups[1:]:
|
||||||
|
key, value = group.split()
|
||||||
|
setattr(self, unicode(key).lower(), int(value))
|
||||||
|
matches = re.findall(b"(TUPLTYPE\s+\w+)", self.header)
|
||||||
|
self.tupltypes = [s.split(None, 1)[1] for s in matches]
|
||||||
|
|
||||||
|
def _read_pnm_header(self, data):
|
||||||
|
"""Read PNM header and initialize instance."""
|
||||||
|
bpm = data[1:2] in b"14"
|
||||||
|
regroups = re.search(b"".join((
|
||||||
|
b"(^(P[123456]|P7 332)\s+(?:#.*[\r\n])*",
|
||||||
|
b"\s*(\d+)\s+(?:#.*[\r\n])*",
|
||||||
|
b"\s*(\d+)\s+(?:#.*[\r\n])*" * (not bpm),
|
||||||
|
b"\s*(\d+)\s(?:\s*#.*[\r\n]\s)*)")), data).groups() + (1, ) * bpm
|
||||||
|
self.header = regroups[0]
|
||||||
|
self.magicnum = regroups[1]
|
||||||
|
self.width = int(regroups[2])
|
||||||
|
self.height = int(regroups[3])
|
||||||
|
self.maxval = int(regroups[4])
|
||||||
|
self.depth = 3 if self.magicnum in b"P3P6P7 332" else 1
|
||||||
|
self.tupltypes = [self._types[self.magicnum]]
|
||||||
|
|
||||||
|
def _read_data(self, fh, byteorder='>'):
|
||||||
|
"""Return image data from open file as numpy array."""
|
||||||
|
fh.seek(len(self.header))
|
||||||
|
data = fh.read()
|
||||||
|
dtype = 'u1' if self.maxval < 256 else byteorder + 'u2'
|
||||||
|
depth = 1 if self.magicnum == b"P7 332" else self.depth
|
||||||
|
shape = [-1, self.height, self.width, depth]
|
||||||
|
size = numpy.prod(shape[1:])
|
||||||
|
if self.magicnum in b"P1P2P3":
|
||||||
|
data = numpy.array(data.split(None, size)[:size], dtype)
|
||||||
|
data = data.reshape(shape)
|
||||||
|
elif self.maxval == 1:
|
||||||
|
shape[2] = int(math.ceil(self.width / 8))
|
||||||
|
data = numpy.frombuffer(data, dtype).reshape(shape)
|
||||||
|
data = numpy.unpackbits(data, axis=-2)[:, :, :self.width, :]
|
||||||
|
else:
|
||||||
|
data = numpy.frombuffer(data, dtype)
|
||||||
|
data = data[:size * (data.size // size)].reshape(shape)
|
||||||
|
if data.shape[0] < 2:
|
||||||
|
data = data.reshape(data.shape[1:])
|
||||||
|
if data.shape[-1] < 2:
|
||||||
|
data = data.reshape(data.shape[:-1])
|
||||||
|
if self.magicnum == b"P7 332":
|
||||||
|
rgb332 = numpy.array(list(numpy.ndindex(8, 8, 4)), numpy.uint8)
|
||||||
|
rgb332 *= [36, 36, 85]
|
||||||
|
data = numpy.take(rgb332, data, axis=0)
|
||||||
|
return data
|
||||||
|
|
||||||
|
def _fromdata(self, data, maxval=None):
|
||||||
|
"""Initialize instance from numpy array."""
|
||||||
|
data = numpy.array(data, ndmin=2, copy=True)
|
||||||
|
if data.dtype.kind not in "uib":
|
||||||
|
raise ValueError("not an integer type: %s" % data.dtype)
|
||||||
|
if data.dtype.kind == 'i' and numpy.min(data) < 0:
|
||||||
|
raise ValueError("data out of range: %i" % numpy.min(data))
|
||||||
|
if maxval is None:
|
||||||
|
maxval = numpy.max(data)
|
||||||
|
maxval = 255 if maxval < 256 else 65535
|
||||||
|
if maxval < 0 or maxval > 65535:
|
||||||
|
raise ValueError("data out of range: %i" % maxval)
|
||||||
|
data = data.astype('u1' if maxval < 256 else '>u2')
|
||||||
|
self._data = data
|
||||||
|
if data.ndim > 2 and data.shape[-1] in (3, 4):
|
||||||
|
self.depth = data.shape[-1]
|
||||||
|
self.width = data.shape[-2]
|
||||||
|
self.height = data.shape[-3]
|
||||||
|
self.magicnum = b'P7' if self.depth == 4 else b'P6'
|
||||||
|
else:
|
||||||
|
self.depth = 1
|
||||||
|
self.width = data.shape[-1]
|
||||||
|
self.height = data.shape[-2]
|
||||||
|
self.magicnum = b'P5' if maxval > 1 else b'P4'
|
||||||
|
self.maxval = maxval
|
||||||
|
self.tupltypes = [self._types[self.magicnum]]
|
||||||
|
self.header = self._header()
|
||||||
|
|
||||||
|
def _tofile(self, fh, pam=False):
|
||||||
|
"""Write Netbm file."""
|
||||||
|
fh.seek(0)
|
||||||
|
fh.write(self._header(pam))
|
||||||
|
data = self.asarray(copy=False)
|
||||||
|
if self.maxval == 1:
|
||||||
|
data = numpy.packbits(data, axis=-1)
|
||||||
|
data.tofile(fh)
|
||||||
|
|
||||||
|
def _header(self, pam=False):
|
||||||
|
"""Return file header as byte string."""
|
||||||
|
if pam or self.magicnum == b'P7':
|
||||||
|
header = "\n".join((
|
||||||
|
"P7",
|
||||||
|
"HEIGHT %i" % self.height,
|
||||||
|
"WIDTH %i" % self.width,
|
||||||
|
"DEPTH %i" % self.depth,
|
||||||
|
"MAXVAL %i" % self.maxval,
|
||||||
|
"\n".join("TUPLTYPE %s" % unicode(i) for i in self.tupltypes),
|
||||||
|
"ENDHDR\n"))
|
||||||
|
elif self.maxval == 1:
|
||||||
|
header = "P4 %i %i\n" % (self.width, self.height)
|
||||||
|
elif self.depth == 1:
|
||||||
|
header = "P5 %i %i %i\n" % (self.width, self.height, self.maxval)
|
||||||
|
else:
|
||||||
|
header = "P6 %i %i %i\n" % (self.width, self.height, self.maxval)
|
||||||
|
if sys.version_info[0] > 2:
|
||||||
|
header = bytes(header, 'ascii')
|
||||||
|
return header
|
||||||
|
|
||||||
|
def __str__(self):
|
||||||
|
"""Return information about instance."""
|
||||||
|
return unicode(self.header)
|
||||||
|
|
||||||
|
|
||||||
|
if sys.version_info[0] > 2:
|
||||||
|
basestring = str
|
||||||
|
unicode = lambda x: str(x, 'ascii')
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
# Show images specified on command line or all images in current directory
|
||||||
|
from glob import glob
|
||||||
|
from matplotlib import pyplot
|
||||||
|
files = sys.argv[1:] if len(sys.argv) > 1 else glob('*.p*m')
|
||||||
|
for fname in files:
|
||||||
|
try:
|
||||||
|
pam = NetpbmFile(fname)
|
||||||
|
img = pam.asarray(copy=False)
|
||||||
|
if False:
|
||||||
|
pam.write('_tmp.pgm.out', pam=True)
|
||||||
|
img2 = imread('_tmp.pgm.out')
|
||||||
|
assert numpy.all(img == img2)
|
||||||
|
imsave('_tmp.pgm.out', img)
|
||||||
|
img2 = imread('_tmp.pgm.out')
|
||||||
|
assert numpy.all(img == img2)
|
||||||
|
pam.close()
|
||||||
|
except ValueError as e:
|
||||||
|
print(fname, e)
|
||||||
|
continue
|
||||||
|
_shape = img.shape
|
||||||
|
if img.ndim > 3 or (img.ndim > 2 and img.shape[-1] not in (3, 4)):
|
||||||
|
img = img[0]
|
||||||
|
cmap = 'gray' if pam.maxval > 1 else 'binary'
|
||||||
|
pyplot.imshow(img, cmap, interpolation='nearest')
|
||||||
|
pyplot.title("%s %s %s %s" % (fname, unicode(pam.magicnum),
|
||||||
|
_shape, img.dtype))
|
||||||
|
pyplot.show()
|
||||||
|
|
@ -48,14 +48,10 @@ def common_subarrays(X, axis=0):
|
||||||
assert X.ndim == 2 and axis in (0,1), "Only implemented for 2D arrays"
|
assert X.ndim == 2 and axis in (0,1), "Only implemented for 2D arrays"
|
||||||
subarrays = defaultdict(list)
|
subarrays = defaultdict(list)
|
||||||
cnt = count()
|
cnt = count()
|
||||||
logger = logging.getLogger("common_subarrays")
|
|
||||||
def accumulate(x, s, c):
|
def accumulate(x, s, c):
|
||||||
logger.debug("creating tuple")
|
|
||||||
t = tuple(x)
|
t = tuple(x)
|
||||||
logger.debug("tuple done")
|
|
||||||
col = c.next()
|
col = c.next()
|
||||||
iadd(s[t], [col])
|
iadd(s[t], [col])
|
||||||
logger.debug("added col {}".format(col))
|
|
||||||
return None
|
return None
|
||||||
if axis == 0: [accumulate(x, subarrays, cnt) for x in X]
|
if axis == 0: [accumulate(x, subarrays, cnt) for x in X]
|
||||||
else: [accumulate(x, subarrays, cnt) for x in X.T]
|
else: [accumulate(x, subarrays, cnt) for x in X.T]
|
||||||
|
|
|
||||||
|
|
@ -40,6 +40,37 @@ def std_norm_cdf(x):
|
||||||
weave.inline(code, arg_names=['x', 'cdf_x', 'N'], support_code=support_code)
|
weave.inline(code, arg_names=['x', 'cdf_x', 'N'], support_code=support_code)
|
||||||
return cdf_x
|
return cdf_x
|
||||||
|
|
||||||
|
def std_norm_cdf_np(x):
|
||||||
|
"""
|
||||||
|
Cumulative standard Gaussian distribution
|
||||||
|
Based on Abramowitz, M. and Stegun, I. (1970)
|
||||||
|
Around 3 times slower when x is a scalar otherwise quite a lot slower
|
||||||
|
"""
|
||||||
|
x_shape = np.asarray(x).shape
|
||||||
|
|
||||||
|
if len(x_shape) == 0 or x_shape[0] == 1:
|
||||||
|
sign = np.sign(x)
|
||||||
|
x *= sign
|
||||||
|
x /= np.sqrt(2.)
|
||||||
|
t = 1.0/(1.0 + 0.3275911*x)
|
||||||
|
erf = 1. - np.exp(-x**2)*t*(0.254829592 + t*(-0.284496736 + t*(1.421413741 + t*(-1.453152027 + t*(1.061405429)))))
|
||||||
|
cdf_x = 0.5*(1.0 + sign*erf)
|
||||||
|
return cdf_x
|
||||||
|
else:
|
||||||
|
x = np.atleast_1d(x).copy()
|
||||||
|
cdf_x = np.zeros_like(x)
|
||||||
|
sign = np.ones_like(x)
|
||||||
|
neg_x_ind = x<0
|
||||||
|
sign[neg_x_ind] = -1.0
|
||||||
|
x[neg_x_ind] = -x[neg_x_ind]
|
||||||
|
x /= np.sqrt(2.)
|
||||||
|
t = 1.0/(1.0 + 0.3275911*x)
|
||||||
|
erf = 1. - np.exp(-x**2)*t*(0.254829592 + t*(-0.284496736 + t*(1.421413741 + t*(-1.453152027 + t*(1.061405429)))))
|
||||||
|
cdf_x = 0.5*(1.0 + sign*erf)
|
||||||
|
cdf_x = cdf_x.reshape(x_shape)
|
||||||
|
return cdf_x
|
||||||
|
|
||||||
|
|
||||||
def inv_std_norm_cdf(x):
|
def inv_std_norm_cdf(x):
|
||||||
"""
|
"""
|
||||||
Inverse cumulative standard Gaussian distribution
|
Inverse cumulative standard Gaussian distribution
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue