biiig changes in tieing, and printing -> hirarchy now always shown

This commit is contained in:
Max Zwiessele 2013-11-11 17:46:33 +00:00
parent ed6a45fd9b
commit 6c60eef4bd
10 changed files with 153 additions and 88 deletions

View file

@ -22,7 +22,7 @@ class GP(GPBase):
"""
def __init__(self, X, likelihood, kernel, normalize_X=False):
GPBase.__init__(self, X, likelihood, kernel, normalize_X=normalize_X)
super(GP, self).__init__(X, likelihood, kernel, normalize_X=normalize_X)
#self._set_params(self._get_params())
def getstate(self):

View file

@ -13,6 +13,7 @@ from domains import _POSITIVE, _REAL
from numpy.linalg.linalg import LinAlgError
from index_operations import ParameterIndexOperations
import itertools
from GPy.core.parameterized import UNFIXED
# import numdifftools as ndt
class Model(Parameterized):
@ -277,8 +278,17 @@ class Model(Parameterized):
# param_names = self._get_param_names()
for s in positive_strings:
paramlist = self.grep_param_names(".*"+s)
if paramlist:
self.__getitem__(None, paramlist).constrain_positive(warning=warning)
for param in paramlist:
for p in param.flattened_parameters:
rav_i = set(self._raveled_index_for(p))
for constraint in self.constraints.iter_properties():
rav_i -= set(self._constraint_indices(p, constraint))
rav_i -= set(np.nonzero(self._fixes_for(p)!=UNFIXED)[0])
ind = self._backtranslate_index(p, np.array(list(rav_i), dtype=int))
if ind.size != 0:
p[np.unravel_index(ind, p.shape)].constrain_positive()
# if paramlist:
# self.__getitem__(None, paramlist).constrain_positive(warning=warning)
# currently_constrained = self.all_constrained_indices()
# to_make_positive = []
# for s in positive_strings:

View file

@ -20,6 +20,8 @@ __print_threshold__ = 5
class ListArray(numpy.ndarray):
"""
ndarray which can be stored in lists and checked if it is in.
WARNING: This overrides the functionality of x==y!!!
Use numpy.equal(x,y) for element-wise equality testing.
"""
def __new__(cls, input_array):
obj = numpy.asanyarray(input_array).view(cls)
@ -43,10 +45,16 @@ class ObservableArray(ListArray, Observable):
if obj is None: return
self._observers_ = getattr(obj, '_observers_', None)
def __setitem__(self, s, val, update=True):
if not numpy.all(numpy.equal(self[s], val)):
super(ObservableArray, self).__setitem__(s, val)
if update:
self._notify_observers()
if self.ndim:
if not numpy.all(numpy.equal(self[s], val)):
super(ObservableArray, self).__setitem__(s, val)
if update:
self._notify_observers()
else:
if not numpy.all(numpy.equal(self, val)):
super(ObservableArray, self).__setitem__(Ellipsis, val)
if update:
self._notify_observers()
def __getslice__(self, start, stop):
return self.__getitem__(slice(start, stop))
def __setslice__(self, start, stop, val):
@ -77,6 +85,7 @@ class Param(ObservableArray, Nameable, Pickleable):
See :py:class:`GPy.core.parameterized.Parameterized` for more details.
"""
__array_priority__ = -numpy.inf # Never give back Param
_fixes_ = None
def __new__(cls, name, input_array, *args, **kwargs):
obj = numpy.atleast_1d(super(Param, cls).__new__(cls, input_array=input_array))
obj._direct_parent_ = None
@ -289,9 +298,9 @@ class Param(ObservableArray, Nameable, Pickleable):
"""
assert isinstance(param, Param), "Argument {1} not of type {0}".format(Param,param.__class__)
try:
if self._original_: # this happens when indexing created a copy of the array
if self._original_:
self[:] = param
else:
else: # this happens when indexing created a copy of the array
self._direct_parent_._get_original(self)[self._current_slice_] = param
except ValueError:
raise ValueError("Trying to tie {} with shape {} to {} with shape {}".format(self.name, self.shape, param.name, param.shape))
@ -304,8 +313,9 @@ class Param(ObservableArray, Nameable, Pickleable):
param._add_tie_listener(self)
self._highest_parent_._set_fixed(self)
for t in self._tied_to_me_.iterkeys():
t.untie()
t.tie_to(param)
if t is not self:
t.untie(self)
t.tie_to(param)
# self._direct_parent_._add_tie(self, param)
def untie(self, *ties):
@ -433,6 +443,8 @@ class Param(ObservableArray, Nameable, Pickleable):
view = super(Param, self).round(decimals, out).view(Param)
view.__array_finalize__(self)
return view
def _has_fixes(self):
return False
round.__doc__ = numpy.round.__doc__
def _get_original(self, param):
return self
@ -494,7 +506,7 @@ class Param(ObservableArray, Nameable, Pickleable):
def _max_len_names(self, gen, header):
return reduce(lambda a, b:max(a, len(b)), gen, len(header))
def _max_len_values(self):
return reduce(lambda a, b:max(a, len("{x:=.{0}g}".format(__precision__, x=b))), self.flat, len(self.name))
return reduce(lambda a, b:max(a, len("{x:=.{0}g}".format(__precision__, x=b))), self.flat, len(self.name_hirarchical))
def _max_len_index(self, ind):
return reduce(lambda a, b:max(a, len(str(b))), ind, len(__index_name__))
def _short(self):
@ -518,7 +530,7 @@ class Param(ObservableArray, Nameable, Pickleable):
if lx is None: lx = self._max_len_values()
if li is None: li = self._max_len_index(indices)
if lt is None: lt = self._max_len_names(ties, __tie_name__)
header = " {i:^{2}s} | \033[1m{x:^{1}s}\033[0;0m | {c:^{0}s} | {t:^{3}s}".format(lc,lx,li,lt, x=self.name, c=__constraints_name__, i=__index_name__, t=__tie_name__) # nice header for printing
header = " {i:^{2}s} | \033[1m{x:^{1}s}\033[0;0m | {c:^{0}s} | {t:^{3}s}".format(lc,lx,li,lt, x=self.name_hirarchical, c=__constraints_name__, i=__index_name__, t=__tie_name__) # nice header for printing
if not ties: ties = itertools.cycle([''])
return "\n".join([header]+[" {i!s:^{3}s} | {x: >{1}.{2}g} | {c:^{0}s} | {t:^{4}s} ".format(lc,lx,__precision__,li,lt, x=x, c=" ".join(map(str,c)), t=(t or ''), i=i) for i,x,c,t in itertools.izip(indices,vals,constr_matrix,ties)]) # return all the constraints with right indices
#except: return super(Param, self).__str__()

View file

@ -37,7 +37,9 @@ class Pickleable(object):
"""
Returns the state of this class in a memento pattern.
The state must be a list-like structure of all the fields
this class need to run
this class needs to run.
See python doc "pickling" (`__getstate__` and `__setstate__`) for details.
"""
raise NotImplementedError, "To be able to use pickling you need to implement this method"
def setstate(self, state):
@ -47,6 +49,8 @@ class Pickleable(object):
an object is a copy of another when calling
copy = <classname>.__new__(*args,**kw).setstate(<to_be_copied>.getstate())
See python doc "pickling" (`__getstate__` and `__setstate__`) for details.
"""
raise NotImplementedError, "To be able to use pickling you need to implement this method"
@ -61,8 +65,9 @@ class Observable(object):
[callble(self) for callble in self._observers_.itervalues()]
def _adjust_name_for_printing(name):
return name.replace(" ", "_").replace(".", "_")
if name is not None:
return name.replace(" ", "_").replace(".", "_").replace("-","").replace("+","").replace("!","").replace("*","").replace("/","")
return ''
from parameter import ParamConcatenation, Param
from index_operations import ParameterIndexOperations,\
index_empty
@ -126,11 +131,10 @@ class Parameterized(Nameable, Pickleable, Observable):
super(Parameterized, self).__init__(name)
self._in_init_ = True
self._constraints_ = None#ParameterIndexOperations()
self._fixes_ = None
if not hasattr(self, "_parameters_"):
self._parameters_ = []
#else:
# self._parameters_.extend(parameters)
if not self._has_fixes():
self._fixes_ = None
self._connect_parameters()
self.gradient_mapping = {}
self._added_names_ = set()
@ -171,6 +175,10 @@ class Parameterized(Nameable, Pickleable, Observable):
# else:
# self._parameters_.insert(index, Param(name, array, gradient))
# self._connect_parameters(gradient_parent=gradient_parent)
def _has_fixes(self):
return hasattr(self, "_fixes_") and self._fixes_ is not None
def add_parameter(self, parameter, gradient=None, index=None):
"""
:param parameters: the parameters to add
@ -184,16 +192,52 @@ class Parameterized(Nameable, Pickleable, Observable):
at any given index using the :py:func:`list.insert` syntax
"""
if parameter in self._parameters_ and index is not None:
# make sure fixes and constraints are indexed right
if self._has_fixes():
param_slice = slice(self._offset_for(parameter),self._offset_for(parameter)+parameter.size)
dest_index = sum((p.size for p in self._parameters_[:index]))
dest_slice = slice(dest_index,dest_index+parameter.size)
fixes_param = self._fixes_[param_slice].copy()
self._fixes_[param_slice] = self._fixes_[dest_slice]
self._fixes_[dest_slice] = fixes_param
del self._parameters_[parameter._parent_index_]
self._parameters_.insert(index, parameter)
elif parameter not in self._parameters_:
# make sure the size is set
if not hasattr(self, 'size'):
self.size = sum(p.size for p in self._parameters_)
if index is None:
self._parameters_.append(parameter)
# make sure fixes and constraints are indexed right
if parameter._has_fixes(): fixes_param = parameter._fixes_.copy()
else: fixes_param = numpy.ones(parameter.size, dtype=bool)
if self._has_fixes(): self._fixes_ = np.r_[self._fixes_, fixes_param]
elif parameter._has_fixes(): self._fixes_ = np.r_[np.ones(self.size, dtype=bool), fixes_param]
else:
self._parameters_.insert(index, parameter)
self._connect_parameters()
# make sure fixes and constraints are indexed right
if parameter._has_fixes(): fixes_param = parameter._fixes_.copy()
else: fixes_param = numpy.ones(parameter.size, dtype=bool)
ins = sum((p.size for p in self._parameters_[:index]))
if self._has_fixes(): self._fixes_ = np.r_[self._fixes_[:ins], fixes_param, self._fixes[ins:]]
elif not np.all(fixes_param):
self._fixes_ = np.ones(self.size+parameter.size, dtype=bool)
self._fixes_[ins:ins+parameter.size] = fixes_param
self.size += parameter.size
if gradient:
self.gradient_mapping[parameter] = gradient
self._connect_parameters()
# make sure the constraints are pulled over:
if hasattr(parameter, "_constraints_") and parameter._constraints_ is not None:
for t, ind in parameter._constraints_.iteritems():
self.constraints.add(t, ind+self._offset_for(parameter))
parameter._constraints_.clear()
if self._has_fixes() and np.all(self._fixes_): # ==UNFIXED
self._fixes_= None
def add_parameters(self, *parameters):
"""
@ -234,6 +278,9 @@ class Parameterized(Nameable, Pickleable, Observable):
# no parameters for this class
return
i = 0
sizes = [0]
#self.size = sum(p.size for p in self._parameters_)
self._param_slices_ = []
for p in self._parameters_:
#if p._parent_ is None:
p._direct_parent_ = self
@ -242,12 +289,11 @@ class Parameterized(Nameable, Pickleable, Observable):
for pi in p.flattened_parameters:
pi._highest_parent_ = self
not_unique = []
# for k,v in self.__dict__.iteritems():
# try:
# if fast_array_equal(v,p):
# self.__dict__[k] = p
# except: # parameter comparison, just for convenience
# pass
sizes.append(p.size+sizes[-1])
self._param_slices_.append(slice(sizes[-2], sizes[-1]))
# if p._fixes_ is not None:
# self._fixes_[p._raveled_index_for(p)] = p._fixes_
# p._fixes_ = None
pname = _adjust_name_for_printing(p.name)
if pname in self.__dict__:
if isinstance(self.__dict__[pname], (Parameterized, Param)):
@ -257,9 +303,15 @@ class Parameterized(Nameable, Pickleable, Observable):
elif not (pname in not_unique):
self.__dict__[pname] = p
self._added_names_.add(pname)
sizes = numpy.cumsum([0] + self._parameter_sizes_)
self.size = sizes[-1]
self._param_slices_ = [slice(start, stop) for start,stop in zip(sizes, sizes[1:])]
# for p in self._parameters_:
# if hasattr(p, '_constraints_') and p._constraints_ is not None:
# for t, ind in p._constraints_.iteritems():
# self.constraints.add(t, ind+self._offset_for(p))
# p._constraints_.clear()
# if np.all(self._fixes_): # ==UNFIXED
# self._fixes_= None
# else:
# self.constraints.add(__fixed__, np.nonzero(~self._fixes_)[0])
# self.parameters_changed()
#===========================================================================
# Pickling operations
@ -341,14 +393,14 @@ class Parameterized(Nameable, Pickleable, Observable):
# to_remove = np.hstack((self.fixed_indices + [t[1:] for t in self.tied_indices]))
# return np.delete(g, to_remove)
# else:
if self._fixes_ is not None: return g[self._fixes_]
if self._has_fixes(): return g[self._fixes_]
return g
#===========================================================================
# Optimization handles:
#===========================================================================
def _get_param_names_transformed(self):
n = numpy.array([p.name_hirarchical+'['+str(i)+']' for p in self.flattened_parameters for i in p._indices()])
if self._fixes_ is not None:
if self._has_fixes():
return n[self._fixes_]
return n
def _get_params(self):
@ -361,12 +413,12 @@ class Parameterized(Nameable, Pickleable, Observable):
def _get_params_transformed(self):
p = self._get_params()
[numpy.put(p, ind, c.finv(p[ind])) for c,ind in self.constraints.iteritems() if c != __fixed__]
if self._fixes_ is not None:
if self._has_fixes():
return p[self._fixes_]
return p
def _set_params_transformed(self, p):
p = p.copy()
if self._fixes_ is not None: tmp = self._get_params(); tmp[self._fixes_] = p; p = tmp; del tmp
if self._has_fixes(): tmp = self._get_params(); tmp[self._fixes_] = p; p = tmp; del tmp
[numpy.put(p, ind, c.f(p[ind])) for c,ind in self.constraints.iteritems() if c != __fixed__]
self._set_params(p)
def _name_changed(self, param, old_name):
@ -389,19 +441,22 @@ class Parameterized(Nameable, Pickleable, Observable):
return ind
def _offset_for(self, param):
# get the offset in the parameterized index array for param
if param._direct_parent_._get_original(param) in self._parameters_:
return self._param_slices_[param._direct_parent_._get_original(param)._parent_index_].start
if param.has_parent():
if param._direct_parent_._get_original(param) in self._parameters_:
return self._param_slices_[param._direct_parent_._get_original(param)._parent_index_].start
return self._offset_for(param._direct_parent_) + param._direct_parent_._offset_for(param)
return 0
def _raveled_index_for(self, param):
return param._raveled_index() + self._offset_for(param)
def _raveled_index(self):
return numpy.r_[:self.size]
#===========================================================================
# Handle ties:
#===========================================================================
def _set_fixed(self, param_or_index):
if self._fixes_ is None: self._fixes_ = numpy.ones(self.size, dtype=bool)
if not self._has_fixes(): self._fixes_ = numpy.ones(self.size, dtype=bool)
try:
param_or_index = self._raveled_index_for(param_or_index)
except AttributeError:
@ -409,40 +464,20 @@ class Parameterized(Nameable, Pickleable, Observable):
self._fixes_[param_or_index] = FIXED
if numpy.all(self._fixes_): self._fixes_ = None # ==UNFIXED
def _set_unfixed(self, param_or_index):
if self._fixes_ is None: self._fixes_ = numpy.ones(self.size, dtype=bool)
if not self._has_fixes(): self._fixes_ = numpy.ones(self.size, dtype=bool)
try:
param_or_index = self._raveled_index_for(param_or_index)
except AttributeError:
pass
self._fixes_[param_or_index] = UNFIXED
for constr, ind in self.constraints.iteritems():
if constr is __fixed__:
self._fixes_[ind] = FIXED
if numpy.all(self._fixes_): self._fixes_ = None # ==UNFIXED
# def _add_tie(self, param, tied_to):
# # tie param to tie_to, if the values match (with broadcasting)
# self._remove_tie(param) # delete if multiple ties should be allowed
# f, _ = self._fixes_.add(param, tied_to)
# if self._fixes_ is None: self._fixes_ = numpy.ones(self.size, dtype=bool)
# self._fixes_[f] = False
# def _remove_tie(self, param, *params):
# # remove the tie from param to all *params (can be None, so all ties get deleted for param)
# if len(params) == 0:
# params = self._fixes_.properties()
# for p in params:
# _, t = self._fixes_.remove(param, p)
# self._fixes_[t] = True
# if numpy.all(self._fixes_): self._fixes_ = None # ==UNFIXED
# def _ties_iter_items(self, param):
# for tied_to, ind in self._fixes_.iter_from_items():
# ind = self._backtranslate_index(param, ind)
# if not index_empty(ind):
# yield tied_to, ind
# def _ties_iter(self, param):
# for constr, _ in self._ties_iter_items(param):
# yield constr
# def _ties_iter_indices(self, param):
# for _, ind in self._ties_iter_items(param):
# yield ind
# def _ties_for(self, param, rav_index):
# return self._fixes_.from_to_for(rav_index+self._offset_for(param))
def _fixes_for(self, param):
if self._has_fixes():
return self._fixes_[self._raveled_index_for(param)]
return numpy.ones(self.size, dtype=bool)[self._raveled_index_for(param)]
#===========================================================================
# Fixing parameters:
#===========================================================================
@ -450,7 +485,7 @@ class Parameterized(Nameable, Pickleable, Observable):
f = self._add_constrain(param, __fixed__, warning)
self._set_fixed(f)
def _unfix(self, param):
if self._fixes_ is not None:
if self._has_fixes():
f = self._remove_constrain(param, __fixed__)
self._set_unfixed(f)
#===========================================================================
@ -458,9 +493,10 @@ class Parameterized(Nameable, Pickleable, Observable):
#===========================================================================
def _is_fixed(self, param):
# returns if the whole parameter is fixed
if self._fixes_ is None:
if not self._has_fixes():
return False
return not self._fixes_[self._offset_for(param): self._offset_for(param)+param._realsize_].any()
return not self._fixes_[self._raveled_index_for(param)].any()
#return not self._fixes_[self._offset_for(param): self._offset_for(param)+param._realsize_].any()
@property
def is_fixed(self):
for p in self._parameters_:
@ -510,11 +546,14 @@ class Parameterized(Nameable, Pickleable, Observable):
for constr, _ in self._constraints_iter_items(param):
yield constr
def _contraints_iter_indices(self, param):
# iterate through all constraints belonging to param
for _, ind in self._constraints_iter_items(param):
yield ind
def _constraint_indices(self, param, constraint):
return self._backtranslate_index(param, self.constraints[constraint])
# indices in model range for parameter and constraint
return self._backtranslate_index(param, self.constraints[constraint]) + self._offset_for(param)
def _constraints_for(self, param, rav_index):
# constraint for param given its internal rav_index
return self.constraints.properties_for(rav_index+self._offset_for(param))
#===========================================================================
# Get/set parameters:
@ -580,7 +619,7 @@ class Parameterized(Nameable, Pickleable, Observable):
return [x.size for x in self._parameters_]
@property
def size_transformed(self):
if self._fixes_ is not None:
if self._has_fixes():
return sum(self._fixes_)
return self.size
@property
@ -598,7 +637,7 @@ class Parameterized(Nameable, Pickleable, Observable):
def __str__(self, header=True):
constrs = self._constraints_str; ts = self._ties_str
desc = self._description_str; names = self.parameter_names
nl = max([len(str(x)) for x in names + ["Name"]])
nl = max([len(str(x)) for x in names + [_adjust_name_for_printing(self.name)]])
sl = max([len(str(x)) for x in desc + ["Value"]])
cl = max([len(str(x)) if x else 0 for x in constrs + ["Constraint"]])
tl = max([len(str(x)) if x else 0 for x in ts + ["Tied to"]])
@ -609,7 +648,7 @@ class Parameterized(Nameable, Pickleable, Observable):
#to_print = [format_spec.format(p=p, const=c, t=t) if isinstance(p, Param) else p.__str__(header=False) for p, c, t in itertools.izip(self._parameters_, constrs, ts)]
sep = '-'*(nl+sl+cl+tl+8*2+3)
if header:
header = " {{0:<{0}s}} | {{1:^{1}s}} | {{2:^{2}s}} | {{3:^{3}s}}".format(nl, sl, cl, tl).format("Name", "Value", "Constraint", "Tied to")
header = " {{0:<{0}s}} | {{1:^{1}s}} | {{2:^{2}s}} | {{3:^{3}s}}".format(nl, sl, cl, tl).format(_adjust_name_for_printing(self.name), "Value", "Constraint", "Tied to")
#header += '\n' + sep
to_print.insert(0, header)
return '\n'.format(sep).join(to_print)

View file

@ -228,7 +228,7 @@ def Matern52(input_dim, variance=1., lengthscale=None, ARD=False):
part = parts.Matern52.Matern52(input_dim, variance, lengthscale, ARD)
return kern(input_dim, [part])
def bias(input_dim, variance=1.):
def bias(input_dim, variance=1., name='bias'):
"""
Construct a bias kernel.
@ -238,7 +238,7 @@ def bias(input_dim, variance=1.):
:type variance: float
"""
part = parts.bias.Bias(input_dim, variance)
part = parts.bias.Bias(input_dim, variance, name=name)
return kern(input_dim, [part])
def finite_dimensional(input_dim, F, G, variances=1., weights=None):

View file

@ -605,7 +605,7 @@ class Kern_check_model(Model):
self.dL_dK = dL_dK
#self.constrained_indices=[]
#self.constraints=[]
Model.__init__(self)
Model.__init__(self, 'kernel_test_model')
def is_positive_definite(self):
v = np.linalg.eig(self.kernel.K(self.X))[0]
@ -614,14 +614,14 @@ class Kern_check_model(Model):
else:
return True
# def _get_params(self):
# return self.kernel._get_params()
#
# def _get_param_names(self):
# return self.kernel._get_param_names()
#
# def _set_params(self, x):
# self.kernel._set_params(x)
def _get_params(self):
return self.kernel._get_params()
def _get_param_names(self):
return self.kernel._get_param_names()
def _set_params(self, x):
self.kernel._set_params(x)
def log_likelihood(self):
return (self.dL_dK*self.kernel.K(self.X, self.X2)).sum()

View file

@ -8,14 +8,14 @@ import hashlib
from GPy.core.parameter import Param
class Bias(Kernpart):
def __init__(self,input_dim,variance=1.):
def __init__(self,input_dim,variance=1.,name=None):
"""
:param input_dim: the number of input dimensions
:type input_dim: int
:param variance: the variance of the kernel
:type variance: float
"""
super(Bias, self).__init__(input_dim, 'bias')
super(Bias, self).__init__(input_dim, name)
self.variance = Param("variance", variance)
self.add_parameter(self.variance)
#self._set_params(np.array([variance]).flatten())

View file

@ -54,7 +54,7 @@ class RBF(Kernpart):
self.lengthscale = Param('lengthscale', lengthscale)
self.lengthscale.add_observer(self, self.update_lengthscale)
self.add_parameters(self.variance, self.lengthscale)
self.parameters_changed()
self.parameters_changed() # initializes cache
#self.update_inv_lengthscale(self.lengthscale)
#self.parameters_changed()

View file

@ -36,7 +36,9 @@ class RBFInv(RBF):
def __init__(self, input_dim, variance=1., inv_lengthscale=None, ARD=False, name='inverse rbf'):
#self.input_dim = input_dim
#self.name = 'rbf_inv'
super(RBFInv, self).__init__(input_dim, variance=variance, lengthscale=1./np.array(inv_lengthscale), ARD=ARD, name=name)
if inv_lengthscale is not None: lengthscale = 1./np.array(inv_lengthscale)
else: lengthscale = None
super(RBFInv, self).__init__(input_dim, variance=variance, lengthscale=lengthscale, ARD=ARD, name=name)
self.ARD = ARD
if not ARD:
self.num_params = 2

View file

@ -17,11 +17,13 @@ except ImportError:
class KernelTests(unittest.TestCase):
def test_kerneltie(self):
K = GPy.kern.rbf(5, ARD=True)
K.tie_params('.*[01]')
K.constrain_fixed('2')
K.rbf.lengthscale[:2].tie_to(K.rbf.lengthscale[2:4])
K.rbf.lengthscale[2].constrain_fixed()
import ipdb;ipdb.set_trace()
X = np.random.rand(5,5)
Y = np.ones((5,1))
m = GPy.models.GPRegression(X,Y,K)
import ipdb;ipdb.set_trace()
self.assertTrue(m.checkgrad())
def test_rbfkernel(self):