mirror of
https://github.com/SheffieldML/GPy.git
synced 2026-05-12 13:32:39 +02:00
All gradients now gradcheck
This commit is contained in:
parent
6e405319b0
commit
e36ffcba6e
2 changed files with 82 additions and 77 deletions
|
|
@ -291,6 +291,7 @@ class StudentT(LikelihoodFunction):
|
|||
"""
|
||||
assert y.shape == f.shape
|
||||
e = y - f
|
||||
#FIXME: OUT BY SOME FUNCTION OF N
|
||||
dlik_dvar = self.v*(e**2 - self.sigma2)/(2*self.sigma2*(self.sigma2*self.v + e**2))
|
||||
return dlik_dvar
|
||||
|
||||
|
|
@ -442,7 +443,7 @@ class Gaussian(LikelihoodFunction):
|
|||
self.I = np.eye(self.N)
|
||||
self.covariance_matrix = self.I * self._variance
|
||||
self.Ki = self.I*(1.0 / self._variance)
|
||||
self.ln_K = np.trace(self.covariance_matrix)
|
||||
self.ln_det_K = np.sum(np.log(np.diag(self.covariance_matrix)))
|
||||
|
||||
def link_function(self, y, f, extra_data=None):
|
||||
"""link_function $\ln p(y|f)$
|
||||
|
|
@ -458,11 +459,11 @@ class Gaussian(LikelihoodFunction):
|
|||
e = y - f
|
||||
eeT = np.dot(e, e.T)
|
||||
objective = (- 0.5*self.D*np.log(2*np.pi)
|
||||
- 0.5*self.ln_K
|
||||
#- 0.5*np.sum(np.multiply(self.Ki, eeT))
|
||||
- 0.5*np.dot(np.dot(e.T, self.Ki), e)
|
||||
- 0.5*self.ln_det_K
|
||||
#- 0.5*np.dot(np.dot(e.T, self.Ki), e)
|
||||
- (0.5/self._variance)*np.dot(e.T, e) # As long as K is diagonal
|
||||
)
|
||||
return np.sum(objective) # FIXME: put this back!
|
||||
return np.sum(objective)
|
||||
|
||||
def dlik_df(self, y, f, extra_data=None):
|
||||
"""
|
||||
|
|
@ -514,7 +515,8 @@ class Gaussian(LikelihoodFunction):
|
|||
assert y.shape == f.shape
|
||||
e = y - f
|
||||
s_4 = 1.0/(self._variance**2)
|
||||
dlik_dsigma = -0.5*self.N/self._variance + 0.5*s_4*np.trace(np.dot(e.T, np.dot(self.I, e)))
|
||||
dlik_dsigma = -0.5*self.N/self._variance + 0.5*s_4*np.dot(e.T, e)
|
||||
#dlik_dsigma = -0.5*self.N + 0.5*s_4*np.dot(e.T, e)
|
||||
return dlik_dsigma
|
||||
|
||||
def dlik_df_dvar(self, y, f, extra_data=None):
|
||||
|
|
@ -523,7 +525,7 @@ class Gaussian(LikelihoodFunction):
|
|||
"""
|
||||
assert y.shape == f.shape
|
||||
s_4 = 1.0/(self._variance**2)
|
||||
dlik_grad_dsigma = -np.dot(s_4, np.dot(self.I, y)) + 0.5*np.dot(s_4, np.dot(self.I, f))
|
||||
dlik_grad_dsigma = -np.dot(s_4*self.I, y) + np.dot(s_4*self.I, f)
|
||||
return dlik_grad_dsigma
|
||||
|
||||
def d2lik_d2f_dvar(self, y, f, extra_data=None):
|
||||
|
|
@ -533,7 +535,7 @@ class Gaussian(LikelihoodFunction):
|
|||
$$\frac{d}{d\sigma}(\frac{d^{2}p(y_{i}|f_{i})}{d^{2}f}) = \frac{2\sigma v(v + 1)(\sigma^2 v - 3(y-f)^2)}{((y-f)^2 + \sigma^2 v)^3}$$
|
||||
"""
|
||||
assert y.shape == f.shape
|
||||
dlik_hess_dsigma = 0.5*np.diag((1.0/(self._variance**2))*self.I)[:, None]
|
||||
dlik_hess_dsigma = np.diag((1.0/(self._variance**2))*self.I)[:, None]
|
||||
return dlik_hess_dsigma
|
||||
|
||||
def _gradients(self, y, f, extra_data=None):
|
||||
|
|
|
|||
|
|
@ -3,6 +3,7 @@ import unittest
|
|||
import GPy
|
||||
from GPy.models import GradientChecker
|
||||
import functools
|
||||
import inspect
|
||||
|
||||
def dparam_partial(inst_func, *args):
|
||||
"""
|
||||
|
|
@ -22,66 +23,71 @@ def dparam_partial(inst_func, *args):
|
|||
return inst_func(*args)
|
||||
return functools.partial(param_func, inst_func=inst_func, args=args)
|
||||
|
||||
def grad_checker_wrt_params(func, dfunc, params, args, randomize=False, verbose=False):
|
||||
def dparam_checkgrad(func, dfunc, params, args, constrain_positive=True, randomize=False, verbose=False):
|
||||
"""
|
||||
checkgrad expects a f: R^N -> R^1 and df: R^N -> R^N
|
||||
However if we are holding other parameters fixed and moving something else
|
||||
We need to check the gradient of each of the fixed parameters (f and y for example) seperately
|
||||
Whilst moving another parameter. otherwise f: gives back R^N and df: gives back R^NxM where M is
|
||||
We need to check the gradient of each of the fixed parameters
|
||||
(f and y for example) seperately.
|
||||
Whilst moving another parameter. otherwise f: gives back R^N and
|
||||
df: gives back R^NxM where M is
|
||||
The number of parameters and N is the number of data
|
||||
Need to take a slice out from f and a slice out of df
|
||||
"""
|
||||
print "{} likelihood: {} vs {}".format(func.im_self.__class__.__name__,
|
||||
func.__name__, dfunc.__name__)
|
||||
#print "\n{} likelihood: {} vs {}".format(func.im_self.__class__.__name__,
|
||||
#func.__name__, dfunc.__name__)
|
||||
partial_f = dparam_partial(func, *args)
|
||||
partial_df = dparam_partial(dfunc, *args)
|
||||
gradchecked = False
|
||||
gradchecking = True
|
||||
for param in params:
|
||||
fnum = np.atleast_1d(partial_f(param)).shape[0]
|
||||
dfnum = np.atleast_1d(partial_df(param)).shape[0]
|
||||
for fixed_val in range(dfnum):
|
||||
f_ind = min(fnum, fixed_val+1) - 1 #dlik and dlik_dvar gives back 1 value for each
|
||||
#dlik and dlik_dvar gives back 1 value for each
|
||||
f_ind = min(fnum, fixed_val+1) - 1
|
||||
grad = GradientChecker(lambda x: np.atleast_1d(partial_f(x))[f_ind],
|
||||
lambda x : np.atleast_1d(partial_df(x))[fixed_val],
|
||||
param, 'p')
|
||||
grad.constrain_positive('p')
|
||||
if constrain_positive:
|
||||
grad.constrain_positive('p')
|
||||
if randomize:
|
||||
grad.randomize()
|
||||
print grad
|
||||
if verbose:
|
||||
grad.checkgrad(verbose=1)
|
||||
cg = grad.checkgrad()
|
||||
print cg
|
||||
if cg:
|
||||
print "True"
|
||||
gradchecked = True
|
||||
else:
|
||||
print "False"
|
||||
return False
|
||||
print str(gradchecked)
|
||||
return gradchecked
|
||||
if not grad.checkgrad():
|
||||
gradchecking = False
|
||||
|
||||
return gradchecking
|
||||
|
||||
|
||||
class LaplaceTests(unittest.TestCase):
|
||||
def setUp(self):
|
||||
self.N = 5
|
||||
self.D = 1
|
||||
self.N = 1
|
||||
self.D = 5
|
||||
self.X = np.linspace(0, 1, self.N)[:, None]
|
||||
|
||||
self.real_std = 0.2
|
||||
noise = np.random.randn(*self.X.shape)*self.real_std
|
||||
self.Y = np.sin(self.X*2*np.pi) + noise
|
||||
#self.Y = np.array([[1.0]])#np.sin(self.X*2*np.pi) + noise
|
||||
|
||||
self.f = np.random.rand(self.N, 1)
|
||||
#self.f = np.array([[3.0]])#np.sin(self.X*2*np.pi) + noise
|
||||
|
||||
self.var = 0.1
|
||||
self.var = np.random.rand(1)
|
||||
self.stu_t = GPy.likelihoods.functions.StudentT(deg_free=5, sigma2=self.var)
|
||||
self.gauss = GPy.likelihoods.functions.Gaussian(self.var, self.D, self.N)
|
||||
|
||||
def tearDown(self):
|
||||
self.stu_t = None
|
||||
self.gauss = None
|
||||
self.Y = None
|
||||
self.f = None
|
||||
self.X = None
|
||||
|
||||
def test_gaussian_dlik_df(self):
|
||||
print "\n{}".format(inspect.stack()[0][3])
|
||||
link = functools.partial(self.gauss.link_function, self.Y)
|
||||
dlik_df = functools.partial(self.gauss.dlik_df, self.Y)
|
||||
grad = GradientChecker(link, dlik_df, self.f.copy(), 'f')
|
||||
|
|
@ -90,6 +96,7 @@ class LaplaceTests(unittest.TestCase):
|
|||
self.assertTrue(grad.checkgrad())
|
||||
|
||||
def test_gaussian_d2lik_d2f(self):
|
||||
print "\n{}".format(inspect.stack()[0][3])
|
||||
dlik_df = functools.partial(self.gauss.dlik_df, self.Y)
|
||||
d2lik_d2f = functools.partial(self.gauss.d2lik_d2f, self.Y)
|
||||
grad = GradientChecker(dlik_df, d2lik_d2f, self.f.copy(), 'f')
|
||||
|
|
@ -98,6 +105,7 @@ class LaplaceTests(unittest.TestCase):
|
|||
self.assertTrue(grad.checkgrad())
|
||||
|
||||
def test_gaussian_d3lik_d3f(self):
|
||||
print "\n{}".format(inspect.stack()[0][3])
|
||||
d2lik_d2f = functools.partial(self.gauss.d2lik_d2f, self.Y)
|
||||
d3lik_d3f = functools.partial(self.gauss.d3lik_d3f, self.Y)
|
||||
grad = GradientChecker(d2lik_d2f, d3lik_d3f, self.f.copy(), 'f')
|
||||
|
|
@ -106,28 +114,31 @@ class LaplaceTests(unittest.TestCase):
|
|||
self.assertTrue(grad.checkgrad())
|
||||
|
||||
def test_gaussian_dlik_dvar(self):
|
||||
#link = dparam_partial(self.gauss.link_function, self.Y, self.f)
|
||||
#dlik_dvar = dparam_partial(self.gauss.dlik_dvar, self.Y, self.f)
|
||||
#grad = GradientChecker(link, dlik_dvar, self.var, 'v')
|
||||
#grad.constrain_positive('v')
|
||||
#grad.randomize()
|
||||
#grad.checkgrad(verbose=1)
|
||||
#self.assertTrue(grad.checkgrad())
|
||||
self.assertTrue(grad_checker_wrt_params(self.gauss.link_function, self.gauss.dlik_dvar,
|
||||
[self.var], args=(self.Y, self.f), randomize=True, verbose=True))
|
||||
print "\n{}".format(inspect.stack()[0][3])
|
||||
self.assertTrue(
|
||||
dparam_checkgrad(self.gauss.link_function, self.gauss.dlik_dvar,
|
||||
[self.var], args=(self.Y, self.f), constrain_positive=True,
|
||||
randomize=False, verbose=True)
|
||||
)
|
||||
|
||||
def test_gaussian_dlik_df_dvar(self):
|
||||
#dlik_df = dparam_partial(self.gauss.dlik_df, self.Y, self.f)
|
||||
#dlik_df_dvar = dparam_partial(self.gauss.dlik_df_dvar, self.Y, self.f)
|
||||
#grad = GradientChecker(dlik_df, dlik_df_dvar, self.var, 'v')
|
||||
#grad.constrain_positive('v')
|
||||
#grad.randomize()
|
||||
#grad.checkgrad(verbose=1)
|
||||
#self.assertTrue(grad.checkgrad())
|
||||
self.assertTrue(grad_checker_wrt_params(self.gauss.dlik_df, self.gauss.dlik_df_dvar,
|
||||
[self.var], args=(self.Y, self.f), randomize=True, verbose=True))
|
||||
print "\n{}".format(inspect.stack()[0][3])
|
||||
self.assertTrue(
|
||||
dparam_checkgrad(self.gauss.dlik_df, self.gauss.dlik_df_dvar,
|
||||
[self.var], args=(self.Y.copy(), self.f.copy()), constrain_positive=True,
|
||||
randomize=False, verbose=True)
|
||||
)
|
||||
|
||||
def test_gaussian_d2lik_d2f_dvar(self):
|
||||
print "\n{}".format(inspect.stack()[0][3])
|
||||
self.assertTrue(
|
||||
dparam_checkgrad(self.gauss.d2lik_d2f, self.gauss.d2lik_d2f_dvar,
|
||||
[self.var], args=(self.Y, self.f), constrain_positive=True,
|
||||
randomize=True, verbose=True)
|
||||
)
|
||||
|
||||
def test_studentt_dlik_df(self):
|
||||
print "\n{}".format(inspect.stack()[0][3])
|
||||
link = functools.partial(self.stu_t.link_function, self.Y)
|
||||
dlik_df = functools.partial(self.stu_t.dlik_df, self.Y)
|
||||
grad = GradientChecker(link, dlik_df, self.f.copy(), 'f')
|
||||
|
|
@ -135,6 +146,7 @@ class LaplaceTests(unittest.TestCase):
|
|||
grad.checkgrad(verbose=1)
|
||||
|
||||
def test_studentt_d2lik_d2f(self):
|
||||
print "\n{}".format(inspect.stack()[0][3])
|
||||
dlik_df = functools.partial(self.stu_t.dlik_df, self.Y)
|
||||
d2lik_d2f = functools.partial(self.stu_t.d2lik_d2f, self.Y)
|
||||
grad = GradientChecker(dlik_df, d2lik_d2f, self.f.copy(), 'f')
|
||||
|
|
@ -142,6 +154,7 @@ class LaplaceTests(unittest.TestCase):
|
|||
grad.checkgrad(verbose=1)
|
||||
|
||||
def test_studentt_d3lik_d3f(self):
|
||||
print "\n{}".format(inspect.stack()[0][3])
|
||||
d2lik_d2f = functools.partial(self.stu_t.d2lik_d2f, self.Y)
|
||||
d3lik_d3f = functools.partial(self.stu_t.d3lik_d3f, self.Y)
|
||||
grad = GradientChecker(d2lik_d2f, d3lik_d3f, self.f.copy(), 'f')
|
||||
|
|
@ -149,39 +162,29 @@ class LaplaceTests(unittest.TestCase):
|
|||
grad.checkgrad(verbose=1)
|
||||
|
||||
def test_studentt_dlik_dvar(self):
|
||||
#link = dparam_partial(self.stu_t.link_function, self.Y, self.f)
|
||||
#dlik_dvar = dparam_partial(self.stu_t.dlik_dvar, self.Y, self.f)
|
||||
#grad = GradientChecker(link, dlik_dvar, self.var, 'v')
|
||||
#grad.constrain_positive('v')
|
||||
#grad.randomize()
|
||||
#grad.checkgrad(verbose=1)
|
||||
#self.assertTrue(grad.checkgrad())
|
||||
self.assertTrue(grad_checker_wrt_params(self.stu_t.link_function, self.stu_t.dlik_dvar,
|
||||
[self.var], args=(self.Y.copy(), self.f.copy()), randomize=True, verbose=True))
|
||||
print "\n{}".format(inspect.stack()[0][3])
|
||||
self.assertTrue(
|
||||
dparam_checkgrad(self.stu_t.link_function, self.stu_t.dlik_dvar,
|
||||
[self.var], args=(self.Y.copy(), self.f.copy()),
|
||||
constrain_positive=True, randomize=True, verbose=True)
|
||||
)
|
||||
|
||||
def test_studentt_dlik_df_dvar(self):
|
||||
#dlik_df = dparam_partial(self.stu_t.dlik_df, self.Y, self.f)
|
||||
#dlik_df_dvar = dparam_partial(self.stu_t.dlik_df_dvar, self.Y, self.f)
|
||||
#grad = GradientChecker(dlik_df, dlik_df_dvar, self.var, 'v')
|
||||
#grad.constrain_positive('v')
|
||||
#grad.randomize()
|
||||
#grad.checkgrad(verbose=1)
|
||||
#self.assertTrue(grad.checkgrad())
|
||||
self.assertTrue(grad_checker_wrt_params(self.stu_t.dlik_df, self.stu_t.dlik_df_dvar,
|
||||
[self.var], args=(self.Y.copy(), self.f.copy()), randomize=True, verbose=True))
|
||||
print "\n{}".format(inspect.stack()[0][3])
|
||||
self.assertTrue(
|
||||
dparam_checkgrad(self.stu_t.dlik_df, self.stu_t.dlik_df_dvar,
|
||||
[self.var], args=(self.Y.copy(), self.f.copy()),
|
||||
constrain_positive=True, randomize=True, verbose=True)
|
||||
)
|
||||
|
||||
def test_studentt_d2lik_d2f_dvar(self):
|
||||
print "\n{}".format(inspect.stack()[0][3])
|
||||
self.assertTrue(
|
||||
dparam_checkgrad(self.stu_t.d2lik_d2f, self.stu_t.d2lik_d2f_dvar,
|
||||
[self.var], args=(self.Y.copy(), self.f.copy()),
|
||||
constrain_positive=True, randomize=True, verbose=True)
|
||||
)
|
||||
|
||||
if __name__ == "__main__":
|
||||
#N = 5
|
||||
#D = 1
|
||||
#X = np.linspace(0, 1, N)[:, None]
|
||||
#real_std = 0.2
|
||||
#noise = np.random.randn(*X.shape)*real_std
|
||||
#Y = np.sin(X*2*np.pi) + noise
|
||||
#f = np.random.rand(N, 1)
|
||||
#var = 0.1
|
||||
#stu_t = GPy.likelihoods.functions.StudentT(deg_free=5, sigma2=var)
|
||||
|
||||
#print grad_checker_wrt_params(stu_t.dlik_df, stu_t.dlik_df_dvar, [var], args=(Y, f), randomize=True, verbose=False)
|
||||
|
||||
print "Running unit tests"
|
||||
unittest.main()
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue