diff --git a/GPy/core/model.py b/GPy/core/model.py index 4a1791bd..145a607f 100644 --- a/GPy/core/model.py +++ b/GPy/core/model.py @@ -303,7 +303,7 @@ class model(parameterised): return '\n'.join(s) - def checkgrad(self, verbose=False, include_priors=False, step=1e-6, tolerance = 1e-3, *args): + def checkgrad(self, verbose=False, include_priors=False, step=1e-6, tolerance = 1e-3, return_ratio=False, *args): """ Check the gradient of the model by comparing to a numerical estimate. If the overall gradient fails, invividual components are tested. @@ -323,12 +323,12 @@ class model(parameterised): gradient = self._log_likelihood_gradients_transformed() numerical_gradient = (f1-f2)/(2*dx) - ratio = (f1-f2)/(2*np.dot(dx,gradient)) + global_ratio = (f1-f2)/(2*np.dot(dx,gradient)) if verbose: - print "Gradient ratio = ", ratio, '\n' + print "Gradient ratio = ", global_ratio, '\n' sys.stdout.flush() - if (np.abs(1.-ratio)