[objective grads] undid the change, as this would lead to dramatic problems in reloading old models using the pickle module

This commit is contained in:
Max Zwiessele 2015-01-13 13:53:28 +00:00
parent 20eff02061
commit 53dcd3f9fa
2 changed files with 12 additions and 12 deletions

View file

@ -257,7 +257,7 @@ class Model(Parameterized):
opt = optimizer(start, model=self, max_iters=max_iters, **kwargs)
with VerboseOptimization(self, opt, maxiters=max_iters, verbose=messages, ipython_notebook=ipython_notebook):
opt.run(f_fp=self._objective_and_grads, f=self._objective, fp=self._objective_grads)
opt.run(f_fp=self._objective_grads, f=self._objective, fp=self._grads)
self.optimization_runs.append(opt)
@ -314,7 +314,7 @@ class Model(Parameterized):
# evaulate around the point x
f1 = self._objective(x + dx)
f2 = self._objective(x - dx)
gradient = self._objective_grads(x)
gradient = self._grads(x)
dx = dx[transformed_index]
gradient = gradient[transformed_index]
@ -360,7 +360,7 @@ class Model(Parameterized):
print "No free parameters to check"
return
gradient = self._objective_grads(x).copy()
gradient = self._grads(x).copy()
np.where(gradient == 0, 1e-312, gradient)
ret = True
for nind, xind in itertools.izip(param_index, transformed_index):