From 85ae035e32bee114b7230cc18665fd9ef864cccd Mon Sep 17 00:00:00 2001 From: mzwiessele Date: Tue, 24 Feb 2015 11:55:22 +0000 Subject: [PATCH 1/6] [var_dtc] constant jitter 1e-10 --- GPy/inference/latent_function_inference/var_dtc.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/GPy/inference/latent_function_inference/var_dtc.py b/GPy/inference/latent_function_inference/var_dtc.py index d61e7f0f..eaa02009 100644 --- a/GPy/inference/latent_function_inference/var_dtc.py +++ b/GPy/inference/latent_function_inference/var_dtc.py @@ -21,7 +21,7 @@ class VarDTC(LatentFunctionInference): For efficiency, we sometimes work with the cholesky of Y*Y.T. To save repeatedly recomputing this, we cache it. """ - const_jitter = 1e-6 + const_jitter = 1e-10 def __init__(self, limit=1): #self._YYTfactor_cache = caching.cache() from ...util.caching import Cacher From 72c6b6698376ba5b3772647309d88401caba45be Mon Sep 17 00:00:00 2001 From: mzwiessele Date: Tue, 24 Feb 2015 11:56:06 +0000 Subject: [PATCH 2/6] [updateable] update field in observable --- GPy/core/parameterization/updateable.py | 1 - 1 file changed, 1 deletion(-) diff --git a/GPy/core/parameterization/updateable.py b/GPy/core/parameterization/updateable.py index 278ba8cd..379e92e1 100644 --- a/GPy/core/parameterization/updateable.py +++ b/GPy/core/parameterization/updateable.py @@ -11,7 +11,6 @@ class Updateable(Observable): A model can be updated or not. Make sure updates can be switched on and off. """ - _updates = True def __init__(self, *args, **kwargs): super(Updateable, self).__init__(*args, **kwargs) From 924899069e583819d634198627eca68a30d44c4a Mon Sep 17 00:00:00 2001 From: mzwiessele Date: Tue, 24 Feb 2015 11:56:36 +0000 Subject: [PATCH 3/6] [optimization] experimental auto detect of ipython notebook --- GPy/core/model.py | 15 ++++++++++++--- 1 file changed, 12 insertions(+), 3 deletions(-) diff --git a/GPy/core/model.py b/GPy/core/model.py index c63a29e5..35b046fd 100644 --- a/GPy/core/model.py +++ b/GPy/core/model.py @@ -255,7 +255,16 @@ class Model(Parameterized): else: optimizer = optimization.get_optimizer(optimizer) opt = optimizer(start, model=self, max_iters=max_iters, **kwargs) - + + try: + from IPython.display import display + from IPython.html import widgets + display(widgets.TextWidget()) + ipython_notebook = True + except: + # Not in Ipython notebook + ipython_notebook = False + with VerboseOptimization(self, opt, maxiters=max_iters, verbose=messages, ipython_notebook=ipython_notebook) as vo: opt.run(f_fp=self._objective_grads, f=self._objective, fp=self._grads) vo.finish(opt) @@ -402,7 +411,7 @@ class Model(Parameterized): model_details = [['Model', self.name + '
'], ['Log-likelihood', '{}
'.format(float(self.log_likelihood()))], ["Number of Parameters", '{}
'.format(self.size)], - ["Updates", '{}
'.format(self._updates)], + ["Updates", '{}
'.format(self._update_on)], ] from operator import itemgetter to_print = ["""