GPy/GPy/examples/classification.py

169 lines
4.4 KiB
Python
Raw Normal View History

# Copyright (c) 2012, GPy authors (see AUTHORS.txt).
# Licensed under the BSD 3-clause license (see LICENSE.txt)
2012-11-29 16:27:46 +00:00
"""
Gaussian Processes classification
2012-11-29 16:27:46 +00:00
"""
import pylab as pb
import numpy as np
import GPy
2013-05-17 17:17:30 +01:00
default_seed = 10000
def crescent_data(seed=default_seed): # FIXME
2012-11-29 16:27:46 +00:00
"""Run a Gaussian process classification on the crescent data. The demonstration calls the basic GP classification model and uses EP to approximate the likelihood.
:param model_type: type of model to fit ['Full', 'FITC', 'DTC'].
:param seed : seed value for data generation.
:type seed: int
:param inducing : number of inducing variables (only used for 'FITC' or 'DTC').
:type inducing: int
"""
2012-11-29 16:27:46 +00:00
data = GPy.util.datasets.crescent_data(seed=seed)
Y = data['Y']
Y[Y.flatten()==-1] = 0
2013-06-05 14:39:32 +01:00
m = GPy.models.GPClassification(data['X'], Y)
2013-03-11 11:41:46 +00:00
m.ensure_default_constraints()
m.update_likelihood_approximation()
m.optimize()
2012-11-29 16:27:46 +00:00
print(m)
m.plot()
return m
def oil():
"""
Run a Gaussian process classification on the oil data. The demonstration calls the basic GP classification model and uses EP to approximate the likelihood.
"""
2012-11-29 16:27:46 +00:00
data = GPy.util.datasets.oil()
Y = data['Y'][:, 0:1]
Y[Y.flatten()==-1] = 0
# Create GP model
2013-06-05 14:39:32 +01:00
m = GPy.models.GPClassification(data['X'], Y)
# Contrain all parameters to be positive
2012-11-29 16:27:46 +00:00
m.constrain_positive('')
2013-06-04 18:54:29 +01:00
m.tie_params('.*len')
m.update_likelihood_approximation()
2012-11-29 16:27:46 +00:00
# Optimize
2012-11-29 16:27:46 +00:00
m.optimize()
print(m)
return m
def toy_linear_1d_classification(seed=default_seed):
"""
Simple 1D classification example
2012-11-29 16:27:46 +00:00
:param seed : seed value for data generation (default is 4).
:type seed: int
"""
2012-11-29 16:27:46 +00:00
data = GPy.util.datasets.toy_linear_1d_classification(seed=seed)
Y = data['Y'][:, 0:1]
2013-06-04 16:23:04 +01:00
Y[Y.flatten() == -1] = 0
2012-11-29 16:27:46 +00:00
# Model definition
2013-06-05 14:39:32 +01:00
m = GPy.models.GPClassification(data['X'], Y)
2013-03-11 11:41:46 +00:00
m.ensure_default_constraints()
2012-11-29 16:27:46 +00:00
# Optimize
2013-03-11 11:41:46 +00:00
m.update_likelihood_approximation()
# Parameters optimization:
m.optimize()
# Plot
pb.subplot(211)
m.plot_f()
pb.subplot(212)
2013-02-01 17:58:21 +00:00
m.plot()
2012-11-29 16:27:46 +00:00
print(m)
2012-11-29 16:27:46 +00:00
return m
2013-03-11 14:05:56 +00:00
def sparse_toy_linear_1d_classification(seed=default_seed):
"""
2013-05-15 18:12:10 +01:00
Sparse 1D classification example
2013-03-11 14:05:56 +00:00
:param seed : seed value for data generation (default is 4).
:type seed: int
"""
data = GPy.util.datasets.toy_linear_1d_classification(seed=seed)
Y = data['Y'][:, 0:1]
Y[Y.flatten() == -1] = 0
2013-03-11 14:05:56 +00:00
# Model definition
2013-06-05 14:39:32 +01:00
m = GPy.models.SparseGPClassification(data['X'], Y)
2013-06-04 18:54:29 +01:00
m['.*len']= 2.
2013-03-11 14:05:56 +00:00
m.ensure_default_constraints()
# Optimize
m.update_likelihood_approximation()
# Parameters optimization:
m.optimize()
# Plot
pb.subplot(211)
m.plot_f()
pb.subplot(212)
m.plot()
print(m)
return m
def sparse_crescent_data(inducing=10, seed=default_seed):
2013-06-05 16:30:57 +01:00
"""
Run a Gaussian process classification with DTC approxiamtion on the crescent data. The demonstration calls the basic GP classification model and uses EP to approximate the likelihood.
2013-03-11 14:05:56 +00:00
:param model_type: type of model to fit ['Full', 'FITC', 'DTC'].
:param seed : seed value for data generation.
:type seed: int
:param inducing : number of inducing variables (only used for 'FITC' or 'DTC').
:type inducing: int
"""
data = GPy.util.datasets.crescent_data(seed=seed)
Y = data['Y']
Y[Y.flatten()==-1]=0
2013-03-11 14:05:56 +00:00
2013-06-05 14:39:32 +01:00
m = GPy.models.SparseGPClassification(data['X'], Y)
2013-03-11 14:05:56 +00:00
m.ensure_default_constraints()
2013-06-04 18:54:29 +01:00
m['.*len'] = 10.
2013-03-11 14:05:56 +00:00
m.update_likelihood_approximation()
m.optimize()
print(m)
m.plot()
return m
2013-06-05 14:11:28 +01:00
def FITC_crescent_data(inducing=10, seed=default_seed):
2013-06-05 16:30:57 +01:00
"""
Run a Gaussian process classification with FITC approximation on the crescent data. The demonstration uses EP to approximate the likelihood.
2013-06-05 14:11:28 +01:00
:param model_type: type of model to fit ['Full', 'FITC', 'DTC'].
:param seed : seed value for data generation.
:type seed: int
:param inducing : number of inducing variables (only used for 'FITC' or 'DTC').
:type inducing: int
"""
2013-06-05 16:30:57 +01:00
data = GPy.util.datasets.crescent_data(seed=seed)
Y = data['Y']
Y[Y.flatten()==-1]=0
2013-06-05 14:11:28 +01:00
data = GPy.util.datasets.crescent_data(seed=seed)
Y = data['Y']
Y[Y.flatten()==-1]=0
2013-06-05 14:39:32 +01:00
m = GPy.models.FITCClassification(data['X'], Y)
2013-06-05 14:11:28 +01:00
m.ensure_default_constraints()
2013-06-05 16:30:57 +01:00
m['.*len'] = 3.
2013-06-05 14:11:28 +01:00
m.update_likelihood_approximation()
m.optimize()
print(m)
m.plot()
return m