mirror of
https://github.com/SheffieldML/GPy.git
synced 2026-05-08 19:42:39 +02:00
branch 'devel' of github.com:SheffieldML/GPy into devel
This commit is contained in:
commit
60e3d53027
3 changed files with 37 additions and 22 deletions
|
|
@ -6,6 +6,7 @@ The package for the Psi statistics computation of the linear kernel for Bayesian
|
||||||
"""
|
"""
|
||||||
|
|
||||||
import numpy as np
|
import numpy as np
|
||||||
|
from ....util.linalg import tdot
|
||||||
|
|
||||||
def psicomputations(variance, Z, variational_posterior):
|
def psicomputations(variance, Z, variational_posterior):
|
||||||
"""
|
"""
|
||||||
|
|
@ -19,9 +20,9 @@ def psicomputations(variance, Z, variational_posterior):
|
||||||
mu = variational_posterior.mean
|
mu = variational_posterior.mean
|
||||||
S = variational_posterior.variance
|
S = variational_posterior.variance
|
||||||
|
|
||||||
psi0 = np.einsum('q,nq->n',variance,np.square(mu)+S)
|
psi0 = (variance*(np.square(mu)+S)).sum(axis=1)
|
||||||
psi1 = np.einsum('q,mq,nq->nm',variance,Z,mu)
|
psi1 = np.dot(mu,(variance*Z).T)
|
||||||
psi2 = np.einsum('q,mq,oq,nq->mo',np.square(variance),Z,Z,S) + np.einsum('nm,no->mo',psi1,psi1)
|
psi2 = np.dot(S.sum(axis=0)*np.square(variance)*Z,Z.T)+ tdot(psi1.T)
|
||||||
|
|
||||||
return psi0, psi1, psi2
|
return psi0, psi1, psi2
|
||||||
|
|
||||||
|
|
@ -33,10 +34,12 @@ def psiDerivativecomputations(dL_dpsi0, dL_dpsi1, dL_dpsi2, variance, Z, variati
|
||||||
|
|
||||||
# Compute for psi0 and psi1
|
# Compute for psi0 and psi1
|
||||||
mu2S = np.square(mu)+S
|
mu2S = np.square(mu)+S
|
||||||
dL_dvar += np.einsum('n,nq->q',dL_dpsi0,mu2S) + np.einsum('nm,mq,nq->q',dL_dpsi1,Z,mu)
|
dL_dpsi0_var = dL_dpsi0[:,None]*variance[None,:]
|
||||||
dL_dmu += np.einsum('n,q,nq->nq',dL_dpsi0,2.*variance,mu) + np.einsum('nm,q,mq->nq',dL_dpsi1,variance,Z)
|
dL_dpsi1_mu = np.dot(dL_dpsi1.T,mu)
|
||||||
dL_dS += np.einsum('n,q->nq',dL_dpsi0,variance)
|
dL_dvar += (dL_dpsi0[:,None]*mu2S).sum(axis=0)+ (dL_dpsi1_mu*Z).sum(axis=0)
|
||||||
dL_dZ += np.einsum('nm,q,nq->mq',dL_dpsi1, variance,mu)
|
dL_dmu += 2.*dL_dpsi0_var*mu+np.dot(dL_dpsi1,Z)*variance
|
||||||
|
dL_dS += dL_dpsi0_var
|
||||||
|
dL_dZ += dL_dpsi1_mu*variance
|
||||||
|
|
||||||
return dL_dvar, dL_dZ, dL_dmu, dL_dS
|
return dL_dvar, dL_dZ, dL_dmu, dL_dS
|
||||||
|
|
||||||
|
|
@ -55,17 +58,20 @@ def _psi2computations(dL_dpsi2, variance, Z, mu, S):
|
||||||
# _psi2_dS NxQ
|
# _psi2_dS NxQ
|
||||||
|
|
||||||
variance2 = np.square(variance)
|
variance2 = np.square(variance)
|
||||||
common_sum = np.einsum('q,mq,nq->nm',variance,Z,mu) # NxM
|
common_sum = np.dot(mu,(variance*Z).T)
|
||||||
Z_expect = np.einsum('mo,mq,oq->q',dL_dpsi2,Z,Z)
|
Z_expect = (np.dot(dL_dpsi2,Z)*Z).sum(axis=0)
|
||||||
common_expect = np.einsum('mo,mq,no->nq',dL_dpsi2+dL_dpsi2.T,Z,common_sum)
|
dL_dpsi2T = dL_dpsi2+dL_dpsi2.T
|
||||||
|
common_expect = np.dot(common_sum,np.dot(dL_dpsi2T,Z))
|
||||||
|
Z2_expect = np.inner(common_sum,dL_dpsi2T)
|
||||||
|
Z1_expect = np.dot(dL_dpsi2T,Z)
|
||||||
|
|
||||||
dL_dvar = np.einsum('q,nq,q->q',Z_expect,2.*S,variance)+ np.einsum('nq,nq->q',common_expect,mu)
|
dL_dvar = 2.*S.sum(axis=0)*variance*Z_expect+(common_expect*mu).sum(axis=0)
|
||||||
|
|
||||||
dL_dmu = np.einsum('nq,q->nq',common_expect,variance)
|
dL_dmu = common_expect*variance
|
||||||
|
|
||||||
dL_dS = np.empty(S.shape)
|
dL_dS = np.empty(S.shape)
|
||||||
dL_dS[:] = np.einsum('q,q->q',Z_expect,variance2)
|
dL_dS[:] = Z_expect*variance2
|
||||||
|
|
||||||
dL_dZ = 2.*(np.einsum('om,q,mq,nq->oq',dL_dpsi2,variance2,Z,S)+np.einsum('om,q,nq,nm->oq',dL_dpsi2,variance,mu,common_sum))
|
dL_dZ = variance2*S.sum(axis=0)*Z1_expect+np.dot(Z2_expect.T,variance*mu)
|
||||||
|
|
||||||
return dL_dvar, dL_dmu, dL_dS, dL_dZ
|
return dL_dvar, dL_dmu, dL_dS, dL_dZ
|
||||||
|
|
|
||||||
|
|
@ -5,6 +5,8 @@
|
||||||
The package for the Psi statistics computation of the linear kernel for SSGPLVM
|
The package for the Psi statistics computation of the linear kernel for SSGPLVM
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
from ....util.linalg import tdot
|
||||||
|
|
||||||
import numpy as np
|
import numpy as np
|
||||||
|
|
||||||
def psicomputations(variance, Z, variational_posterior):
|
def psicomputations(variance, Z, variational_posterior):
|
||||||
|
|
@ -20,10 +22,9 @@ def psicomputations(variance, Z, variational_posterior):
|
||||||
S = variational_posterior.variance
|
S = variational_posterior.variance
|
||||||
gamma = variational_posterior.binary_prob
|
gamma = variational_posterior.binary_prob
|
||||||
|
|
||||||
psi0 = np.einsum('q,nq,nq->n',variance,gamma,np.square(mu)+S)
|
psi0 = (gamma*(np.square(mu)+S)*variance).sum(axis=-1)
|
||||||
psi1 = np.einsum('nq,q,mq,nq->nm',gamma,variance,Z,mu)
|
psi1 = np.inner(variance*gamma*mu,Z)
|
||||||
psi2 = np.einsum('nq,q,mq,oq,nq->mo',gamma,np.square(variance),Z,Z,(1-gamma)*np.square(mu)+S) +\
|
psi2 = np.inner(np.square(variance)*(gamma*((1-gamma)*np.square(mu)+S)).sum(axis=0)*Z,Z)+tdot(psi1.T)
|
||||||
np.einsum('nm,no->mo',psi1,psi1)
|
|
||||||
|
|
||||||
return psi0, psi1, psi2
|
return psi0, psi1, psi2
|
||||||
|
|
||||||
|
|
@ -63,9 +64,16 @@ def _psi2computations(dL_dpsi2, variance, Z, mu, S, gamma):
|
||||||
gamma2 = np.square(gamma)
|
gamma2 = np.square(gamma)
|
||||||
variance2 = np.square(variance)
|
variance2 = np.square(variance)
|
||||||
mu2S = mu2+S # NxQ
|
mu2S = mu2+S # NxQ
|
||||||
common_sum = np.einsum('nq,q,mq,nq->nm',gamma,variance,Z,mu) # NxM
|
gvm = np.einsum('nq,nq,q->nq',gamma,mu,variance)
|
||||||
|
common_sum = np.einsum('nq,mq->nm',gvm,Z)
|
||||||
|
# common_sum = np.einsum('nq,q,mq,nq->nm',gamma,variance,Z,mu) # NxM
|
||||||
Z_expect = np.einsum('mo,mq,oq->q',dL_dpsi2,Z,Z)
|
Z_expect = np.einsum('mo,mq,oq->q',dL_dpsi2,Z,Z)
|
||||||
common_expect = np.einsum('mo,mq,no->nq',dL_dpsi2+dL_dpsi2.T,Z,common_sum)
|
dL_dpsi2T = dL_dpsi2+dL_dpsi2.T
|
||||||
|
tmp = np.einsum('mo,oq->mq',dL_dpsi2T,Z)
|
||||||
|
common_expect = np.einsum('mq,nm->nq',tmp,common_sum)
|
||||||
|
# common_expect = np.einsum('mo,mq,no->nq',dL_dpsi2+dL_dpsi2.T,Z,common_sum)
|
||||||
|
Z2_expect = np.einsum('om,nm->no',dL_dpsi2T,common_sum)
|
||||||
|
Z1_expect = np.einsum('om,mq->oq',dL_dpsi2T,Z)
|
||||||
|
|
||||||
dL_dvar = np.einsum('nq,q,q->q',2.*(gamma*mu2S-gamma2*mu2),variance,Z_expect)+\
|
dL_dvar = np.einsum('nq,q,q->q',2.*(gamma*mu2S-gamma2*mu2),variance,Z_expect)+\
|
||||||
np.einsum('nq,nq,nq->q',common_expect,gamma,mu)
|
np.einsum('nq,nq,nq->q',common_expect,gamma,mu)
|
||||||
|
|
@ -78,6 +86,7 @@ def _psi2computations(dL_dpsi2, variance, Z, mu, S, gamma):
|
||||||
|
|
||||||
dL_dS = np.einsum('q,nq,q->nq',Z_expect,gamma,variance2)
|
dL_dS = np.einsum('q,nq,q->nq',Z_expect,gamma,variance2)
|
||||||
|
|
||||||
dL_dZ = 2.*(np.einsum('om,nq,q,mq,nq->oq',dL_dpsi2,gamma,variance2,Z,(mu2S-gamma*mu2))+np.einsum('om,nq,q,nq,nm->oq',dL_dpsi2,gamma,variance,mu,common_sum))
|
# dL_dZ = 2.*(np.einsum('om,nq,q,mq,nq->oq',dL_dpsi2,gamma,variance2,Z,(mu2S-gamma*mu2))+np.einsum('om,nq,q,nq,nm->oq',dL_dpsi2,gamma,variance,mu,common_sum))
|
||||||
|
dL_dZ = Z1_expect*np.einsum('nq,q,nq->q',gamma,variance2,(mu2S-gamma*mu2))+np.einsum('nq,q,nq,nm->mq',gamma,variance,mu,Z2_expect)
|
||||||
|
|
||||||
return dL_dvar, dL_dgamma, dL_dmu, dL_dS, dL_dZ
|
return dL_dvar, dL_dgamma, dL_dmu, dL_dS, dL_dZ
|
||||||
|
|
|
||||||
|
|
@ -25,7 +25,7 @@ def add_bar_labels(fig, ax, bars, bottom=0):
|
||||||
c = 'w'
|
c = 'w'
|
||||||
t = TextPath((0, 0), "${xi}$".format(xi=xi), rotation=0, ha='center')
|
t = TextPath((0, 0), "${xi}$".format(xi=xi), rotation=0, ha='center')
|
||||||
transform = transOffset
|
transform = transOffset
|
||||||
if patch.get_extents().height <= t.get_extents().height + 3:
|
if patch.get_extents().height <= t.get_extents().height + 5:
|
||||||
va = 'bottom'
|
va = 'bottom'
|
||||||
c = 'k'
|
c = 'k'
|
||||||
transform = transOffsetUp
|
transform = transOffsetUp
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue