Skip to content

Commit

Permalink
Fix natgrads notebook
Browse files Browse the repository at this point in the history
  • Loading branch information
thomaspinder committed Oct 23, 2022
1 parent 793d1f8 commit 269f670
Show file tree
Hide file tree
Showing 2 changed files with 14 additions and 17 deletions.
2 changes: 1 addition & 1 deletion examples/natgrads.pct.py
Original file line number Diff line number Diff line change
Expand Up @@ -92,7 +92,7 @@
parameter_state=parameter_state,
train_data=D,
n_iters=5000,
batch_size=100,
batch_size=256,
key=jr.PRNGKey(42),
moment_optim=ox.sgd(0.1),
hyper_optim=ox.adam(1e-3),
Expand Down
29 changes: 13 additions & 16 deletions gpjax/variational_families.py
Original file line number Diff line number Diff line change
Expand Up @@ -171,7 +171,7 @@ def predict(
Lz = jnp.linalg.cholesky(Kzz)
μz = self.prior.mean_function(z, params["mean_function"])

def predict_fn(test_inputs: Float[Array, "N D"]) -> dx.MultivariateNormalTri:
def predict_fn(test_inputs: Float[Array, "N D"]) -> dx.MultivariateNormalFullCovariance:
t = test_inputs
n_test = t.shape[0]
Ktt = gram(self.prior.kernel, t, params["kernel"])
Expand All @@ -197,8 +197,7 @@ def predict_fn(test_inputs: Float[Array, "N D"]) -> dx.MultivariateNormalTri:
+ jnp.matmul(Ktz_Kzz_inv_sqrt, Ktz_Kzz_inv_sqrt.T)
)
covariance += I(n_test) * self.jitter
L = jnp.linalg.cholesky(covariance)
return dx.MultivariateNormalTri(jnp.atleast_1d(mean.squeeze()), L)
return dx.MultivariateNormalFullCovariance(jnp.atleast_1d(mean.squeeze()), covariance)

return predict_fn

Expand Down Expand Up @@ -233,7 +232,7 @@ def prior_kl(self, params: Dict) -> Float[Array, "1"]:

def predict(
self, params: Dict
) -> Callable[[Float[Array, "N D"]], dx.MultivariateNormalTri]:
) -> Callable[[Float[Array, "N D"]], dx.MultivariateNormalFullCovariance]:
"""Compute the predictive distribution of the GP at the test inputs t.
This is the integral q(f(t)) = ∫ p(f(t)|u) q(u) du, which can be computed in closed form as
Expand All @@ -255,7 +254,7 @@ def predict(
Kzz += I(m) * self.jitter
Lz = jnp.linalg.cholesky(Kzz)

def predict_fn(test_inputs: Float[Array, "N D"]) -> dx.MultivariateNormalTri:
def predict_fn(test_inputs: Float[Array, "N D"]) -> dx.MultivariateNormalFullCovariance:
t = test_inputs
n_test = t.shape[0]
Ktt = gram(self.prior.kernel, t, params["kernel"])
Expand All @@ -280,7 +279,7 @@ def predict_fn(test_inputs: Float[Array, "N D"]) -> dx.MultivariateNormalTri:
covariance += I(n_test) * self.jitter
L = jnp.linalg.cholesky(covariance)

return dx.MultivariateNormalTri(jnp.atleast_1d(mean.squeeze()), L)
return dx.MultivariateNormalFullCovariance(jnp.atleast_1d(mean.squeeze()), covariance)

return predict_fn

Expand Down Expand Up @@ -364,7 +363,7 @@ def prior_kl(self, params: Dict) -> Float[Array, "1"]:

def predict(
self, params: Dict
) -> Callable[[Float[Array, "N D"]], dx.MultivariateNormalTri]:
) -> Callable[[Float[Array, "N D"]], dx.MultivariateNormalFullCovariance]:
"""Compute the predictive distribution of the GP at the test inputs t.
This is the integral q(f(t)) = ∫ p(f(t)|u) q(u) du, which can be computed in closed form as
Expand Down Expand Up @@ -431,9 +430,8 @@ def predict_fn(test_inputs: Float[Array, "N D"]) -> dx.MultivariateNormalTri:
- jnp.matmul(Lz_inv_Kzt.T, Lz_inv_Kzt)
+ jnp.matmul(Ktz_Kzz_inv_L, Ktz_Kzz_inv_L.T)
)
L = jnp.linalg.cholesky(covariance)

return dx.MultivariateNormalTri(jnp.atleast_1d(mean.squeeze()), L)
return dx.MultivariateNormalFullCovariance(jnp.atleast_1d(mean.squeeze()), covariance)

return predict_fn

Expand Down Expand Up @@ -515,7 +513,7 @@ def prior_kl(self, params: Dict) -> Float[Array, "1"]:

def predict(
self, params: Dict
) -> Callable[[Float[Array, "N D"]], dx.MultivariateNormalTri]:
) -> Callable[[Float[Array, "N D"]], dx.MultivariateNormalFullCovariance]:
"""Compute the predictive distribution of the GP at the test inputs t.
This is the integral q(f(t)) = ∫ p(f(t)|u) q(u) du, which can be computed in closed form as
Expand Down Expand Up @@ -554,7 +552,7 @@ def predict(
Lz = jnp.linalg.cholesky(Kzz)
μz = self.prior.mean_function(z, params["mean_function"])

def predict_fn(test_inputs: Float[Array, "N D"]) -> dx.MultivariateNormalTri:
def predict_fn(test_inputs: Float[Array, "N D"]) -> dx.MultivariateNormalFullCovariance:
t = test_inputs
Ktt = gram(self.prior.kernel, t, params["kernel"])
Kzt = cross_covariance(self.prior.kernel, z, t, params["kernel"])
Expand All @@ -579,7 +577,7 @@ def predict_fn(test_inputs: Float[Array, "N D"]) -> dx.MultivariateNormalTri:
+ jnp.matmul(Ktz_Kzz_inv_sqrt, Ktz_Kzz_inv_sqrt.T)
)
L = jnp.linalg.cholesky(covariance + I(t.shape[0]) * self.jitter)
return dx.MultivariateNormalTri(jnp.atleast_1d(mean.squeeze()), L)
return dx.MultivariateNormalFullCovariance(jnp.atleast_1d(mean.squeeze()), covariance)

return predict_fn

Expand Down Expand Up @@ -617,14 +615,14 @@ def _initialise_params(self, key: PRNGKeyType) -> Dict:

def predict(
self, train_data: Dataset, params: Dict
) -> Callable[[Float[Array, "N D"]], dx.MultivariateNormalTri]:
) -> Callable[[Float[Array, "N D"]], dx.MultivariateNormalFullCovariance]:
"""Compute the predictive distribution of the GP at the test inputs.
Args:
params (Dict): The set of parameters that are to be used to parameterise our variational approximation and GP.
Returns:
Callable[[Float[Array, "N D"]], dx.MultivariateNormalTri]: A function that accepts a set of test points and will return the predictive distribution at those points.
"""
def predict_fn(test_inputs: Float[Array, "N D"]) -> dx.MultivariateNormalTri:
def predict_fn(test_inputs: Float[Array, "N D"]) -> dx.MultivariateNormalFullCovariance:
# TODO - can we cache some of this?
x, y = train_data.X, train_data.y

Expand Down Expand Up @@ -681,8 +679,7 @@ def predict_fn(test_inputs: Float[Array, "N D"]) -> dx.MultivariateNormalTri:
- jnp.matmul(Lz_inv_Kzt.T, Lz_inv_Kzt)
+ jnp.matmul(L_inv_Lz_inv_Kzt.T, L_inv_Lz_inv_Kzt)
)
L = jnp.linalg.cholesky(covariance + I(t.shape[0]) * self.jitter)
return dx.MultivariateNormalTri(jnp.atleast_1d(mean.squeeze()), L)
return dx.MultivariateNormalFullCovariance(jnp.atleast_1d(mean.squeeze()), covariance)

return predict_fn

Expand Down

0 comments on commit 269f670

Please sign in to comment.