GPJax is designed to be a Gaussian process package that provides an
accurate representation of the underlying maths. Variable names are chosen to closely match the notation in {cite}rasmussen2006gaussian.
We here list the notation used in GPJax with its corresponding mathematical quantity.
| On paper | GPJax code | Description |
|---|---|---|
| n | Number of train inputs | |
| x | Train inputs | |
| y | Train labels | |
| t | Test inputs | |
| f | Latent function modelled as a GP | |
| fx | Latent function at inputs |
|
| μx | Prior mean at inputs |
|
| Kxx | Kernel Gram matrix at inputs |
|
| Lx | Lower Cholesky decomposition of |
|
| Ktx | Cross-covariance between inputs |
| On paper | GPJax code | Description |
|---|---|---|
| m | Number of inducing inputs | |
| z | Inducing inputs | |
| u | Inducing outputs |
Prior to building GPJax, the developers of GPJax have benefited greatly from the GPFlow and GPyTorch packages. As such, many of the design principles in GPJax are inspired by the excellent precursory pacakges. Documentation designs have been greatly inspired by the exceptional Flax docs.