# Gaussian processes ###### tags: `Machine Learning II` ## Elements and distributions in probabilistic learning When doing probabilistic machine learning, our **ouput is** not a value, but a **probability density function**. Elements * **Observations/Data** $D$ : >[!TLDR] > >[!missing] >What is the Hypothesis space $H$. Distributions * **Prior** * **Likelihood** * **Posterior** There is also the **posterior predictive distribution**: $$p(y^*| \mathbf{x}^*, D) = \sum_h p(y^* = h(\mathbf{x}^*)|\mathbf{x}^*,h)\cdot p(h|D)$$ >[!tip] >Use priors that are easy to use mathematically ## Bayesian linear regression This form of regression assumes that the data was generated from a data distribution with parameters $\mathbf{w}$, but was contaminated with noise $\sigma_n$. >[!info] >The product of grouping several gaussians is a multivariate gaussian. >Recall that the marginals and conditionals of a Gaussian are also Gaussian themselves.