Gaussian Processes MLM Flashcards
Gaussian Processes (GP)
Gaussian Processes (GP) are used extensively in machine learning and statistics, particularly for regression problems.
- Introduction
Gaussian Processes (GP) is a powerful modeling framework incorporating kernels and Bayesian inference, and is recognized for its robustness and versatility. It is used for both regression and classification tasks, although it’s more commonly used for regression.
- Prior over Functions
One way to think about a Gaussian process is as a distribution over functions. Instead of inferring a distribution over the parameters of a parametric function, GP allows us to infer a distribution over the functions themselves.
- Kernel Functions
The choice of kernel function has a significant impact on the Gaussian process model. It defines the covariance between pairs of random variables and, therefore, determines the smoothness, periodicity, and other properties of the functions sampled from the Gaussian process. Common choices include Radial Basis Function (RBF), Matérn, and Rational Quadratic kernels, among others.
- Noise Model
In Gaussian process regression, a noise model is included to account for the observed variability in the data that isn’t captured by the Gaussian process prior. This is usually done by adding a noise term, typically assumed to be Gaussian, to the diagonal of the covariance matrix.
- Bayesian Inference
Bayesian inference is used in Gaussian processes to infer the posterior distribution over functions given some observed data. This is often done using Markov chain Monte Carlo (MCMC) methods or Variational Inference.
- Hyperparameters Learning
The hyperparameters of the Gaussian process, which might include the parameters of the kernel function and the noise term, are usually learned from the data by maximizing the marginal likelihood.
- Strengths and Limitations
Gaussian Processes are non-parametric, meaning they can represent a wide range of functions, and they provide a measure of uncertainty alongside predictions. However, they can be computationally expensive as the complexity scales cubically with the number of data points. They can also struggle with high-dimensional inputs.
- Applications
Gaussian processes are used in a variety of applications including regression, optimization, and in the design of experiments. In machine learning, one notable application is in the field of reinforcement learning, where GPs can be used to model the unknown dynamics of an environment.
- Extensions
Extensions to Gaussian processes include Sparse Gaussian Processes, which aim to make GPs more scalable, and Deep Gaussian Processes, which stack multiple GPs to create a deep learning model with the benefits of GPs.