Article

Machine Learning: Bayesian Inference

Used in Computational Vision calculations, Bayesian Inference is a method to update model hypotheses z following observations on data x.

It utilizes Bayes’ theorem, that - 

The prior hypothesis on the latent variables z, p(z), quantifies our belief before a new batch of data x is observed. We assume that data x are generated from a constant data-generating process, p(x) and that we can quantify the likelihood of observing x given our hypothesis z through the conditional distribution p(x|z). Despite being a somewhat abstract concept, Bayesian Inference has many real-world applications and is a powerful tool that underpins many machine learning techniques. A non-exhaustive list of applications includes:

  • Probabilistic and generative modelling – for example image standardization and Variational AutoEncoders
  • Uncertainty quantification for parametric models – for example Bayesian neural networks

One of the most ubiquitous applications for Bayesian Inference is uncertainty quantification. In parametric regression for example, our observed data are in fact pairs of datum (x,y) and our task is to predict labels y on new data points x, given that we have a closed-form expression for the posterior distribution p(z|{x,y}). Note that in this setting, the likelihood has a fixed-form of p(y|x,z). Uncertainty is then quantified by the posterior predictive distribution,

which has a closed-form expression for a limited number of scenarios, for example when p(y|x,z) is a normal distribution with isotropic standard deviation and whose expected value is a linear model of z. Otherwise, samples of the latent variable z must be generated from the posterior distribution p(z|{x,y}) and the expected value and variance of p(y│x,{x,y}) must be evaluated numerically.

It is worth concluding this introduction to Bayesian Inference by noting that the application of Bayes’ theorem as described here, assumes that we have a closed-form expression for evaluating the posterior distribution p(z│x). This is not always the case and approximations of p(z│x) must often be made in order for model predictions to be used, for example via the posterior predictive distribution. A large branch of machine learning called approximate inference, in fact exists to achieve just this.

At Element, we use the Bayesian Inference method within our Digital Engineering operations. Contact our Digital Engineering team for more information. 

Find related Resources

Our team of over 9,000 Engaged Experts in North America, Europe, The Middle East, Australia, Asia and Africa are ready to help you.