Bayesian Regression
Learn about Bayesian regression techniques.
We'll cover the following...
Chapter Goals:
- Learn about Bayesian regression techniques
A. Bayesian techniques
So far, we've discussed hyperparameter optimization through cross-validation. Another way to optimize the hyperparameters of a regularized regression model is with Bayesian techniques.
In Bayesian statistics, the main idea is to make certain assumptions about the probability distributions of a model's parameters before being fitted on data. These initial distribution assumptions are called priors for the model's parameters.
In a Bayesian ridge regression model, there are two hyperparameters to optimize: α and λ. The α hyperparameter serves the same exact purpose as it does for regular ridge regression; namely, it acts as a scaling factor for the penalty term.
The λ hyperparameter acts as the precision of the model's weights. Basically, the smaller the λ value, the greater the variance between the individual weight values.
B. Hyperparameter priors
Both the α and λ hyperparameters have gamma distribution priors, meaning we assume both values ...