Generative Learning

Familiarize yourself with the basic principles behind generative learning and latent variable models.

Discriminative vs generative models

Machine Learning models are often categorized into discriminative and generative models. This distinction arises from the probabilistic formulation that we use to build and train those models.

Discriminative models learn the probability of a label yy based on a data point xx. In mathematical terms, this is denoted as p(yx)p( y | x). In order to categorize a data point into a class, we need to learn a mapping between the data and the classes. This mapping can be described as a probability distribution. Each label will “compete” with the other ones for probability density over a specific data point.

Generative models, on the other hand, learn a probability distribution over the data points without external labels. Mathematically, this is formulated as p(x)p(x). In this case, we have the data themselves “compete” for probability density.

Conditional Generative models are another category of models that try to learn the probability distribution of the data xx conditioned on the labels yy. As you can probably tell, this is denoted as p(xy)p(x|y). Here, we again have the data “compete” for density, but for each possible label.

It is important to clarify the notion of competition here. The probability density function pp is a normalized function whose integral overall value is equal to 1.

Xp(x)dx=1\int_{X} p(x) dx =1

It is evident that each data point ...