官术网_书友最值得收藏!

Comparing discriminative and generative models

Learning the conditional distribution is easier, because you do not have to make assumptions about the marginal distribution of x or y.

We will use the following diagram to illustrate the differences between discriminative and generative models. We can see two plots with 13 points of a two-dimensional dataset; let's call the blue class labels , and the yellow class labels :

When training a discriminative model, we want to estimate the hidden parameters of the model that describe the conditional probability distribution that provides a decision boundary with an optimal split between the classes at hand. When training a generative model, , we want to estimate the parameters that describe the joint probability distribution of x and y.

In addition to predicting the conditional probability, learning the joint probability distribution allows us to sample the learned model to generate new data from , where is conditioned on and is conditioned on . In the preceding diagram, for example, you could model the joint probability by learning the hidden parameters of a mixture distribution; for example, a Gaussian mixture with one component per class.

Another way to visualize the difference between generative and discriminative models is to look at a graphical depiction of the distribution that is being modeled. In the following diagram, we can see that the depiction of the discriminative model shows a decision boundary that can be used to define the class label, given some fixed data. In this case, predicting  can be seen as finding a decision boundary from which the distance of a datapoint to the boundary is proportional to the probability of that datapoint belonging to a class.

In a binary classification task on a single variable-let's call it -the simplest form of such a model is to find the boundary at which more samples are labeled correctly. In the following figure, the value of  that maximizes the number of correct labels is around 50. 

The following depiction of the generative model shows the exact distribution of  in the presence and absence of . Naturally, given that we know the exact distribution of and , we can sample it to generate new data:

Since generative models handle the hard task of modeling all dependencies and patterns that are in the input and output data, applications of generative models are uncountable. The deep learning field has produced state-of-the-art generative models for applications such as image generation, speech synthesis, and model-based control.

A fascinating aspect of generative models is that they are potentially capable of learning large and complex data distributions with a relatively small number of parameters. Unlike discriminative models, generative models can learn meaningful features from large, unlabeled datasets, a process that requires little to no labeling or human supervision.

Most recent work in generative models has been focused on GANs and likelihood-based methods, including autoregressive models, Variational Autoencoders (VAEs), and flow-based models. In the following paragraphs, we will describe likelihood-based models and variations thereof. Later, we will describe the GAN framework in detail.

主站蜘蛛池模板: 南昌县| 哈巴河县| 尼木县| 延安市| 宁晋县| 南安市| 汤阴县| 漳州市| 富裕县| 成武县| 濮阳县| 包头市| 牡丹江市| 法库县| 湟中县| 南雄市| 德昌县| 青冈县| 永昌县| 泗洪县| 疏勒县| 大新县| 平乐县| 承德市| 五指山市| 怀远县| 天水市| 济宁市| 改则县| 疏勒县| 土默特右旗| 壤塘县| 会宁县| 雷山县| 时尚| 公主岭市| 汝阳县| 报价| 和林格尔县| 潞城市| 长岭县|