Metrics
Get familiar with the application of Jensen-Shannon divergence (JSD) and the Kolmogorov-Smirnov (KS) two-sample test in comparing real and GAN-generated samples.
We'll cover the following
We are going to use the Jensen-Shannon divergence and the Kolgomorov-Smirnov two-sample test for comparing real samples and samples generated with GANs. We are going to use the Kolgomorov-Smirnov two-sample test implementation found on scipy.stats
and ks_2samp
.
Jensen-Shannon divergence (JSD)
As we described in the chapter, “Introduction to Generative Models,” the Jensen-Shannon divergence (JSD) is a symmetric and smoothed version of the Kullback-Leibler (KL) divergence:
Get hands-on with 1400+ tech skills courses.