机器学习-白板推导系列笔记(三十一)-GAN

此文章主要是结合哔站shuhuai008大佬的白板推导视频:生成对抗网络_54min

全部笔记的汇总贴:机器学习-白板推导系列笔记

一、例子

在这里插入图片描述
其中国宝是一个静态的,不会改变,工艺品和这个节目的鉴定水平是动态的,可学习的。

目标:成为高水平、可以以假乱真的大师。

  1. 高水平的鉴赏专家(手段)
  2. 高水平的工艺品大师(目标)
    (高大师(高专家))

在这里插入图片描述

二、数学描述

我们将上图转化为数学符号

古人: { x i } i = 1 N : P d a t a \{x_i\}_{i=1}^N:P_{data} { xi}i=1N:Pdata
工艺品: P g ( x ; θ g ) : g e n e r a t o r ( P z ( z ) + G ( z ; θ g ) )              Z ∼ P Z ( z )                    x = G ( Z ; θ g ) P_g(x;\theta_g):generator(P_z(z)+G(z;\theta_g))\;\;\;\;\;\;Z\sim P_Z(z)\;\;\;\;\;\;\;\;\;x=G(Z;\theta_g) Pg(x;θg):generator(Pz(z)+G(z;θg))ZPZ(z)x=G(Z;θg)

在这里插入图片描述
x x x是国宝的概率: D ( x ; θ d ) D(x;\theta_d) D(x;θd)
在这里插入图片描述

高专家:

如果 x x x来自与 P d a t a P_{data} Pdata,则 D ( x ) D(x) D(x)相对较高。(可以改写为 log ⁡ D ( x ) \log D(x) logD(x)
如果 x x x来自与 P g P_{g} Pg(相当于 Z Z Z来自于 P z P_z Pz),则 D ( x ) D(x) D(x)相对较低(可以改写为 log ⁡ ( 1 − D ( G ( z ) ) ) \log (1-D(G(z))) log(1D(G(z))),则这个应该较高)

max ⁡ D [ E x ∼ P d a t a [ log ⁡ D ( x ) ] + E z ∼ P z [ log ⁡ ( 1 − D ( G ( z ) ) ) ] ] \max_D\Bigg[E_{x\sim P_{data}}\Big[\log D(x)\Big]+E_{z\sim P_{z}}\Big[\log (1-D(G(z)))\Big]\Bigg] Dmax[ExPdata[logD(x)]+EzPz[log(1D(G(z)))]]

扫描二维码关注公众号,回复: 12192665 查看本文章

高大师:

如果 x x x来自与 P g P_{g} Pg(相当于 Z Z Z来自于 P z P_z Pz),则 D ( x ) D(x) D(x)相对较高(可以改写为 log ⁡ ( 1 − D ( G ( z ) ) ) \log (1-D(G(z))) log(1D(G(z))),则这个应该较低)
min ⁡ G E z ∼ P z [ log ⁡ ( 1 − D ( G ( z ) ) ) ] \min_GE_{z\sim P_{z}}\Big[\log (1-D(G(z)))\Big] GminEzPz[log(1D(G(z)))]

总目标:

min ⁡ G max ⁡ D [ E x ∼ P d a t a [ log ⁡ D ( x ) ] + E z ∼ P z [ log ⁡ ( 1 − D ( G ( z ) ) ) ] ] \min_G\max_D\Bigg[E_{x\sim P_{data}}\Big[\log D(x)\Big]+E_{z\sim P_{z}}\Big[\log (1-D(G(z)))\Big]\Bigg] GminDmax[ExPdata[logD(x)]+EzPz[log(1D(G(z)))]]

三、全局最优解

y ∣ x : d i s c r i m i n a t o r y|x:discriminator yx:discriminator

y / x y/x y/x 1 0
p p p D ( x ) D(x) D(x) 1 − D ( x ) 1-D(x) 1D(x)

V ( D , G ) = E x ∼ P d a t a [ log ⁡ D ( x ) ] + E x ∼ P g [ log ⁡ ( 1 − D ( x ) ) ] V(D,G)=E_{x\sim P_{data}}\Big[\log D(x)\Big]+E_{x\sim P_{g}}\Big[\log (1-D(x))\Big] V(D,G)=ExPdata[logD(x)]+ExPg[log(1D(x))]

固定 G G G,求 D ∗ D^* D,记作 D G ∗ D^*_G DG

max ⁡ D V ( D , G ) \max_DV(D,G) DmaxV(D,G)
max ⁡ D V ( D , G ) = ∫ P d a t a ⋅ log ⁡ D d x + ∫ P g ⋅ log ⁡ ( 1 − D ) d x = ∫ [ P d a t a ⋅ log ⁡ D + P g ⋅ log ⁡ ( 1 − D ) ] d x \max_DV(D,G)=\int P_{data}\cdot\log D{d}x+\int P_g\cdot\log(1-D){d}x\\=\int\Big[P_{data}\cdot\log D+ P_g\cdot\log(1-D)\Big]{d}x DmaxV(D,G)=PdatalogDdx+Pglog(1D)dx=[PdatalogD+Pglog(1D)]dx
关于 D D D求偏导:
∂ ∂ D ( max ⁡ D V ( D , G ) ) = ∂ ∂ D ∫ [ P d a t a ⋅ log ⁡ D + P g ⋅ log ⁡ ( 1 − D ) ] d x = ∫ ∂ ∂ D [ P d a t a ⋅ log ⁡ D + P g ⋅ log ⁡ ( 1 − D ) ] d x = ∫ [ P d a t a ⋅ 1 D + P g ⋅ − 1 1 − D ] d x \frac{\partial }{\partial D}(\max_DV(D,G))=\frac{\partial }{\partial D}\int\Big[P_{data}\cdot\log D+ P_g\cdot\log(1-D)\Big]{d}x\\=\int\frac{\partial }{\partial D}\Big[P_{data}\cdot\log D+ P_g\cdot\log(1-D)\Big]{d}x\\=\int\Big[P_{data}\cdot\frac1D+ P_g\cdot\frac{-1}{1-D}\Big]{d}x D(DmaxV(D,G))=D[PdatalogD+Pglog(1D)]dx=D[PdatalogD+Pglog(1D)]dx=[PdataD1+Pg1D1]dx
令导数为 0 0 0,得到:
D G ∗ = P d a t a P d a t a + P g D^*_G=\frac{P_{data}}{P_{data}+P_g} DG=Pdata+PgPdata

D G ∗ D^*_G DG代入,则有:

min ⁡ G max ⁡ D V ( D , G ) = min ⁡ G V ( D G ∗ , G ) = min ⁡ G E x ∼ P d a t a [ log ⁡ P d a t a P d a t a + P g ] + E x ∼ P g [ log ⁡ ( 1 − P d a t a P d a t a + P g ) ] = min ⁡ G E x ∼ P d a t a [ log ⁡ P d a t a P d a t a + P g ] + E x ∼ P g [ log ⁡ P g P d a t a + P g ] = min ⁡ G E x ∼ P d a t a [ log ⁡ P d a t a P d a t a + P g 2 ⋅ 1 2 ] + E x ∼ P g [ log ⁡ P g P d a t a + P g 2 ⋅ 1 2 ] = min ⁡ G K L ( P d a t a ∣ ∣ P d a t a + P g 2 ) + K L ( P g ∣ ∣ P d a t a + P g 2 ) − log ⁡ 4 ≥ − log ⁡ 4 \min_G\max_D V(D,G)=\min_G V(D_G^*,G)\\=\min_GE_{x\sim P_{data}}\Big[\log \frac{P_{data}}{P_{data}+P_g}\Big]+E_{x\sim P_{g}}\Big[\log (1-\frac{P_{data}}{P_{data}+P_g})\Big]\\=\min_GE_{x\sim P_{data}}\Big[\log \frac{P_{data}}{P_{data}+P_g}\Big]+E_{x\sim P_{g}}\Big[\log \frac{P_{g}}{P_{data}+P_g}\Big]\\=\min_GE_{x\sim P_{data}}\Big[\log \frac{P_{data}}{\frac{P_{data}+P_g}2}\cdot\frac12\Big]+E_{x\sim P_{g}}\Big[\log \frac{P_{g}}{\frac{P_{data}+P_g}2}\cdot\frac12\Big]\\=\min_G KL(P_{data}||\frac{P_{data}+P_g}2)+KL(P_g||\frac{P_{data}+P_g}2)-\log 4\\\ge -\log 4 GminDmaxV(D,G)=GminV(DG,G)=GminExPdata[logPdata+PgPdata]+ExPg[log(1Pdata+PgPdata)]=GminExPdata[logPdata+PgPdata]+ExPg[logPdata+PgPg]=GminExPdata[log2Pdata+PgPdata21]+ExPg[log2Pdata+PgPg21]=GminKL(Pdata2Pdata+Pg)+KL(Pg2Pdata+Pg)log4log4
P d a t a = P d a t a + P g 2 = P g P_{data}=\frac{P_{data}+P_g}2=P_g Pdata=2Pdata+Pg=Pg时,“=”成立。    \;
此时, P g ∗ = P d a t a , D g ∗ = 1 2 P^*_g=P_{data},D^*_g=\frac12 Pg=Pdata,Dg=21

   \;
   \;
   \;
下一章传送门:白板推导系列笔记(三十二)-变分自编码器

猜你喜欢

转载自blog.csdn.net/qq_41485273/article/details/112397006
今日推荐