|
123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208 |
- # -*- coding: utf-8 -*-
- # ---
- # jupyter:
- # jupytext_format_version: '1.2'
- # kernelspec:
- # display_name: Python 3
- # language: python
- # name: python3
- # language_info:
- # codemirror_mode:
- # name: ipython
- # version: 3
- # file_extension: .py
- # mimetype: text/x-python
- # name: python
- # nbconvert_exporter: python
- # pygments_lexer: ipython3
- # version: 3.5.2
- # ---
-
- # # 变分自动编码器
- # 变分编码器是自动编码器的升级版本,其结构跟自动编码器是类似的,也由编码器和解码器构成。
- #
- # 回忆一下,自动编码器有个问题,就是并不能任意生成图片,因为我们没有办法自己去构造隐藏向量,需要通过一张图片输入编码我们才知道得到的隐含向量是什么,这时我们就可以通过变分自动编码器来解决这个问题。
- #
- # 其实原理特别简单,只需要在编码过程给它增加一些限制,迫使其生成的隐含向量能够粗略的遵循一个标准正态分布,这就是其与一般的自动编码器最大的不同。
- #
- # 这样我们生成一张新图片就很简单了,我们只需要给它一个标准正态分布的随机隐含向量,这样通过解码器就能够生成我们想要的图片,而不需要给它一张原始图片先编码。
- #
- # 一般来讲,我们通过 encoder 得到的隐含向量并不是一个标准的正态分布,为了衡量两种分布的相似程度,我们使用 KL divergence,利用其来表示隐含向量与标准正态分布之间差异的 loss,另外一个 loss 仍然使用生成图片与原图片的均方误差来表示。
- #
- # KL divergence 的公式如下
- #
- # $$
- # D{KL} (P || Q) = \int_{-\infty}^{\infty} p(x) \log \frac{p(x)}{q(x)} dx
- # $$
-
- # ## 重参数
- # 为了避免计算 KL divergence 中的积分,我们使用重参数的技巧,不是每次产生一个隐含向量,而是生成两个向量,一个表示均值,一个表示标准差,这里我们默认编码之后的隐含向量服从一个正态分布的之后,就可以用一个标准正态分布先乘上标准差再加上均值来合成这个正态分布,最后 loss 就是希望这个生成的正态分布能够符合一个标准正态分布,也就是希望均值为 0,方差为 1
- #
- # 所以标准的变分自动编码器如下
- #
- # 
-
- # 所以最后我们可以将我们的 loss 定义为下面的函数,由均方误差和 KL divergence 求和得到一个总的 loss
- #
- # ```
- # def loss_function(recon_x, x, mu, logvar):
- # """
- # recon_x: generating images
- # x: origin images
- # mu: latent mean
- # logvar: latent log variance
- # """
- # MSE = reconstruction_function(recon_x, x)
- # # loss = 0.5 * sum(1 + log(sigma^2) - mu^2 - sigma^2)
- # KLD_element = mu.pow(2).add_(logvar.exp()).mul_(-1).add_(1).add_(logvar)
- # KLD = torch.sum(KLD_element).mul_(-0.5)
- # # KL divergence
- # return MSE + KLD
- # ```
-
- # 下面我们用 mnist 数据集来简单说明一下变分自动编码器
-
- # + {"ExecuteTime": {"start_time": "2018-01-01T10:41:05.215490Z", "end_time": "2018-01-01T10:41:05.738797Z"}}
- import os
-
- import torch
- from torch.autograd import Variable
- import torch.nn.functional as F
- from torch import nn
- from torch.utils.data import DataLoader
-
- from torchvision.datasets import MNIST
- from torchvision import transforms as tfs
- from torchvision.utils import save_image
-
- # + {"ExecuteTime": {"start_time": "2018-01-01T10:41:05.741302Z", "end_time": "2018-01-01T10:41:05.769643Z"}}
- im_tfs = tfs.Compose([
- tfs.ToTensor(),
- tfs.Normalize([0.5, 0.5, 0.5], [0.5, 0.5, 0.5]) # 标准化
- ])
-
- train_set = MNIST('./mnist', transform=im_tfs)
- train_data = DataLoader(train_set, batch_size=128, shuffle=True)
-
- # + {"ExecuteTime": {"start_time": "2018-01-01T10:41:06.306479Z", "end_time": "2018-01-01T10:41:06.397118Z"}}
- class VAE(nn.Module):
- def __init__(self):
- super(VAE, self).__init__()
-
- self.fc1 = nn.Linear(784, 400)
- self.fc21 = nn.Linear(400, 20) # mean
- self.fc22 = nn.Linear(400, 20) # var
- self.fc3 = nn.Linear(20, 400)
- self.fc4 = nn.Linear(400, 784)
-
- def encode(self, x):
- h1 = F.relu(self.fc1(x))
- return self.fc21(h1), self.fc22(h1)
-
- def reparametrize(self, mu, logvar):
- std = logvar.mul(0.5).exp_()
- eps = torch.FloatTensor(std.size()).normal_()
- if torch.cuda.is_available():
- eps = Variable(eps.cuda())
- else:
- eps = Variable(eps)
- return eps.mul(std).add_(mu)
-
- def decode(self, z):
- h3 = F.relu(self.fc3(z))
- return F.tanh(self.fc4(h3))
-
- def forward(self, x):
- mu, logvar = self.encode(x) # 编码
- z = self.reparametrize(mu, logvar) # 重新参数化成正态分布
- return self.decode(z), mu, logvar # 解码,同时输出均值方差
-
- # + {"ExecuteTime": {"start_time": "2018-01-01T10:41:06.430817Z", "end_time": "2018-01-01T10:41:10.056600Z"}}
- net = VAE() # 实例化网络
- if torch.cuda.is_available():
- net = net.cuda()
-
- # + {"ExecuteTime": {"start_time": "2018-01-01T10:41:10.059597Z", "end_time": "2018-01-01T10:41:10.409900Z"}}
- x, _ = train_set[0]
- x = x.view(x.shape[0], -1)
- if torch.cuda.is_available():
- x = x.cuda()
- x = Variable(x)
- _, mu, var = net(x)
-
- # + {"ExecuteTime": {"start_time": "2018-01-01T10:41:29.749178Z", "end_time": "2018-01-01T10:41:29.753678Z"}}
- print(mu)
- # -
-
- # 可以看到,对于输入,网络可以输出隐含变量的均值和方差,这里的均值方差还没有训练
- #
- # 下面开始训练
-
- # + {"ExecuteTime": {"start_time": "2018-01-01T10:13:54.530108Z", "end_time": "2018-01-01T10:13:54.560436Z"}}
- reconstruction_function = nn.MSELoss(size_average=False)
-
- def loss_function(recon_x, x, mu, logvar):
- """
- recon_x: generating images
- x: origin images
- mu: latent mean
- logvar: latent log variance
- """
- MSE = reconstruction_function(recon_x, x)
- # loss = 0.5 * sum(1 + log(sigma^2) - mu^2 - sigma^2)
- KLD_element = mu.pow(2).add_(logvar.exp()).mul_(-1).add_(1).add_(logvar)
- KLD = torch.sum(KLD_element).mul_(-0.5)
- # KL divergence
- return MSE + KLD
-
- optimizer = torch.optim.Adam(net.parameters(), lr=1e-3)
-
- def to_img(x):
- '''
- 定义一个函数将最后的结果转换回图片
- '''
- x = 0.5 * (x + 1.)
- x = x.clamp(0, 1)
- x = x.view(x.shape[0], 1, 28, 28)
- return x
-
- # + {"ExecuteTime": {"start_time": "2018-01-01T10:13:54.562533Z", "end_time": "2018-01-01T10:35:01.115877Z"}}
- for e in range(100):
- for im, _ in train_data:
- im = im.view(im.shape[0], -1)
- im = Variable(im)
- if torch.cuda.is_available():
- im = im.cuda()
- recon_im, mu, logvar = net(im)
- loss = loss_function(recon_im, im, mu, logvar) / im.shape[0] # 将 loss 平均
- optimizer.zero_grad()
- loss.backward()
- optimizer.step()
-
- if (e + 1) % 20 == 0:
- print('epoch: {}, Loss: {:.4f}'.format(e + 1, loss.data[0]))
- save = to_img(recon_im.cpu().data)
- if not os.path.exists('./vae_img'):
- os.mkdir('./vae_img')
- save_image(save, './vae_img/image_{}.png'.format(e + 1))
- # -
-
- # 可以看看使用变分自动编码器得到的结果,可以发现效果比一般的编码器要好很多
- #
- # 
- #
- # 我们可以输出其中的均值看看
-
- # + {"ExecuteTime": {"start_time": "2018-01-01T10:40:36.463332Z", "end_time": "2018-01-01T10:40:36.481622Z"}}
- x, _ = train_set[0]
- x = x.view(x.shape[0], -1)
- if torch.cuda.is_available():
- x = x.cuda()
- x = Variable(x)
- _, mu, _ = net(x)
-
- # + {"ExecuteTime": {"start_time": "2018-01-01T10:40:37.485127Z", "end_time": "2018-01-01T10:40:37.490484Z"}}
- print(mu)
- # -
-
- # 变分自动编码器虽然比一般的自动编码器效果要好,而且也限制了其输出的编码 (code) 的概率分布,但是它仍然是通过直接计算生成图片和原始图片的均方误差来生成 loss,这个方式并不好,在下一章生成对抗网络中,我们会讲一讲这种方式计算 loss 的局限性,然后会介绍一种新的训练办法,就是通过生成对抗的训练方式来训练网络而不是直接比较两张图片的每个像素点的均方误差
|