PyTorch实例:基于自编码器的图形去噪
Posted candyren
tags:
篇首语:本文由小常识网(cha138.com)小编为大家整理,主要介绍了PyTorch实例:基于自编码器的图形去噪相关的知识,希望对你有一定的参考价值。
去噪自编码器模拟人类视觉机制能够自动忍受图像的噪声来识别图片。自编码器的目标是要学习一个近似的恒等函数,使得输出近似等于输入。去噪自编码器采用随机的部分带噪输入来解决恒等函数问题,自编码器能够获得输入的良好表征,该表征使得自编码器能进行去噪或恢复。
下面是代码:
#加载库和配置参数 #去噪自编码器 import torch import torch.nn as nn import torch.utils as utils from torch.autograd import Variable import torchvision.datasets as dset import torchvision.transforms as transforms import numpy as np import matplotlib.pyplot as plt #配置参数 torch.manual_seed(1) n_epoch=200 batch_size=100 learning_rate=0.002 #下载图片库训练集 mnist_train=dset.MNIST("./",train=True,transform=transforms.ToTensor(),target_transform=None,download=True) train_loader=torch.utils.data.DataLoader(dataset=mnist_train,batch_size=batch_size,shuffle=True) #Encoder和Decoder模型设置 class Encoder(nn.Module): def __init__(self): super(Encoder,self).__init__() self.layer1=nn.Sequential( nn.Conv2d(1,32,3,padding=1),#batch*32*28*28 nn.ReLU(), nn.BatchNorm2d(32), nn.Conv2d(32,32,3,padding=1),#batch*32*28*28 nn.ReLU(), nn.BatchNorm2d(32), nn.Conv2d(32,64,3,padding=1),#batch*64*28*28 nn.ReLU(), nn.BatchNorm2d(64), nn.Conv2d(64,64,3,padding=1),#batch*64*28*28 nn.ReLU(), nn.BatchNorm2d(64), nn.MaxPool2d(2,2)#batch*64*14*14 ) self.layer2=nn.Sequential( nn.Conv2d(64,128,3,padding=1),#batch*128*14*14 nn.ReLU(), nn.BatchNorm2d(128), nn.Conv2d(128,128,3,padding=1),#batch*128*14*14 nn.ReLU(), nn.BatchNorm2d(128), nn.MaxPool2d(2,2), nn.Conv2d(128,256,3,padding=1),#batch*256*14*14 nn.ReLU() ) def forward(self,x): out=self.layer1(x) out=self.layer2(out) out=out.view(batch_size,-1) return out encoder=Encoder().cuda() #decoder设置 class Decoder(nn.Module): def __init__(self): super(Decoder,self).__init__() self.layer1=nn.Sequential( nn.ConvTranspose2d(256,128,3,2,1,1),#batch*128*14*14 nn.ReLU(), nn.BatchNorm2d(128), nn.ConvTranspose2d(128,128,3,1,1),#batch*128*14*14 nn.ReLU(), nn.BatchNorm2d(128), nn.ConvTranspose2d(128,64,3,1,1),#batch*64*14*14 nn.ReLU(), nn.BatchNorm2d(64), nn.ConvTranspose2d(64,64,3,1,1),#batch*64*14*14 nn.ReLU(), nn.BatchNorm2d(64) ) self.layer2=nn.Sequential( nn.ConvTranspose2d(64,32,3,1,1),#batch*32*14*14 nn.ReLU(), nn.BatchNorm2d(32), nn.ConvTranspose2d(32,32,3,1,1),#batch*32*14*14 nn.ReLU(), nn.BatchNorm2d(32), nn.ConvTranspose2d(32,1,3,2,1,1),#batch*1*28*28 nn.ReLU() ) def forward(self,x): out=x.view(batch_size,256,7,7) out=self.layer1(out) out=self.layer2(out) return out decoder=Decoder().cuda() ###Loss 函数和优化器 parameters=list(encoder.parameters())+list(decoder.parameters()) loss_func=nn.MSELoss() optimizer=torch.optim.Adam(parameters,lr=learning_rate) ###自编码器训练 #添加噪声 noise=torch.rand(batch_size,1,28,28) for I in range(n_epoch): for image,label in train_loader: image_n=torch.mul(image+0.25,0.1*noise) image=Variable(image).cuda() image_n=Variable(image_n).cuda() optimizer.zero_grad() output=encoder(image_n) output=decoder(output) loss=loss_func(output,image) loss.backward() optimizer.step() break print(‘epoch[{}/{}],loss:{:.4f}‘.format(I+1,n_epoch,loss.item())) ####带噪图片和去噪图片对比 img=image[0].cpu() input_img=image_n[0].cpu() output_img=output[0].cpu() origin=img.data.numpy() inp=input_img.data.numpy() out=output_img.data.numpy() plt.figure(‘denoising autoencoder‘) plt.subplot(131) plt.imshow(origin[0],cmap=‘gray‘) plt.subplot(132) plt.imshow(inp[0],cmap=‘gray‘) plt.subplot(133) plt.imshow(out[0],cmap=‘gray‘) plt.show() print(label[0])
以上是关于PyTorch实例:基于自编码器的图形去噪的主要内容,如果未能解决你的问题,请参考以下文章
基于词向量使用Pytorch常规自编码器对句子进行向量表示与降维
pytorch: 图像恢复问题的代码实现详解(derain,dehaze,deblur,denoise等通用)
TensorFlow实现去噪自编码器(Denoising Autoencoder)