Research On Extended Image Data Set Based On Deep Convolution Generative Adversarial Network
Research On Extended Image Data Set Based On Deep Convolution Generative Adversarial Network
Authorized licensed use limited to: University of Canberra. Downloaded on June 06,2020 at 16:15:18 UTC from IEEE Xplore. Restrictions apply.
high quality handwritten digital image [6]. Zhang Jialiang variable of unknown distribution.
and others used the Variational Auto Encoder (VAE) to extract
According to the cross-entropy loss, the following loss
the real image information, and regarded the reconstructed
functions can be constructed:
image as a false image to improve the BEGAN [7]. Zhang
Guanghua and others first used GAN to generate astronomical V(D,G)=𝐸𝐸𝑥𝑥 ∼𝑃𝑃𝑑𝑑𝑑𝑑𝑑𝑑𝑑𝑑 (𝑥𝑥) [ln 𝐷𝐷(𝑥𝑥)]+ 𝐸𝐸𝑧𝑧 ∼𝑃𝑃𝑧𝑧 (𝑧𝑧) [ln(1 − 𝐷𝐷(𝐺𝐺(𝑧𝑧)))]( 1 )
images, and gave the GAN model structure of astronomical The specific steps of the experiment are as follows:
image generation [8].
Define the generator. The generator has two convolution
GAN does not have any structural restrictions on G and D, networks, and tanh is used as the activation function.
while Deep Convolutional Generative Adversarial Networks
Define the discriminator. The input image of the
(DCGAN) [9] uses a special structure in D and G to effectively
discriminator is 28 × 28, 64 convolution kernels and 5 × 5
model pictures.
kernels. 64 5 × 5 convolution kernels are used to extract
III. EXPERIMENT different features, and the default step size is 1. There are four
In this paper, DCGAN is used to generate images in convolution networks. Tanh is used as the activation function of
TensorFlow. The data set is MNIST data set. Suppose that the the first, second and third layers, and sigmoid is used as the
real picture data used for training is x, and the distribution of activation function of the fourth layer.
picture data is 𝑃𝑃𝑑𝑑𝑑𝑑𝑑𝑑𝑑𝑑 (𝑥𝑥). Previously, it was said that G can Define the training function. Select 128 for
"generate pictures", and the real G can learn the real data Batch_size, 100 for epoch, 0.0005 for learning rate, and SGD
distribution 𝑃𝑃𝑑𝑑𝑑𝑑𝑑𝑑𝑑𝑑 (𝑥𝑥) . The distribution of noise Z is set as random gradient descent algorithm for loss function.
𝑃𝑃𝑧𝑧 (𝑧𝑧), 𝑃𝑃𝑧𝑧 (𝑧𝑧) is known, and 𝑃𝑃𝑑𝑑𝑑𝑑𝑑𝑑𝑑𝑑 (𝑥𝑥) is unknown. In an ideal
Select part of the epoch output image experimental results,
case, the distribution of 𝐺𝐺(𝑧𝑧) should be as close as possible to
as follows:
𝑃𝑃𝑑𝑑𝑑𝑑𝑑𝑑𝑑𝑑 (𝑥𝑥). G maps the Z variable of known distribution to the X
48
Authorized licensed use limited to: University of Canberra. Downloaded on June 06,2020 at 16:15:18 UTC from IEEE Xplore. Restrictions apply.
Fig. 3. 40_epoch Fig. 4. 99_epoch
1 0.0313 0.9919
12 0.0373 0.9923
29 0.0422 0.9928
38 0.0274 0.9995
The trained model is saved and tested. The test result is Convolution Neural Network is used to test the new MNIST
0.9935. data set. The results show that the test accuracy of the
Convolution Neural Network to the new data set is equivalent
IV. SUMMARY
to that of the Convolution Neural Network on the original
In this paper, the concept and some applications of Deep MNIST data set, that is to say, the Deep Convolution Generative
Convolution Generative Adversarial Network and Generative Adversarial Network is proved. The new data generated can be
Adversarial Network are introduced. In this paper, the original used in practice. This experiment also proves that the Deep
MNIST data set is expanded to generate a new data set by using Convolution Generative Adversarial Network can use the
the Deep Convolution Generative Adversarial Network, and the
49
Authorized licensed use limited to: University of Canberra. Downloaded on June 06,2020 at 16:15:18 UTC from IEEE Xplore. Restrictions apply.
original data set to generate new data to meet the requirements
in the case of few data sets, which solves the problem of finding
data difficult and data less can not be effectively trained.
REFERENCES
[1] Waibel, A., 1987. Phoneme recognition using time-delay neural
networks. Meeting of the Institute of Electrical, Information and
Communication Engineers (IEICE). Tokyo, Japan.
[2] Miyo. Research on radar image enhancement technology based on
generated countermeasure network [D]. Beijing University of Posts and
telecommunications, 2019.
[3] Wang gege, Guo Tao, Li Guiyang. Multi layer perceptron deep
convolution generation countermeasure network [J]. Computer science,
2019,46(09):243-249.
[4] Lu Ping, Dong Husheng. Face image generation based on deep
convolution generation network [J]. Modern computer, 2019 (21):56-
58+64.
[5] Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David
Warde-Farley, Sherjil Ozair, Aaron Courville, Yoshua Bengio, 2014,
Generative Adversarial Networks.
[6] Chen Lifang, Lu Guojun. MNIST handwritten digital image generation
method based on VAE and Gan fusion network [J]. Journal of Langfang
Normal University (NATURAL SCIENCE EDITION), 2019, 19 (02):
25-29.
[7] zhangjialiang, he Zhipeng, Wang Yuanyuan, Zeng Bing, Shen Yi, Jia
Yu. An improved method of fast image generation based on limited data
set [J]. Communication technology, 2019, 52 (05): 1119-1125.
[8] Zhang Guanghua, Wang Fubao, Duan Weijun. Astronomical image
generation based on GaN [J]. Journal of Northwest University of
technology, 2019,37(02):315-322.
[9] Alec Radford, Luke Metz, Soumith Chintala, 2015, Unsupervised
Representation Learning with Deep Convolutional Generative
Adversarial Networks.
50
Authorized licensed use limited to: University of Canberra. Downloaded on June 06,2020 at 16:15:18 UTC from IEEE Xplore. Restrictions apply.