Artificial Intelligence Image Recognition Method Based On Convolutional Neural Network Algorithm
Artificial Intelligence Image Recognition Method Based On Convolutional Neural Network Algorithm
Received June 7, 2020, accepted June 21, 2020, date of publication June 30, 2020, date of current version July 21, 2020.
Digital Object Identifier 10.1109/ACCESS.2020.3006097
ABSTRACT As an algorithm with excellent performance, convolutional neural network has been widely
used in the field of image processing and achieved good results by relying on its own local receptive fields,
weight sharing, pooling, and sparse connections. In order to improve the convergence speed and recognition
accuracy of the convolutional neural network algorithm, this paper proposes a new convolutional neural
network algorithm. First, a recurrent neural network is introduced into the convolutional neural network, and
the deep features of the image are learned in parallel using the convolutional neural network and the recurrent
neural network. Secondly, according to the idea of ResNet’s skip convolution layer, a new residual module
ShortCut3-ResNet is constructed. Then, a dual optimization model is established to realize the integrated
optimization of the convolution and full connection process. Finally, the effects of various parameters of
the convolutional neural network on the network performance are analyzed through simulation experiments,
and the optimal network parameters of the convolutional neural network are finally set. Experimental results
show that the convolutional neural network algorithm proposed in this paper can learn the diverse features of
the image, and improve the accuracy of feature extraction and image recognition ability of the convolutional
neural network.
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
VOLUME 8, 2020 125731
Y. Tian: Artificial Intelligence Image Recognition Method Based on Convolutional Neural Network Algorithm
In previous years, most methods used shallow structural convolutional neural network (Scatter-Net) based on wavelet
models to process data, and the structural model had at most transform, which uses wavelet transform to extract image
one or two layers of nonlinear features. The most represen- high-frequency information hierarchically instead of the
tative shallow structures are Gaussian Mixture Model [10], parameter learning process, which shows good performance
K-means clustering [11], Support Vector Machine [12], and in image recognition and classification tasks. Hu et al. [32]
Logistic Regression [13]. Convolutional neural network can proposed a model PCAnet that initializes the CNN con-
extract the connection and spatial information between its volution layer parameters by extracting the features of the
layers from the image, and can express the relevant charac- image principal component, and has achieved good results in
teristics inside the image [14], [15]. The image recognition image recognition tasks. Sadr et al. [34] pointed out that the
process based on deep learning is mainly to input the image structure of the convolutional neural network itself is the main
into the neural network, and use the deep learning forward factor for the network to extract multi-level and multi-scale
propagation and back propagation error algorithms to min- features. Guo et al. [35] combined a convolutional neural
imize the loss function. After updating the weights, a better network and a recurrent neural network to propose a new deep
recognition model is obtained. Then use this model to identify learning structure. The convolutional neural network learns
new images. the shallow features of the original image and uses it as the
In practical applications, CNN has been used in many input of the recurrent neural network. Using the recurrent
visual pattern recognition systems. Morvan et al. [16] pro- neural network to learn the high-level features, it achieves
posed the CNN structure LeNet for handwritten digit recog- a good recognition rate in color-depth image recognition.
nition. Convolutional neural networks are also used for facial Bernal et al. [36] added the input of the convolutional neu-
recognition and facial localization [17]. Parmar et al. [18] ral network based on the [35], and proposed a multi-scale
used convolutional neural networks to detect faces and facial convolutional recurrent neural network. After local contrast
expressions. Grossberg [19] used shunt suppression convo- normalization and sampling, it was used as the input of the
lutional neural networks for the detection of eye and face recurrent neural network to extract more abstract high-level
images. Nguyen et al. [20] used convolutional neural net- features.
works to detect text images. Wang et al. [21] achieved Although there are many image recognition algorithms
victory in the large-scale image recognition competition based on convolutional neural networks, and the recognition
using the classic AlexNet model, and successfully reduced effect is very good. However, many recognition algorithms
the false recognition rate to 17%. After the success of are now based on a specific database to design the depth
AlexNet, the researchers proposed other network models, and level of the network. Through continuous exploration,
such as VGGNet [22], GoogleNet [23], and ResNet [24]. the best parameters and optimization algorithms are found.
AlexNet uses the ReLU function to replace the activation The human factor is relatively large, and there is no system-
function commonly used in traditional neural networks [25]. atic theory to affect the recognition effect of the convolutional
Compared with sigmoid function and Tanh function, ReLU nerve. Especially when classifying and recognizing natural
has no exponential calculation, the calculation amount is images, the selection of the initial state parameters of the
small and the network will not be saturated, and because of convolutional neural network and the optimization algorithm
the linear unsaturated form of ReLU, it will speed up the will have a great impact on the network training. If the
network convergence speed. To solve the problem that the selection is not good, the network will not work, or it may fall
ReLU function is not derivable at the origin, Montanelli and into the local minimal, under-fitting, over-fitting, and many
Du [26] proposed a sparse ReLU function. Wang et al. [27] other problems [37], [38].
proposed parametric ReLU, and proved through experiments In order to improve the ability of the convolutional
that the PReLU function has achieved good results in the neural network to classify and recognize two-dimensional
big data classification task. In recent years, the study of images, speed up the convergence of the algorithm, reduce
convolutional neural networks has been inseparable from the number of iterations and shorten the training period,
transfer learning. Transfer learning is a method that can use and achieve good classification results, this paper pro-
the knowledge that has been learned to solve problems in poses a new convolutional neural network algorithm. First,
new fields [28]. Feng et al. [29] used large image datasets a recurrent neural network is introduced into the convolu-
to perform pre-training on convolutional neural networks, tional neural network, and the deep features of the image
and then trained and tested the trained networks on the are learned in parallel using the convolutional neural net-
image datasets to be classified. Compared with the traditional work and the recurrent neural network. Secondly, accord-
method of training the network directly on the target data set, ing to the idea of ResNet’s skip convolution layer, a new
the image recognition rate of this method is greatly improved. residual module ShortCut3-ResNet is constructed. Finally,
Zhang et al. [30] proposed an algorithm based on hierarchi- a dual optimization model is established to achieve inte-
cal sparse coding (HSC), which extracts features through grated optimization of the convolution and full connection
hierarchical pooling and sparse processing, and obtains process.
good results in handwriting recognition and multi-class Specifically, the technical contributions of our paper can
object recognition. Carroll et al. [31] proposed a scattering be concluded as follows:
the network will automatically learn the features without the neural network can adapt to more problems that are com-
manual selection of features, which avoids time and effort. plex. Commonly used activation functions include sigmoid
Suppose there is an image with a size of 4 × 4, a 2 × 2 convo- function, Tanh function, ReLU function, and leaky ReLU
lution kernel is used, and the sliding step of the convolution function. The formula of sigmoid function is shown as for-
kernel is one, then the convolution operation process can be mula (1). The Tanh function can be expressed as formula (2).
expressed as shown in Figure 2. The ReLU function is described by formula as shown in equa-
tion (3). The improvement of the gradient problem caused by
the negative ReLU input results in the Leaky ReLU function,
whose function formula can be expressed as formula (4).
2) POOLING LAYER
The pooling layer is also a very common type of hidden layer
used in CNN. Because the local features in the image are
related, pooling the image can greatly reduce the amount of
calculation but will not lose the main features of the image.
Assuming that the size of the image is 4 × 4, a convolution
kernel of 2 × 2 size is used, and the sliding step of the con-
volution kernel is set to two, the common pooling methods,
maximum pooling, average pooling and random pooling the
process is shown in Figure 3.
C. ResNet INTRODUCTION
FIGURE 3. The process of pooling operation.
Since AlexNet, some cutting-edge CNNs are constantly deep-
ening [44], [45]. For example, AlexNet has five convolutional
3) ACTIVE LAYER layers; VGGNet and GoogleNet have 19 and 22 convolutional
Another important hidden layer of CNN is the active layer. layers, respectively. However, it is not feasible to increase the
When solving more complex problems, the activation func- network depth by directly superimposing layers. Assuming
tion adds nonlinear factors into the neural network so that, there is a shallow network, multiple maps are stacked on
this network to form a deep network. In theory, the training Among them, the variable xi is the input vector. The vari-
error of this deep network will not be higher than that of able xi+1 is the output vector. The variable F represents the
the shallow network. However, the experimental results show residual mapping that the residual structure needs to learn
that such a deep network cannot be found. After doing many and can be expressed as F = W2 σ (W1 x). The variable
experiments, it is also found that the deep network shows f represents the activation function operation. The variable
higher training error than the shallow network on the same h(xi ) = xi is a cross-layer connection.
data set, as shown in Figure 5. Nevertheless, in CNN, as the network deepens, the number
of convolution kernels also increases. Therefore, for the case
where these two dimensions do not match, a special con-
volution kernel Ws can be used to perform the convolution
operation to ensure the matching of the two dimensions.
Assuming that the input data is xi, first calculate the mean
value of the data, as shown in equation (9).
m
1X
µB = xi (9)
m
i=1
In the network structure described above, the first layer of the two channels of F(x) and x. The number of the two
uses a 3 × 3 convolutional layer. Nevertheless, the last layer channels at the solid line connection is the same and can be
no longer uses the fully connected layer in VGGNet, but bor- added directly. The dotted line indicates that the two channels
rows from the global average-pooling layer in network. This have different numbers, and the dimension of x needs to be
can effectively avoid the problems of excessive parameter adjusted by a convolution operation, that is, the convolution
quantity, low training speed, and easy overfitting in the fully kernel Ws in formula (8).
connected layer.
When the value of n is three, we can get the number of C. DOUBLE OPTIMIZATION
hidden layers as 18, plus the first layer of convolutional layer The design principle of the convolution optimization model is
and the last layer of global average pooling layer, we can get to realize the weight optimization of the convolution kernel.
a 20-layer ShortCut3-ResNet. Table 1 describes the original We can learn the data set weights and bias parameters from
ResNet and ShortCut3-ResNet network configurations for the small data blocks to obtain a sparse feature matrix. The
20-layer network. Based on this structure, 6n + 2 (n >= 3) convolution kernel is initialized by convolution coefficient
layer network structure can be obtained, as shown in Figure 9. control. Let matrix X is the sample data set. A is the base
matrix used to transform X from sample space to feature
TABLE 1. Two different ResNet structures. space. Matrix S is the feature table of the data set. Setting
the objective function J (A, S) and assigning the initial value
of S, the process of reducing the objective function through
iteration is the process of optimizing S. Giving a good initial
value of S can avoid the situation of poor convergence during
the iteration process, and at the same time obtain a faster
convergence and more optimized results. The process of S
initialization and feature update is as follows:
S = G(W T X ) (13)
Sc0 = Sc /||Ac || (14)
Among them, the variable W T is a random orthogonal
matrix. Using W T to extract sample X through weighted
transformation, the initial value of matrix S is generated. The
variable Sc represents the c-th feature matrix of the matrix S.
The variable Ac represents the corresponding base matrix of
Sc in matrix A. Let variable M be m × n matrix, then:
X n
m X
||M ||k = ( |mij |k )1/k (15)
i=1 j=1
Step 3 Use the gradient descent method to obtain the The optimization process of fully connected parameters
local minimum value of the objective function J (A, S) again is similar to convolution optimization, and the process is as
according to the value of variable S 0 . Moreover, obtain the follows.
value of variable A0 at this time, α is the step size, and Let Opt3 be a fully connected parameter matrix randomly
control the change amount of gradient direction each time. 2 . According to
initialized using parameters ncag and nf limg
The calculation process is the operation process and classification results of the convo-
lutional neural network, the parameter settings of the fully
∂J (A, S)
A0 = A − α (18) connected layer are affected by factors such as the number
∂S of iterations of the convolutional neural network. According
Step 4 Use A ’and S’ to replace A and S respectively, repeat to the interpolation principle, the constructor optimizes the
step 2 ∼ step 4. parameters of the full connection, and ρ is the optimization
As the number of iterations increases, the objective func- coefficient:
tion will gradually decrease in the reverse direction of the
gradient until the gradient vector approaches zero, and the ρ = ncag (w − εk ε−1 )/2 (23)
objective function no longer decreases or the change can be Among them, the variable ω is a factor that affects the
ignored. Randomly sample the obtained feature matrix S to optimization coefficient, which is determined by factors such
construct the initial weight of the convolution kernel. At the as the amount of data processed by the single-layer percep-
same time, the dynamically determined value µ is used in the tron and the number of classifications. Let variable θ2 be the
convolutional neural network to replace the constant µ0 as correction error term,
the convolution coefficient, to realize the optimization of the q
convolution kernel. ε
w = λ( nf limg
2 −n
cag ) − k + θ2 (24)
Suppose the convolutional neural network contains a total
ε−1
of k convolutional layers, the size of each convolution kernel X
λ=k+ i (25)
is lker ×lker . The input image size of the convolutional layer is
i=0
a matrix of variable limg × limg . The input and output feature
maps or images are nin and nout respectively. Let Opt4 be the parameter matrix of the last fully connected
Let the matrix Opt1 be the feature matrix S when the layer, and the optimized fully connected layer parameter
objective function obtains the minimum value after multiple expression is:
iterations. Use the convolution coefficient to optimize the
ρ
s
convolution kernel, analyze the original convolution result Opt4 = 2 × Opt3 (26)
2
ncag + nf limg
through the dichotomy, and construct the function expression
according to the interpolation principle. The dynamic convo-
lution coefficient µ is expressed as follows D. CONVOLUTIONAL NEURAL NETWORK TRAINING
PROCESS
2
nin nout limg
µ= + θ1 (19) Convolutional neural network is essentially a mapping from
2k input to output, which can learn many features that do not
Among them, the variable θ1 is the correction error term. require any precise mathematical expression between input
The expressions of the number of parameters required for the and output, and realize the mapping between input and out-
input data and output data corresponding to the convolution put. Because the network performs supervised learning, its
kernel are as follows: sample set is a vector pair of input vectors and ideal output
vectors. The network training process is shown in Figure 10.
2
fin = nin lker (20)
2
fout = nout lker (21)
same time, set the learning speed and the corresponding variable δjl+1 from left to right and from top to bottom. More-
accuracy control parameters. over, corresponding to the value of the variable xil multiplied,
Because each convolutional layer has its threshold that can after accumulation to obtain the derivative of the convolution
be trained, the weight of each convolution kernel is a learn- kernel function. After the derivative of the weight value of the
able parameter. Therefore, the focus of CNN weight update convolution kernel is obtained, it is updated to the Kij position
is the update of convolution kernel weight and convolution corresponding to the original convolution kernel.
layer threshold. To update the bias of the threshold of the convolutional
layer, simply add the error sensitivity of the j-th feature map
1) UPDATE OF WEIGHT AND THRESHOLD OF NETWORK of the above l + 1 layer.
The reverse adjustment of neural network is the idea of gradi- The threshold update derivative is calculated as shown in
ent descent. For the network weight update in the algorithm, equation (28).
the adjustment of parameters always proceeds in the direction
of error reduction. ∂loss X l+1
= (δj ) (28)
∂kij u,v
u,v
memory, four NVIDIA Geforce TITAN X 12GB GPU and TABLE 3. Experimental results of different sampling methods.
Ubuntu14.04 operating system.
The evaluation criteria of the experiment in this paper are
the test accuracy and the size of the model finally generated
by training.
2) SAMPLING METHOD
TABLE 4. Classification results of several pooling methods on the
After the feature map passes through the convolutional layer, CIFAR-10 database.
the dimension is generally very large, which can easily cause
the dimension disaster. Therefore, each convolutional layer
in the convolutional neural network will be connected to a
sampling layer to down sample the feature map to reduce the
dimension of the feature map and reduce the computational
complexity. Therefore, the sampling layer is also an essential
part of the convolutional neural network structure. Choosing
the appropriate sampling method will greatly improve the
performance of the convolutional neural network. By sam-
pling the feature map, the convolutional neural network can
tolerate small deformations. Common sampling methods are
maximum sampling, mean sampling, and random sampling.
Choosing the most suitable sampling method can improve the
recognition efficiency and accuracy of the network. In this
paper, three different network models were constructed using
three sampling methods. Except for the different sampling
methods, the remaining parameters are all the same. The
experimental results are shown in Table 3 below.
Three experiments were carried out for each model, and
the average value of the three experiments was selected as
the final recognition result. It can be seen from the above FIGURE 13. Classification results of different pooling sizes on the
table that the recognition effect of maximum sampling and CIFAR-10 database.
cause overfitting, and a larger pooling size will increase the The mean square error curve varies with the training batch
error due to too much noise in down sampling. at different iteration times. When the training times are once,
twice and three times, the convergence curve of each algo-
C. MODEL CONVERGENCE TRAINING rithm is shown in Figure 15.
The training set samples are used for training, and the initial
standard deviation of the initial weights of the network is
0.01, with a Gaussian distribution with an average value
of zero. Suppose the number of sample iterations is 3000,
the initial learning rate of the weight parameter is 0.001, and
the momentum factor is 0.9. The training results are shown
in Figure 14.
FIGURE 15. Convergence curve of each algorithm when training times are
different.
TABLE 5. Performance comparison of different algorithms on the [5] L. Wen, K. Zhou, and S. Yang, ‘‘A shape-based clustering method for
CIFAR-10 dataset. pattern recognition of residential electricity consumption,’’ J. Cleaner
Prod., vol. 212, pp. 475–488, Mar. 2019.
[6] T. Zan, Z. Liu, H. Wang, M. Wang, and X. Gao, ‘‘Control chart pattern
recognition using the convolutional neural network,’’ J. Intell. Manuf.,
vol. 31, no. 3, pp. 703–716, Mar. 2020.
[7] J. Yu, X. Zheng, and S. Wang, ‘‘A deep autoencoder feature learning
method for process pattern recognition,’’ J. Process Control, vol. 79,
pp. 1–15, Jul. 2019.
[8] D. Freire-Obregón, F. Narducci, S. Barra, and M. Castrillón-Santana,
‘‘Deep learning for source camera identification on mobile devices,’’ Pat-
tern Recognit. Lett., vol. 126, pp. 86–91, Sep. 2019.
[9] X. Zhe, S. Chen, and H. Yan, ‘‘Directional statistics-based deep metric
learning for image classification and retrieval,’’ Pattern Recognit., vol. 93,
pp. 113–123, Sep. 2019.
[10] A. O’Hagan, T. B. Murphy, L. Scrucca, and I. C. Gormley, ‘‘Investigation
of parameter uncertainty in clustering using a Gaussian mixture model via
random forest is inferior to the traditional CNN in network jackknife, bootstrap and weighted likelihood bootstrap,’’ Comput. Statist.,
vol. 34, no. 4, pp. 1779–1813, Dec. 2019.
performance, and the algorithm proposed in this paper is [11] S. Wang, A. Gittens, and M. W. Mahoney, ‘‘Scalable kernel K-means
higher than the single structure algorithm in network test clustering with Nyström approximation: Relative-error bounds,’’ J. Mach.
accuracy. The AlexNet network uses a three-layer fully con- Learn. Res., vol. 20, no. 1, pp. 431–479, 2019.
[12] F. Karimi, S. Sultana, A. Shirzadi Babakan, and S. Suthaharan,
nected layer, so the network has a large number of training ‘‘An enhanced support vector machine model for urban expansion predic-
parameters, resulting in more storage resources occupied by tion,’’ Comput., Environ. Urban Syst., vol. 75, pp. 61–75, May 2019.
the final training model. In a comprehensive comparison, [13] P. Sur and E. J. Candès, ‘‘A modern maximum-likelihood theory for high-
dimensional logistic regression,’’ Proc. Nat. Acad. Sci. USA, vol. 116,
the algorithm proposed in this paper is more diversified when
no. 29, pp. 14516–14525, Jul. 2019.
extracting features and the test accuracy of the network has [14] D. Zhu, F. Zhang, S. Wang, Y. Wang, X. Cheng, Z. Huang, and Y. Liu,
been improved. After combining the ultra-lightweight net- ‘‘Understanding place characteristics in geographic contexts through graph
work structure, the amount of parameters is appropriately convolutional neural networks,’’ Ann. Amer. Assoc. Geographers, vol. 110,
no. 2, pp. 408–420, Mar. 2020.
reduced. [15] A. Jati and P. Georgiou, ‘‘Neural predictive coding using convolutional
neural networks toward unsupervised learning of speaker characteris-
tics,’’ IEEE/ACM Trans. Audio, Speech, Lang. Process., vol. 27, no. 10,
V. CONCLUSION
pp. 1577–1589, Oct. 2019.
In order to improve the ability of the convolutional neural [16] M. Morvan, D. Arangalage, G. Franck, F. Perez, L. Cattan-Levy,
network to classify and recognize two-dimensional images I. Codogno, M.-P. Jacob-Lenet, C. Deschildre, C. Choqueux, G. Even,
and speed up the convergence of the algorithm, this paper pro- J.-B. Michel, M. Bäck, D. Messika-Zeitoun, A. Nicoletti, G. Caligiuri,
and J. Laschet, ‘‘Relationship of iron deposition to calcium deposition
poses a new convolutional network algorithm. First, a recur- in human aortic valve leaflets,’’ J. Amer. College Cardiol., vol. 73, no. 9,
rent neural network is introduced into the convolutional neu- pp. 1043–1054, Mar. 2019.
ral network, and the deep features of the image are learned [17] M. F. Hansen, M. L. Smith, L. N. Smith, M. G. Salter, E. M. Baxter,
M. Farish, and B. Grieve, ‘‘Towards on-farm pig face recognition using
in parallel using the convolutional neural network and the convolutional neural networks,’’ Comput. Ind., vol. 98, pp. 145–152,
recurrent neural network. Not only can we use convolutional Jun. 2018.
neural networks to learn high-level features, but also recursive [18] K. Parmar, H. Kher, and M. Gandhi, ‘‘Facial expression recognition using
convolutional neural network,’’ J. Open Source Develop., vol. 6, no. 1,
neural networks to learn the combined features of low-level pp. 18–27, 2019.
features. Secondly, according to ResNet’s idea of skipping [19] S. Grossberg, ‘‘The resonant brain: How attentive conscious seeing reg-
convolutional layers, we construct a new residual module ulates action sequences that interact with attentive cognitive learning,
recognition, and prediction,’’ Attention, Perception, Psychophys., vol. 81,
ShortCut3-ResNet. Finally, the convolutional layer and the no. 7, pp. 2237–2264, Oct. 2019.
full connection process are optimized. Experiments show [20] H. T. Nguyen, C. T. Nguyen, T. Ino, B. Indurkhya, and M. Nakagawa,
that the proposed convolutional neural network algorithm can ‘‘Text-independent writer identification using convolutional neural net-
work,’’ Pattern Recognit. Lett., vol. 121, pp. 104–112, Apr. 2019.
improve the feature extraction accuracy and image recogni- [21] R. Wang, J. Xu, and T. X. Han, ‘‘Object instance detection with pruned
tion ability of convolutional neural network. AlexNet and extended training data,’’ Signal Process., Image Commun.,
vol. 70, pp. 145–156, Feb. 2019.
[22] P. Matlani and M. Shrivastava, ‘‘Hybrid deep VGG-NET convolutional
REFERENCES classifier for video smoke detection,’’ Comput. Model. Eng. Sci., vol. 119,
[1] J. Pan, ‘‘How chinese officials use the Internet to construct their public no. 3, pp. 427–458, 2019.
image,’’ Political Sci. Res. Methods, vol. 7, no. 2, pp. 197–213, Apr. 2019. [23] R. U. Khan, X. Zhang, and R. Kumar, ‘‘Analysis of ResNet and GoogleNet
[2] S. Liansheng, Z. Xiao, H. Chongtian, T. Ailing, and A. Krishna models for malware detection,’’ J. Comput. Virol. Hacking Techn., vol. 15,
Asundi, ‘‘Silhouette-free interference-based multiple-image encryption no. 1, pp. 29–37, Mar. 2019.
using cascaded fractional Fourier transforms,’’ Opt. Lasers Eng., vol. 113, [24] D. McNeely-White, J. R. Beveridge, and B. A. Draper, ‘‘Inception and
pp. 29–37, Feb. 2019. ResNet features are (almost) equivalent,’’ Cognit. Syst. Res., vol. 59,
[3] X. Zhu, Z. Li, X.-Y. Zhang, P. Li, Z. Xue, and L. Wang, ‘‘Deep convo- pp. 312–318, Jan. 2020.
lutional representations and kernel extreme learning machines for image [25] S. Scardapane, S. Van Vaerenbergh, S. Totaro, and A. Uncini, ‘‘Kafnets:
classification,’’ Multimedia Tools Appl., vol. 78, no. 20, pp. 29271–29290, Kernel-based non-parametric activation functions for neural networks,’’
Oct. 2019. Neural Netw., vol. 110, pp. 19–32, Feb. 2019.
[4] F. Wang, D. Jiang, H. Wen, and H. Song, ‘‘AdaBoost-based security level [26] H. Montanelli and Q. Du, ‘‘New error bounds for deep ReLU networks
classification of mobile intelligent terminals,’’ J. Supercomput., vol. 75, using sparse grids,’’ SIAM J. Math. Data Sci., vol. 1, no. 1, pp. 78–92,
no. 11, pp. 7460–7478, Nov. 2019. Jan. 2019.
[27] S.-H. Wang, K. Muhammad, J. Hong, A. K. Sangaiah, and Y.-D. Zhang, [39] W. Jing, T. Jiang, X. Zhang, and L. Zhu, ‘‘The optimisation of speech
‘‘Alcoholism identification via convolutional neural network based on recognition based on convolutional neural network,’’ Int. J. High Perform.
parametric ReLU, dropout, and batch normalization,’’ Neural Comput. Comput. Netw., vol. 13, no. 2, pp. 222–231, 2019.
Appl., vol. 32, no. 3, pp. 665–680, Feb. 2020. [40] S. Bacchi, L. Oakden-Rayner, T. Zerner, T. Kleinig, S. Patel, and J. Jannes,
[28] U. Cote-Allard, C. L. Fall, A. Drouin, A. Campeau-Lecours, C. Gosselin, ‘‘Deep learning natural language processing successfully predicts the cere-
K. Glette, F. Laviolette, and B. Gosselin, ‘‘Deep learning for electromyo- brovascular cause of transient ischemic attack-like presentations,’’ Stroke,
graphic hand gesture signal classification using transfer learning,’’ IEEE vol. 50, no. 3, pp. 758–760, Mar. 2019.
Trans. Neural Syst. Rehabil. Eng., vol. 27, no. 4, pp. 760–771, Apr. 2019. [41] Y. Zhang, X. Zhang, H. Sun, Z. Fan, and X. Zhong, ‘‘Portable brain-
[29] S. Feng, H. Zhou, and H. Dong, ‘‘Using deep neural network with small computer interface based on novel convolutional neural network,’’ Comput.
dataset to predict material defects,’’ Mater. Des., vol. 162, pp. 300–310, Biol. Med., vol. 107, pp. 248–256, Apr. 2019.
Jan. 2019. [42] C. Xu, J. Yang, H. Lai, J. Gao, L. Shen, and S. Yan, ‘‘UP-CNN: Un-
[30] Y. Zhang, Y. Qu, C. Li, Y. Lei, and J. Fan, ‘‘Ontology-driven hierarchi- pooling augmented convolutional neural network,’’ Pattern Recognit. Lett.,
cal sparse coding for large-scale image classification,’’ Neurocomputing, vol. 119, pp. 34–40, Mar. 2019.
vol. 360, pp. 209–219, Sep. 2019. [43] Y.-D. Zhang, Z. Dong, X. Chen, W. Jia, S. Du, K. Muhammad, and
[31] E. L. Carroll, R. Gallego, M. A. Sewell, J. Zeldis, L. Ranjard, H. A. Ross, S.-H. Wang, ‘‘Image based fruit category classification by 13-layer deep
L. K. Tooman, R. O’Rorke, R. D. Newcomb, and R. Constantine, ‘‘Multi- convolutional neural network and data augmentation,’’ Multimedia Tools
locus DNA metabarcoding of zooplankton communities and scat reveal Appl., vol. 78, no. 3, pp. 3613–3632, Feb. 2019.
trophic interactions of a generalist predator,’’ Sci. Rep., vol. 9, no. 1, [44] Y. Wang, Y. Sun, Z. Liu, S. E. Sarma, M. M. Bronstein, and J. M. Solomon,
pp. 1–14, Dec. 2019. ‘‘Dynamic graph CNN for learning on point clouds,’’ ACM Trans. Graph.,
[32] F. Hu, M. Zhou, P. Yan, K. Bian, and R. Dai, ‘‘PCANet: A common solution vol. 38, no. 5, pp. 1–12, Nov. 2019.
for laser-induced fluorescence spectral classification,’’ IEEE Access, vol. 7, [45] S. H. S. Basha, S. R. Dubey, V. Pulabaigari, and S. Mukherjee, ‘‘Impact of
pp. 107129–107141, 2019. fully connected layers on performance of convolutional neural networks for
[33] Y. Wang, G. Wang, C. Chen, and Z. Pan, ‘‘Multi-scale dilated convolution image classification,’’ Neurocomputing, vol. 378, pp. 112–119, Feb. 2020.
of convolutional neural network for image denoising,’’ Multimedia Tools [46] L. Su, L. Ma, N. Qin, D. Huang, and A. H. Kemp, ‘‘Fault diagnosis of high-
Appl., vol. 78, no. 14, pp. 19945–19960, Jul. 2019. speed train bogie by residual-squeeze net,’’ IEEE Trans. Ind. Informat.,
[34] H. Sadr, M. M. Pedram, and M. Teshnehlab, ‘‘A robust sentiment analysis vol. 15, no. 7, pp. 3856–3863, Jul. 2019.
method based on sequential combination of convolutional and recursive
neural networks,’’ Neural Process. Lett., vol. 50, no. 3, pp. 2745–2761,
Dec. 2019.
[35] Z. Guo, X. Lv, L. Yu, Z. Zhang, and S. Tian, ‘‘Identification of hepatitis
b using Raman spectroscopy combined with gated recurrent unit and
multiscale fusion convolutional neural network,’’ Spectrosc. Lett., vol. 53,
no. 4, pp. 277–288, Apr. 2020.
[36] J. Bernal, K. Kushibar, D. S. Asfaw, S. Valverde, A. Oliver, R. Martí, and
YOUHUI TIAN received the master’s degree
X. Lladó, ‘‘Deep convolutional neural networks for brain image analysis
on magnetic resonance imaging: A review,’’ Artif. Intell. Med., vol. 95,
from the Heilongjiang University of Science and
pp. 64–81, Apr. 2019. Technology, in 2013. He is currently a Senior
[37] A. Kamilaris and F. X. Prenafeta-Boldú, ‘‘A review of the use of convo- Engineer with the Jiangsu Vocational Institute of
lutional neural networks in agriculture,’’ J. Agricult. Sci., vol. 156, no. 3, Commerce. His research interests include network
pp. 312–322, Apr. 2018. technology and information systems.
[38] F. Samadi, G. Akbarizadeh, and H. Kaabi, ‘‘Change detection in SAR
images using deep belief network: A new training approach based on mor-
phological images,’’ IET Image Process., vol. 13, no. 12, pp. 2255–2264,
Oct. 2019.