0% found this document useful (0 votes)
55 views

Model-Driven Deep Learning

This document discusses model-driven deep learning, which combines modeling-based and deep learning-based approaches. It begins by noting the successes of deep learning but also its limitations like lack of theoretical foundations and difficulty designing network topology. It then introduces the model-driven approach, which uses models based on task objectives and domain knowledge to guide network design. The rest of the document outlines the process of model-driven deep learning, including constructing a model family based on the task, designing an algorithm family to solve the model family, and using this approach to determine network topology in a way that is theoretically sound while retaining deep learning's ability to learn from data.

Uploaded by

jun zhao
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
55 views

Model-Driven Deep Learning

This document discusses model-driven deep learning, which combines modeling-based and deep learning-based approaches. It begins by noting the successes of deep learning but also its limitations like lack of theoretical foundations and difficulty designing network topology. It then introduces the model-driven approach, which uses models based on task objectives and domain knowledge to guide network design. The rest of the document outlines the process of model-driven deep learning, including constructing a model family based on the task, designing an algorithm family to solve the model family, and using this approach to determine network topology in a way that is theoretically sound while retaining deep learning's ability to learn from data.

Uploaded by

jun zhao
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 3

PERSPECTIVES

INFORMATION SCIENCE

Special Topic: Machine Learning


Model-driven deep-learning
Zongben Xu∗ and Jian Sun∗

Deep learning has been widely recog-


nized as the representative advances of Model family Algorithm family Deep network
machine learning or artificial intelligence
in general nowadays [1,2]. This can be
attributed to the recent breakthroughs Figure 1. Model-driven deep-learning approach.
made by deep learning on a series of
challenging applications. A deep-learning
approach improves the accuracy rate institutions can obviously match these as a data-driven approach because it
of face recognition to be higher than requirements. uses a standard network architecture
99%, beating the human level [3]. For With the arrival of the big data era, as a black box, heavily relying on huge
speech recognition and machine trans- data requirements are gradually no longer data to train the black box. In contrast,
lation, deep learning is approaching the an obstacle (at least for many areas), but the model-driven approach here refers
performance level of a simultaneous in- the determination of network topology to the method using a model (e.g. a
terpreter [4]. For the game of ‘go’, is still a bottleneck. This is mainly due loss function) constructed based on
it successfully beats the human world to the lack of theoretical understandings the objective, physical mechanism and
champion [5]. For diagnosis of some spe- of the relationship between the network domain knowledge for a specific task. A
cific diseases, it has matched the level of topology and performance. In the cur- prominent feature of the model-driven
medium or senior professional physicians rent state, the selection of network topol- approach is that, when the model is
[6]. Until now, it has been hard to find ogy is still an engineering practice instead sufficiently accurate, the solution can be
areas in which the deep-learning tech- of scientific research, leading to the fact generally expected to be optimal, and the
nique has not been tried in their respec- that most of the existing deep-learning minimization algorithm is commonly
tive tasks. approaches lack theoretical foundations. deterministic. A fatal flaw of the model-
One can observe that these break- The difficulties in network design and its driven approach lies in the difficulty in
throughs always take place in large IT interpretation, and a lack of understand- accurately modeling for a specific task
companies or specialized R&D institutes, ing in its generalization ability are the in real applications, and sometimes the
such as Google, Microsoft, Facebook, common limitations of the deep-learning pursuit of accurate modeling is a luxury
etc. This is because deep-learning ap- approach. These limitations may prevent expectation. In recent years, we have
plications require some prerequisites, its widespread use in the trends of ‘stan- studied and implemented a series of
such as a huge volume of labeled data, dardization, commercialization’ of ma- model-driven deep-learning methods
sufficient computational resources and chine learning and artificial intelligence [7–10] combining the modeling-based
the engineering experiences in determin- technology. and deep-learning-based approaches,
ing the network topology, including the A natural question is whether we can which showed their feasibilities and
number of layers, number of neurons design network topology with theoretical effectiveness in real applications.
per layer and non-linear transforms of foundations, and make the network Given a specific task, the basic proce-
neurons. Due to these prerequisites, it structure explainable and predictable. dures of our model-driven deep-learning
requires sufficient knowledge and en- We believe that it is possible to provide a method are shown in Fig. 1 and explained
gineering experience in neural network positive answer to this question through as follows:
design, and takes a long time in accu- combing the model-driven approach (1) A model family is first constructed
mulating and labeling data. Professional and data-driven deep-learning approach. based on the task backgrounds (e.g.
IT companies and specialized R&D Here we take the deep-learning approach objective, physical mechanism and


C The Author(s) 2017. Published by Oxford University Press on behalf of China Science Publishing & Media Ltd. All rights reserved. For permissions, please e-mail:

[email protected]
PERSPECTIVES Xu and Sun 23

Sampling data Reconstructed


M(n-1) M(n) M(n+1)
in k-space MR image

X(1) X(n-1) C(n-1) S(n-1) X(n) C(n) S(n) X(n+1) C(n+1) S(n+1)

Stage n

Figure 2. Topology of ADMM-Net [7]: given under-sampled k-space data, it outputs the reconstructed MRI image after T stages of processing.

prior knowledge). The model family using sub-sampled k-space data lower results using this model-driven deep-
is a family of functions with a large than the Nyquist rate. The model family learning method.
set of unknown parameters, amount- is defined as: The above model-driven deep-
ing to the hypothesis space in ma-  learning approach obviously retains the
1
chine learning. Differently from the x̂ = arg min Ax − y 22 advantages (i.e. determinacy and theo-
accurate model in the model-driven x 2 retical soundness) of the model-driven
approach, this model family only L  approach, and avoids the requirement
provides a very rough and broad def- + λl g (D l x) , (1) for accurate modeling. It also retains
l =1
inition of the solution space. It has the powerful learning ability of the
the advantage of a model-driven ap- where A = P F is the measurement deep-learning approach, and overcomes
proach but greatly reduces the pres- matrix, P is the sampling matrix, F is the difficulties in network topology
sure of accurate modeling. the Fourier transform matrix, Dl is lin-
ear transform for convolution, g (·) is the selection. This makes the deep-learning
(2) An algorithm family is then designed regularization function, λl is the regular- approach designable and predictable,
for solving the model family and the ization parameter and L is the number and it balances well versatility and
convergence theory of the algorithm of linear transforms. All the parameters of pertinence in real applications.
family is established. The algorithm (D l , g , λl , L) are unknown and reflect We point out that the model-driven
family refers to the algorithm with the uncertainty in modeling (notice that approach and data-driven approach are
unknown parameters for minimiz- these parameters are known and fixed not opposed to each other. If the model
ing the model family in the func- in traditional CS-MRI models). Accord-
ing to the ADMM (Alternating Direction is accurate, it provides the essential
tion space. The convergence theory Method of Multipliers) method, the al- description of the problem solutions,
should include the convergence rate gorithm family for solving the model fam- from which infinite ideal samples can
estimation and the constraints on the ily can be designated as: be generated, and vice versa: when the
parameters that assure the conver- ⎧ T sufficient samples are provided, the

T −1
⎪x = F P P + l ρl F Dl Dl F
⎪ (n) T T
gence of the algorithm family. model of the problem is fully (but in


(3) The algorithm family is unfolded to ⎪

 discretized form) represented. This is


a deep network with which parame- ⎨× P T y + l ρl F DlT zl(n−1) + βl(n−1)
,
the essential reason for the effectiveness
ter learning is performed as a deep- ⎪ 

⎪ (n)
= (n)
+ β
(n−1) λl of the model-driven deep-learning
⎪ l
⎪ z S D l x l ; ρl
learning approach. The depth of the ⎪
⎪ approach.

⎩ (n)
network is determined by the con- βl = βl
(n−1)
+ ηl Dl x − zl (n)
(n)
Please refer to [2,8] for the previ-
vergence rate estimation of the al- (2)
ous investigations of the model-driven
gorithm family. The parameter space deep-learning approach. The recent ad-
where S(·) is a non-linear transform re-
of the deep network is determined vances can be found in [7,9–11]. Most
lating to g (·). According to the ADMM
by the parameter constraints. All the of these successful applications lie in the
convergence theory, this algorithm is
parameters of the algorithm family inverse problems in imaging sciences,
linearly convergent. By unfolding the
are learnable. In this way, the topol- for which there exists domain knowl-
algorithm family to a deep network, we
ogy of the deep network is deter- edge that can be well modeled in the
design an ADMM-Net composed of T
mined by the algorithm family, and model family. We believe that this model-
successive stages, as shown in Fig. 2. Each
the deep network can be trained driven deep-learning approach can be
stage consists of a reconstruction layer
through back-propagation. widely applied to the applications where
(R), a convolution layer (C), a non-
Taking [7] as an example, we apply linear transform layer (Z) and a mul- we can design the model family by
the above model-driven deep-learning tiplier update layer (M). We learn the incorporating domain knowledge and
approach to compressive sensing mag- parameters of (S, D l , λl , ρl , ηl ) using then the deep architecture can be corre-
netic resonance imaging (CS-MRI), i.e. a back-propagation algorithm. In [7], spondingly designed following the above
recovering the high-quality MR image we reported the state-of-the-art CS-MRI procedures.
24 Natl Sci Rev, 2018, Vol. 5, No. 1 PERSPECTIVES

Zongben Xu∗ and Jian Sun∗ 2. Gregor K and LeCun Y. ICML 2010. 9. Sun J and Tappen M. IEEE T Image Process 2013;
Xi’an International Academy for Mathematics & 3. Schroff F, Kalenichenko D and Philbin J. CVPR 22: 402–8.
Mathematical Technology, Xi’an Jiaotong 2015. 10. Sun J, Sun J and Xu Z. IEEE T Image Process 2015;
University, China 4. Yonghui W, Schuster M and Zhifeng Chen et al. 24: 4148–59.
∗ Corresponding authors. arXiv:1609.08144, 2016. 11. Sprechmann P, Bronstein AM and Sapiro G. IEEE
E-mails: [email protected]; 5. Silver D, Aja Huang and Chris J. Maddison et al. TPAMI 2015; 37: 1821–33.
[email protected] Nature 2016; 529: 484–9.
6. Gulshan V, Peng L and Coram M et al. Jama 2016; National Science Review
316: 2402–10. 5: 22–24, 2018
REFERENCES 7. Yang Y, Sun J and Li H et al. NIPS 2016. doi: 10.1093/nsr/nwx099
8. Sun J and Tappen M. CVPR 2011. Advance access publication 25 August 2017
1. LeCun Y, Bengio Y and Hinton G. Nature 2015;
521: 436–44.

COMPUTER SCIENCE

Special Topic: Machine Learning


Deep learning for natural language processing: advantages
and challenges
Hang Li

INTRODUCTION however. It might not be sufficient for performances in the first four tasks and
inference and decision making, which becomes the state-of-the-art technology
Deep learning refers to machine learning are essential for complex problems like for the tasks (e.g. [1–8]).
technologies for learning and utilizing multi-turn dialogue. Furthermore, how Table 2 shows the performances of
‘deep’ artificial neural networks, such to combine symbolic processing and neu- example problems in which deep learn-
as deep neural networks (DNN), con- ral processing, how to deal with the long ing has surpassed traditional approaches.
volutional neural networks (CNN) tail phenomenon, etc. are also challenges Among all the NLP problems, progress
and recurrent neural networks (RNN). of deep learning for natural language in machine translation is particularly
Recently, deep learning has been suc- processing. remarkable. Neural machine translation,
cessfully applied to natural language i.e. machine translation using deep
processing and significant progress learning, has significantly outperformed
has been made. This paper summa- traditional statistical machine translation.
rizes the recent advancement of deep PROGRESS IN NATURAL The state-of-the art neural translation
learning for natural language process- LANGUAGE PROCESSING systems employ sequence-to-sequence
ing and discusses its advantages and In our view, there are five major tasks in learning models comprising RNNs
challenges. natural language processing, namely clas- [4–6].
We think that there are five major sification, matching, translation, struc- Deep learning has also, for the first
tasks in natural language processing, in- tured prediction and the sequential deci- time, made certain applications possi-
cluding classification, matching, transla- sion process. Most of the problems in nat- ble. For example, deep learning has been
tion, structured prediction and the se- ural language processing can be formal- successfully applied to image retrieval
quential decision process. For the first ized as these five tasks, as summarized in (also known as text to image), in which
four tasks, it is found that the deep learn- Table 1. In the tasks, words, phrases, sen- query and image are first transformed
ing approach has outperformed or sig- tences, paragraphs and even documents into vector representations with CNNs,
nificantly outperformed the traditional are usually viewed as a sequence of tokens the representations are matched with
approaches. (strings) and treated similarly, although DNN and the relevance of the image to
End-to-end training and represen- they have different complexities. In fact, the query is calculated [3]. Deep learn-
tation learning are the key features of sentences are the most widely used pro- ing is also employed in generation-based
deep learning that make it a powerful cessing units. natural language dialogue, in which, given
tool for natural language process- It has been observed recently an utterance, the system automatically
ing. Deep learning is not almighty, that deep learning can enhance the generates a response and the model

You might also like