0% found this document useful (0 votes)
14 views21 pages

2018 - A Brief Review On Multi-Task Learning - Thung - Wee - Multimedia Tools and Applications

This document provides a brief review of multi-task learning (MTL), a machine learning approach that optimizes multiple related tasks simultaneously, enhancing performance in various applications such as natural language processing and biomedical imaging. It discusses the motivation behind MTL, compares different algorithms, and highlights its effectiveness in addressing data scarcity by leveraging shared information among tasks. The paper also explores MTL's relationship with other learning methods and offers insights into its formulation and application in deep learning contexts.

Uploaded by

yangkunkuo
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
14 views21 pages

2018 - A Brief Review On Multi-Task Learning - Thung - Wee - Multimedia Tools and Applications

This document provides a brief review of multi-task learning (MTL), a machine learning approach that optimizes multiple related tasks simultaneously, enhancing performance in various applications such as natural language processing and biomedical imaging. It discusses the motivation behind MTL, compares different algorithms, and highlights its effectiveness in addressing data scarcity by leveraging shared information among tasks. The paper also explores MTL's relationship with other learning methods and offers insights into its formulation and application in deep learning contexts.

Uploaded by

yangkunkuo
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 21

Multimedia Tools and Applications (2018) 77: 29705–29725

https://doi.org/10.1007/s11042-018-6463-x

A brief review on multi-task learning

Kim-Han Thung1 · Chong-Yaw Wee2

Received: 1 August 2017 / Revised: 4 July 2018 / Accepted: 24 July 2018 /


Published online: 8 August 2018
© Springer Science+Business Media, LLC, part of Springer Nature 2018

Abstract
Multi-task learning (MTL), which optimizes multiple related learning tasks at the same
time, has been widely used in various applications, including natural language process-
ing, speech recognition, computer vision, multimedia data processing, biomedical imaging,
socio-biological data analysis, multi-modality data analysis, etc. MTL sometimes is also
referred to as joint learning, and is closely related to other machine learning subfields like
multi-class learning, transfer learning, and learning with auxiliary tasks, to name a few.
In this paper, we provide a brief review on this topic, discuss the motivation behind this
machine learning method, compare various MTL algorithms, review MTL methods for
incomplete data, and discuss its application in deep learning. We aim to provide the readers
with a simple way to understand MTL without too many complicated equations, and to help
the readers to apply MTL in their applications.

Keywords Multi-task learning · MTL · Transfer learning · Joint learning ·


Multi-class learning · Learning with auxiliary tasks

1 Introduction

1.1 What is a task?

A task is generally referred to the learning of an output target using a single input source. If
the input source consists of a single variable (or feature), we will have a univariate analysis,
if the input source consists of multiple variables (or features), we will have a multivariate
analysis. In this sense, “multiple tasks” could mean the learning of multiple output targets
using a single input source, or the learning of single output target using multiple input

 Chong-Yaw Wee
[email protected]

Kim-Han Thung
[email protected]

1 Department of Radiology, University of North Carolina, Chapel Hill, NC, USA


2 Department of Biomedical Engineering, National University of Singapore, Singapore, Singapore
29706 Multimed Tools Appl (2018) 77:29705–29725

sources, or a mixture of both. Depending on the definition of “multiple tasks”, the multi-
task learning (MTL) could have different objective functions, as we will demonstrate in the
following subsections.

1.2 What is multi-task learning (MTL) and why it is useful?

When the original data representation is high dimensional and the number of examples
provided to solve a regression or classification problem is limited, any learning algorithm
which does not use any sort of prior knowledge will perform poorly due to lack of data
to reliably estimate the model parameters. This issue is particularly crucial for applications
such as medical image analysis since it needs more manual labor to label data instances.
Multi-task learning (MTL) [5, 19, 71], as one type of machine learning method that aims to
solve multiple tasks simultaneously, can be a good recipe by exploiting useful information
from other related learning tasks to help alleviate this data scarcity problem. Rich Caruana
[8] has summarized the goal of MTL succinctly: “MTL is an approach to inductive transfer
that improves generalization by using the domain information contained in the training
signals of related tasks as an inductive bias. It does this by learning tasks in parallel while
using a shared low dimensional representation; what is learned for each task can help other
tasks be learned better”. The basic assumption of MTL is that all the tasks in learning, or at
least a subset of them, are associated with each other, and thus the shared information among
different tasks can lead to better learning performance if all the tasks are learned jointly,
comparing to learn them independently. In other words, it assumes that the learning of one
task can improve the learning of the other tasks. This is generally achieved by learning all
the tasks jointly, utilizing the correlated information among different tasks (which we will
describe in more details in the following sections) to improve the learning of each task.
Although MTL is particularly helpful if the tasks share significant commonalities, it has
also been shown to be beneficial for learning unrelated tasks [67].

1.3 Some application examples of MTL

We describe some application examples of MTL in this subsection. MTL has been used to
tackle feature selection problem that involves multi source data (please refer to Section 2.1.1
for the formulation). In this case, not all attributes (or features) in multi-source data are
useful for classification or regression problem. When the features derived from different
data sources are related in some ways (e.g., when they are originated from the same regions
of interest and thus correspond to each other), then it is better to select the features from all
the data sources jointly by using a joint selection regularizer. Some of the regularizers (or
constraints) that have been introduced include joint sparsity, low-rank, graph sparse coding,
graph self-representation, and so on [32, 99, 105, 110, 111, 115]. The recent studies show
that the inclusion of these additional regularizations during joint feature selection (via MTL)
can improve the performance of their classification models if compared with the model that
selects features from each data source individually.
MTL has also been used to deal with neurodegenerative disease diagnosis problem. For
example, we can use structural Magnetic Resonance Imaging (sMRI) data to predict the val-
ues for different kind of clinical scores and the diagnostic label of a subject. For Alzheimer’s
disease (AD) studies, the clinical scores that are usually used to grade the healthiness
and functionality of a human brain include the Mini-Mental State Examination (MMSE),
Dementia Rating Scale (DRS), Alzheimer’s Disease Assessment Scale (ADAS), etc. In this
case, the prediction of one of the target output using sMRI data is a learning task. As all the
Multimed Tools Appl (2018) 77:29705–29725 29707

target outputs (e.g., clinical scores and diagnostic label) are related, learning all the tasks
together (using MTL strategy) would most probably give us a better prediction results than
the results obtained from learning all the tasks independently (please refer to Section 2.1.2
for the formulation) [73, 84, 113].
Furthermore, MTL has also been used in self-driving automation system, as described in
[59]. Using the image acquired from the camera attached to the car, we would like to detect
different objects on the road, e.g., pedestrian, car, road sign, traffic light, etc. We can use
neural network architecture that takes camera image as input, and the object labels (e.g.,
pedestrian, car, etc.) as outputs. When we train this neural network to learn multiple objects
simultaneously, it becomes a MTL problem. In this case, we hope that building a single
neural network to learn multiple objects is better than learning separate neural network
for each object (please refer to Section 4 for more details). In fact, learning a set of tasks
together in neural network has many advantages, e.g., we could have more training samples,
we can learn shared lower-level features (e.g., edge, line, and shape), and the learning of
one task could benefit from the learning from the other tasks.

1.4 Related research subfields

MTL is very similar to inductive transfer learning [63, 85], where there are two types of
tasks, i.e., the primary task, which is the main goal of the study, and the secondary (or
auxiliary) task, which is associated with or related to the main task, but is not the main goal
of the study. In transfer learning, it is assumed that the associated secondary task can provide
extra information to the primary task, and improves the generalization of the main task.
However, there is no such distinction for MTL, as it generally treats all the tasks equally.
Therefore, transfer learning can be thought of as a special case of MTL.
Besides, MTL is also related to the problem of learning-to-learn (LTL), which aims to
perform a new task by exploiting the knowledge acquired when solving previous tasks. The
capability of LTL is very similar to the ability of human being that learns from experience
when performing new tasks. Hence, a solution or machine constructed based on LTL would
have major impact in general Artificial Intelligence (AI). Recently, learning nonlinear hier-
archical representations from multiple tasks using multilayer deep neural networks [70]
has emerged as one of the hottest research topics in AI. Researchers have shown improved
results in a number of empirical domains, particularly in computer vision [24]. This success
has increased interest in multi-task representation learning in deep neural networks, as we
will discuss in more details in Section 4. In the following sections, we begin our discussion
by introducing different formulations of MTL algorithms.

2 Formulation of MTL algorithms

The typical formulation for a conventional MTL algorithm [5, 8, 71, 101] is given in the
following form:

M
min L(Xm , ym , wm ) + λReg(W), (1)
W=[w1 w2 ···wM ]
m=1
where Xm ∈ RNm ×D denotes the input matrix of the m-th task, ym ∈ RNm ×1 denotes
the corresponding m-th task output vector, and wm ∈ RD×1 denotes the weight vector
(or regression parameters) for the m-th task that maps Xm to ym , e.g., ym ≈ Xm wm (for
regression problem). The scalars Nm , D, and M denote the number of samples for the m-th
29708 Multimed Tools Appl (2018) 77:29705–29725

task, the number of features for each input matrix, and the number of tasks, respectively.
Note that at this stage, we assume that all the input matrices in {Xm , m = 1, 2, · · · , M} are
having the same dimensionality of features (but can have different number of samples for
each task), and the features of all the tasks are corresponding, so that we can concatenate
all the weight vectors in {wm } together to obtain W = [w1 w2 · · · wM ] (i.e., features in
each row of W are corresponding). Based on the prior knowledge of the data and different
assumptions on the relationship among tasks, we can design different constraints for W,
which is normally implemented in the regularizer of W, denoted by Reg(W). In addition,
λ is the regularization parameter that controls the balance between the loss function (first
term) and the regularizer (second term) in (1). If we set the value of λ to zero, we will
have a solution of W that does not use any assumption or prior knowledge about the task
relatedness, which will most probably only perform well on the training data, but not on
the testing data (i.e., data overfitting). In contrary, if we set the value of λ too high, we may
have a general solution of W that satisfies the task-relatedness assumption given in Reg(W),
but does not perform well for all the prediction tasks. Thus, the regularization parameter
(and any other hyper-parameters) is normally determined via inner cross-validation using
the training samples.
In summary, (1) consists of two terms, i.e., 1) the data fidelity term (the first term in (1)),
which computes how well the target predictions match with the ground truth targets, and
2) the regularization term (i.e., the second term in (1)), which regularizes the weight matrix
W to garner the relationship among different learning tasks, to have better prediction model
for each learning task [3, 6, 8, 18, 84, 116, 117]. In the following subsections, we describe
these two terms in more details.

2.1 Different data fidelity terms of MTL

Let us examine the data fitting term in (1), it seems that there are M number of input
matrices and output vectors for M number of tasks. However, in many real life applications,
some of the input matrices or output vectors are shared among different tasks. Based on
the number of unique input in {Xm } and unique output in {ym }, we categorize the MTL
problems into three special cases, namely the multi-input single-output (MISO), the single-
input multi-output (SIMO), and the multi-input multi-output (MIMO). Figure 1 shows the
general form of MTL and its three special cases based on whether the inputs or outputs are
shared among different tasks. Each special case of MTL would have slightly different form

Fig. 1 a General form of Multi-task learning (MTL), and its special cases, i.e., b multi-input single-output
(MISO), where multiple sets of input are mapped to a same set of output target, c single-input multi-output
(SIMO), where one set of input map is mapped to multiple sets of output target, and d multi-input multi-
output (MIMO), where multiple sets of input are mapped to the same multiple sets of output target. (X: Input
data; Y: Target output)
Multimed Tools Appl (2018) 77:29705–29725 29709

of data fitting term in (1). In addition, the definition of task, and how it relates the input data
to the output target would also be slightly different. We will discuss each of these special
cases in more details in the following subsections.

2.1.1 Multi-input single-output (MISO)

In this case, we have, for example, multiple data sources, and all are mapped to the same
target y. A task is thus defined as the prediction of one data source to a common target y.
For instance, this can happen if we have multi-modality data, and each modality can be used
to predict the target y. Another example in multimedia application is that there are multiple
views of the same object, and all the views are used jointly to predict the label of this object.
The mean square loss formulation for the data fidelity term in this case is given as:


S
L(X, y, W) = Xs ws − y2F , (2)
s

where X = {X1 , X2 , · · · , XS } denotes the set of multi-source (also known as multi-view


or multi-modality) data, Xs ∈ RN×D denotes the s-th data source, y ∈ RN×1 denotes the
output feature vector, and W = [w1 w2 · · · wS ] ∈ RD×S denotes the weight matrix with its
s-th column ws denoting the weight vector corresponding to the mapping of Xs to y. The
N , S and D denote the number of samples, the number of data sources, and the number of
features in each data source, respectively. Note that ŷs = Xs ws is the prediction of y for the
s-th task. Without loss of generality, the bias terms are omitted in this review paper, as they
can easily be incorporated into (2) by adding a column of ones to Xs . Examples of research
works using this setting include [47, 48].
Other than mean square loss function, which is suitable for regression task, the data
fidelity term can also take the form of logistic or hinge loss function for classification task,
which are respectively given below as


S 
N
L(X, y, W) = log(1 + exp(−ŷjs yj )), (3)
s j


S 
N
L(X, y, W) = max(0, 1 − ŷjs yj ), (4)
s j

where ŷjs = xsj ws denotes the prediction of the j -th sample of the s-th data source (xsj ), and
yj ∈ {−1, 1} is the corresponding ground truth label.

2.1.2 Single-input multi-output (SIMO)

In this case, there is only one input (or all the tasks share the same input), and it is used to
predict different types of output target. A task is thus defined as the prediction of the input
matrix X to a target vector. For example, given an image, we would like to predict what are
the contents of this image. The mean square loss function for the data fidelity term in this
case is given as:


C
L(X, Y, W) = Xwc − yc 2F = XW − Y2F , (5)
c
29710 Multimed Tools Appl (2018) 77:29705–29725

where X ∈ RN×D and Y = [y1 · · · yC ] ∈ RN×C are the input feature matrix and output tar-
get matrix, respectively, while W = [w1 w2 · · · wC ] ∈ RD×C are the corresponding weight
matrix, with its c-th column is the weight vector wc that corresponds to the prediction of yc
using X. As all the learning tasks share the same input samples, this multi-task learning for
classification problem is also called multi-class learning [36, 46–48, 60, 75]. The logistic
and hinge loss functions for this setting are similar to (3) and (4), respectively, by replacing
the sum over the number of data sources with the sum over the number of targets.

2.1.3 Multi-input multi-output (MIMO)

In MIMO, multiple input sources are used to predict multiple output targets. A task here
is defined as the prediction of one input source to a single target. For example, this can
happen if we have multiple modalities of data, and each modality can be used to predict
several target labels (i.e., multi class classification). The mean square loss function for the
data fidelity term of a MIMO problem is given as:


S 
C 
S
L(X, Y, W) = Xs ws,c − yc 2F = Xs Ws − Y2F , (6)
s c s

where X = {X1 , X2 , · · · , XS } ∈ RN×D , as defined in Section 2.1.1, Y = [y1 y2 · · · yC ] ∈


RN×C , as defined in Section 2.1.2, W = {W1 , W2 , · · · , WS } ∈ RD×C is the set of all
the weight matrices, and Ws = [ws,1 ws,2 · · · ws,C ] ∈ RD×C is the weight matrix for the
s-th modality data Xs . The derivations of logistic and hinge loss functions for this setting
are straight forward and similar to (3)–(4). In some applications, the multi-source data are
concatenated into a single input data, which simplify this problem to SIMO (Section 2.1.2).
Examples of research works with this setting include [47, 48, 93].

2.2 Different regularizations on weight matrix W of MTL

In this section, we adopt the general notation of loss function as in (1), and focus on
discussing some of the commonly used constraints on the weight matrix W of the MTL
formulation in (1).

2.2.1 MTL with lasso

Assuming there is little correlation among different tasks, and we know that the weight
matrix should be sparse for better interpretability and accuracy of the results, the multi-task
learning with Lasso constraint [74] is given as below via 1 -norm regularization on W:


M
min L(Xm , ym , wm ) + λW1 , (7)
W
m=1

where W = [w1 w2 · · · wM ], and λ is the regularization parameter that controls sparsity in


W. Higher value of λ would correspond to a sparser model, i.e., more zero-value elements
in W. Lasso constraint makes sense when not all the features in the tasks are useful, and
there are only weak associations among tasks. As Lasso constraint does not fully utilize the
task relatedness, it is also usually used in combination with other constraints. Works that
based on the variations of this constraint include [43, 45, 51, 52, 109].
Multimed Tools Appl (2018) 77:29705–29725 29711

2.2.2 MTL with group sparsity constraint

2,1 -norm One direct way to extract the task relatedness information from multiple related
tasks is to constrain all models to share a common set of features. This goal is accomplished
by solving the following 2,1 -norm regularized MTL problem [5, 6]

M
min L(Xm , ym , wm ) + λW2,1 , (8)
W
m=1
D
where W2,1 = i=1 wi 2 , with wi ∈ R
1×M denotes i-th row in W. Note that the

2,1 -norm ( · 2,1 ) enforces row-wise sparsity, i.e., it encourages all-zero-value rows in W
for W2,1 . This is equivalent to joint feature selection for all the learning tasks. Research
works that use (8) include [28, 47, 51, 60, 61, 83], where different optimization algorithms
have been proposed to solve it. Some variates for this MTL formulation include [40], where
weighted 2,1 -norm is proposed, and [66], where feature groups can overlap with each other.

p,q -norm Equation 8 can be generalized to use p,q -norm regularizer to select features, as
given by

M
min L(Xm , ym , wm ) + λWp,q , (9)
W
m=1
where Wp,q = [w1 p · · · wi p · · · wD p ]q . Equation 9 is convex if p > 1, q ≥ 1.
MTL algorithms that use (9) include [45, 58, 76] that uses ∞,1 -norm, and [78] that uses
p,1 -norm.

Capped p,1 -norm In order to obtain a sparser subset of features, Gong et al. [26] proposed
a capped p,1 -norm penalty for wi , as given by

M 
D
min L(Xm , ym , wm ) + λ min(wi p , θ ), (10)
W
m=1 i=1
where θ is a threshold value. The capped p,1 -norm in (10) causes the algorithm to focus
on minimizing the rows of W with p -norm less than θ , thus encouraging sparser solution.
In other words, the smaller the value of θ , the sparser is the solution of (10), and vice versa.
When the value of θ is large enough, the capped p,1 -norm in (10) will become p,1 -norm.

Multi-level lasso Up to this point, the regularization is imposed on the weight matrix W
directly to garner the task relatedness. Another line of research decomposes the weight
matrix into several components and imposes different regularization on them. By using the
right regularizers on these components, we can get similar group sparsity effect induced by
p,q -norm of weight matrix W. For example, Lozano et al. [52] multi-level Lasso, given as

M
min L(Xm , ym , wm ) + λ1 θ1 + λ2 W̃1 , s.t.W = diag(θ)W̃, (11)
W̃,θ m=1

where θ ∈ RD×1 is a non-negative coefficient vector that controls the feature-level group
sparsity. When one of the elements in θ (e.g., θi ) is zero, then the corresponding feature
weight (e.g., wi , i-th row in W) would be zero as well, resulting group sparsity effect on W.
It was shown in [52] that the regularizer in (11) is equivalent to 1, 1 . Wang et al. generalized
2
29712 Multimed Tools Appl (2018) 77:29705–29725

(11) by using p,1 -norm for W̃ (i.e., transpose of W̃) and q -norm for θ . Other variations
of (11) include [31, 36–38, 49].

Structured group lasso On the other hand, instead of assuming all tasks are related using
p,1 -norm on the weight matrix W, some studies, such as [38, 49], utilize the prior knowl-
edge of the data and task relatedness to impose group sparsity on the same group of
  Gv
tasks. More specifically, the regularizer of W is given by D i=1 v∈V λv wi 2 , where
V denotes the number of group, and Gv denotes a set of related tasks in group v. In this
way, only tasks within the same group will have joint sparsity constraint and may share a
common set of relevant input features, while the weakly related tasks (tasks from different
groups) will less likely to be affected by the same set of features. If the grouping of tasks
follows hierarchical tree structure, we will have tree-structured MTL [29, 38]. More specif-
ically, in tree-guided group Lasso [38], V denotes the number of nodes in a tree (assuming
a task is denoted by a node in a tree), and Gv denotes the set of tasks in the subtree rooted at
node v.

Temporal group Lasso In cases where the learning tasks involve time such as longitudinal
study, there could exist temporal relationship among tasks. For example, when we predict
the same taget using data at different time points, or when we use the same data to predict
targets at different time points. In this case, it may be beneficial to add temporal smoothness
constraint on the W to ensure the weight vectors at adjacent time points are consistent [108].
The MTL with temporal smoothness term is given as

M 
M−1
min L(Xm , ym , wm ) + λ1 W2F + λ2 wm − wm+1 22 , (12)
W
m=1 m=1
where the weight vectors of adjacent tasks are assumed to be similar and not to differ too
much. The temporal smoothness term can be rewritten as WH2F , where H ∈ RM×(M−1)
is a matrix with Hij = 1 if i = j, Hij = −1 if i = j + 1, and Hij = 0 otherwise. The
variants of MTL that utilize temporal smoothness prior include sparse fused group Lasso
[107], variant of fused Lasso [104], etc [41, 82, 88, 90].

2.2.3 MTL with low-rank constraint

Instead of using group sparsity constraint, another way to extract task relationship is to
constrain the prediction models from different tasks to share a low-dimensional subspace,
i.e., W is of low rank. This type of MTL can be accomplished by solving the following
surrogate approximation of rank minimization problem [22, 64]:

M
min L(Xm , ym , wm ) + λW∗ , (13)
W
m=1
where  · ∗ denotes the trace norm (or nuclear norm), i.e., the sum of the singular values,

W∗ = min(M,D) σi (W). There are some variations of trace norm regularization such
i=1

as [26, 30], which uses a capped trace regularizer min(M,D)
i=1 min(σi (W), θ ) that penalizes
only small singular values of W that determine the rank of W, and [56] that introduces
spectral k-support norm. There are also other ways to formulate low rank constraint rather
than nuclear norm, for example by finding two low-rank matrices, and let W equal to the
product of these two matrices. We will discuss this case in more details in Section 2.3.2.
Multimed Tools Appl (2018) 77:29705–29725 29713

2.2.4 MTL for unrelated tasks

Some studies have shown that unrelated groups of tasks can be exploited to improve the
learning of certain task [17, 77, 81]. Thus, it is still possible that the MTL can be beneficial
to the main task even if the other tasks (e.g., auxiliary tasks) are not related to the main
task. For example, if we have prior knowledge that the tasks are not related, we can have
regularization that promotes task exclusiveness. In [109], Zhou et al. assume that all tasks
are exclusive and use 1,2 -norm to regularize W, i.e., W21,2 . In [67], Bernardino et al.
introduce a regularization term that penalizes the inner product between the weight matrices
of two different groups of tasks (details in Section 2.3.2).

2.2.5 MTL with graph Laplacian regularization

Other than regularization that utilizes the relationship among tasks, which normally assumes
that the weight parameters for related task are similar, we can also use graph level regular-
ization that utilizes the relationship among samples [114]. Specifically, we can introduce
a graph Laplacian regularization to preserve the local topological relation among samples,
i.e., the local structural information of the sample is preserved after the transformation using
the weight matrix W. The graph Laplacian regularization is given as


N
sij xi W − xj W22 = tr(W X LXW), (14)
i,j

where L = D − S ∈ RN×N denotes the Laplacian matrix, S = [sij ] ∈ RN×N denotes


xi and xj , and D ∈ RN×N is a
the similarity matrix between every pair of sample points 
diagonal matrix with its diagonal element defined as dii = j sij .

2.3 Different regularizers for decomposed weight matrix W of MTL

Recently, more advanced MTL algorithms have been proposed by decomposing the weight
matrix W in (1) into summation or/and product of two or more matrices. For example, in
MTL dirty model [35], the weight matrix is decomposed into the summation of two matrices
(i.e., W = P + Q), each with different constraint to model a more complicated relation-
ship among tasks. In [6], the weight matrix is decomposed into a product of two matrices
(i.e., W = BA), also each with different constraint, to enable feature transformation while
performing MTL. In general, the weight matrix W can be decomposed into

W = P + BA, (15)

where P and A can be seen as the coefficient matrix in the original and the transformed fea-
ture space, respectively, while B is the transformation matrix. In the following, we discuss
these cases in more details.

2.3.1 MTL with W = P + Q

Dirty model MTL based on group sparsity penalization (1 /q -norm regularization) per-
forms well in ideal cases or applications in which the prediction parameters of all the
learning tasks share the same structure, i.e., they share the same features. However, simply
using the 1 /q -norm regularization may not be effective for many practical applications
29714 Multimed Tools Appl (2018) 77:29705–29725

that deal with prediction tasks that may not fall into a single structure. This can be accom-
plished by decomposing W into two components P and Q while solving the dirty multitask
least squares problem [35]

M
min L(Xm , ym , (pm + qm )) + λ1 P1,∞ + λ2 Q1 , (16)
P,Q
m=1

where P = [p1 · · · pM ] ∈ RD×M is the group sparsity component and Q = [q1 · · · qM ] ∈


RD×M is the element-wise sparse component, λ1 controls the group sparsity regularization
on P and λ2 controls the sparsity regularization on Q. In other words, (17) encourages all the
tasks to select the same set of features (via group sparsity component) while each individual
task can still select features which are not common to other tasks, but are discriminative to
its own task. Studies that follow this line of work include [34, 72, 106].

Robust multi-task feature learning In [27], Gong et al. propose a robust multi-task feature
learning, where not only row-sparsity is imposed on W to select common features across
tasks, but column sparsity is also imposed on W to identify outlier tasks that do not share
the same structure as the other tasks. This is acheived by letting W = P + Q, and imposing
2,1 -norm on P and Q , respectively, as given below

M
min L(Xm , ym , (pm + qm )) + λ1 P2,1 + λ2 Q 2,1 . (17)
P,Q
m=1

A mixture of sparse and low-rank model Assumption in Section 2.2.3 that all models
share the common low-dimensional subspace is too restrictive in some applications. To
enable simultaneous learning of incoherent sparse and low-rank patterns, an extension to
(13) has been proposed by decomposing the task model W into two components, i.e., a
sparse component P and a low-rank component Q. This incoherent sparse and low-rank
least squares problem is given as follows [10].

M
min L(Xm , ym , wm ) + λ1 P1 , s.t. W = P + Q, Q∗ ≤ λ2 , (18)
W,P,Q
m=1

where the regularization parameter λ1 controls the sparsity of the sparse component P, while
the λ2 regularization parameter controls the rank of Q. In [12], 21 -norm (rather than 1 -
norm) is imposed on P, assuming the all the tasks share the same discriminative features.

2.3.2 MTL with W = BA

Multi-task feature transformation In our previous discussions, we mainly assume that


tasks are related in the original feature space. However, there are cases where this assump-
tion does not hold, and it might be better to first transform the original features into another
feature space so that the associations among different tasks can be enhanced, and the shared
representations among different tasks can be learned. One formulation example of MTL
with feature transformation, where W = BA, is given as [5, 6]

M
min L(Xm , ym , Bam ) + λA2,1 , s.t. B B = I, (19)
A,B
m=1
Multimed Tools Appl (2018) 77:29705–29725 29715

where B ∈ RD×D is the orthogonal transformation matrix, am ∈ RD×1 is the model param-
eter for the m-th task after feature transformation, and A = [a1 · · · aM ] ∈ RD×M is the
predictor matrix for all the tasks with row-wise sparsity constraint (via 2,1 -norm). Each
column of B is one of the orthogonal transformation vectors that maps the input data (i.e.,
{Xm } into one of the axes in the shared representation feature space, and the orthogonality
of B is used to prevent redundant feature representation. On the other hand, the m-th col-
umn of A (i.e., am ) is the predictor function for the m-th learning task, and the row-wise
sparsity constraint on A is used to exploit the task relatedness by encouraging the selection
of common features in all the learning tasks. Comparing (19) with (8), it can be seen that
both equations are very similar, and the main difference between them is that there is an
additional step of feature transformation in (19), i.e., the input Xm is first transformed to
Xm B before being mapped to ym via am .
It is interesting that (19) is equivalent to the following optimization problem [6]:

M
min L(Xm , ym , wm ) + λtr(W D+ W), s.t. D 0, tr(D) ≤ 1, (20)
W,D
m=1

where tr(·) denotes the trace of a square matrix, D+ denotes the psudoinverse of D, 0
denotes a zero vector or matrix, and D 0 means that D is positive semidefinite. The
optimal solutions of (19) and (20) are related with the following relationship: W = BA,
D = B diag(ai 2 /A2,1 )D 
i=1 B . D can be seen as the feature covariance matrix for
all the learning tasks, and when W is given, the analytic solution for D is given as
1 1
(W W) 2 /tr((W W) 2 ). Note that (20) is convex and easier to solve than (19), thus many
studies that that based on formulation in (19) will transform their formulations into some-
thing similar to (20), so that their problems can be solved more efficiently. Furthermore, the
second term in (20) can be generalized to λtr(W f (W )W).

Multi-task sparse coding If B is an overcomplete dictionary, i.e., B = [b1 · · · bO ] ∈


RD×O , where O > D is the number of atoms in the dictionary, we will have multi-task
sparse coding [55], given as

M
min L(Xm , ym , Bam ) s.t. {am 1 ≤ λ}M
m=1 , {b 2 ≤ 1}j =1 ,
j O
(21)
A,B
m=1

where am ∈ RO×1 is the sparse coding for the m-th task. Note that dictionary B is shared
by all tasks, with each of its column bounded by 2 -norm, and the group sparsity con-
straint in (19) is different from the lasso constraint in (21). Note also that this sparse
coding method is different from the conventional sparse coding strategies in the literature
[39, 53, 62] where each input Xm is a sparse combination of the atoms in the dictio-
nary, with each atom corresponds to a label, and thus the predicted label for the input Xm
is computed as the same sparse combination of the atom labels. The multi-task learning
for this type of sparse coding is beyond the scope of this review and has been explored
in [44].

Multi-task with low-rank structure Instead of using nuclear norm, we can also use the
product of two low-rank matrices [79, 112] to impose low-rank regularization on W. More
specifically, we set W = BA [79, 112], where B ∈ Rr×D , A ∈ Rr×M , and r < min(D, M)
is the hyper-parameter to be tuned as the rank of matrix W. Then, we can impose 2,1 -
29716 Multimed Tools Appl (2018) 77:29705–29725

norm on B or/and A to simultaneously select features and tasks during the optimization. For
example, in [112],

M
min L(Xm , ym , Bam ) + λ1 B 2,1 + λ2 A2,1 , s.t. r < min(D, M), (22)
B,A
m=1

Shared representation with unrelated task In [67], a MTL method that favors shared
low dimensional representations within each group of tasks, while imposing penalty that
encourage orthogonality between representations from two groups of unrelated tasks. The
formulation is given as
M1 M2
min L(Xm , ym , Bam ) + L(Xm , ym , Bâm )
B,A, m=1 m=1


+λ1 [AÂ]2,1 + λ2 A Â2F , s.t. B B = I, (23)
where the first two terms are the data fitting loss functions for two group of tasks, the third
term is used to encourage shared features between two group of tasks, while the fourth
term encourages orthogonality between two group of tasks, assuming they are unrelated.
Equation 23 can be written into the equivalent form similar to (20) so that it can be solved
more easily [67].

2.3.3 MTL with W = P + BA

A general MTL formulation when W = P + BA is given as



M
min L(Xm , ym , (pm + Bam )) + Reg(P, A), s.t. B B = I. (24)
P,B,A
m=1
In brief, this is a combination of regularizer in Section 2.3.1 and Section 2.3.2. In [3], Ando
et al. assume that part of the model parameters of different tasks share a low-rank subspace,
and their second term in (24) only regularizes P. More specifically, they define P2F for the
regularizer and use B as the shared low-rank subspace by multiple tasks. The orthonormal
constraint on B (i.e., BB = I) is to make the subspace non-redundant [3, 11]. Other studies
that follow this line of formulation include [1, 3, 11, 94, 95].

3 MTL for incomplete data

The discussions so far are only limited to MTL algorithms that use complete data. How-
ever, in many applications involving multimodal or longitudinal data, the dataset could be
incomplete, i.e., some of the modalities or data at certain time points could be missing. In
such cases, we can either 1) use only samples with complete data for MTL study, with the
cost of reduced statistical power of analysis due to smaller dataset, or 2) impute the missing
data before performing the MTL study, where the imputation is very much prone to error
for data missing in blocks, or 3) design a MTL method that is applicable to incomplete data,
such as [87, 91].
In MTL for incomplete data [87, 91], the prediction of one modality combination is
treated as one task, and the aim is to learn multiple tasks simultaneously. Specifically, this
method first groups the incomplete dataset into several subsets of complete data, each subset
of data is comprised of different combination of modalities. Then, treating the prediction or
Multimed Tools Appl (2018) 77:29705–29725 29717

classification using each data subset as task, the models in [87, 91] simultaneously learn all
the tasks jointly. This method avoids the drawback of data imputation while allows the use
of all available data for analysis [73].
Figure 2 shows a typical multi-task learning (MTL) framework for dealing with clas-
sification problems involving incomplete multimodal data [91]. The formulation for this
framework is given as:
1  1
m
min L(Xi , yi , Wi ) + λW2,1 , (25)
W=[W1 W2 ···Wm ] m Ni
i=1

where N i , Xi , yi and Wi denote the number of samples, the subset data, the target, and the
model parameters for the i-th modality combination, respectively, λ is regularization param-
eter, while L(·) is the loss function for the data fitting term, e.g., logistic loss function (i.e.,
Ni i  i
j ln(1 + exp(−yj (β ) )xj )) or least square loss function (i.e., X W − y F ). Note
i i i i 2

that the missing data part of Xi is filled with zeros, while the parameter values in Wi that
corresponding to the missing features in Xi are fixed at zeros, as shown in Fig. 2. The for-
mulation given in (25) is a simplified equivalent version of the formulation given in [91] for
better comprehension. Xiang et al. [87] has extended this formulation by allowing overlap
of samples in each modality combination, and introducing extra regularizing parameters for
each modality in the data fitting term, as shown below
1  1
m
min L(Xi , yi , W, α i ) + λRβ (W), s.t. Rα (α i ) < 1 , (26)
W,α m Ni
i=1

where α i = [α i,1 α i,2 · · · α i,S ] is a weight vector for i-th modality combination, with each
of its element denotes the weight for one modality source (assuming there are a total of
S modality sources, the length of α i would be S). Note that unlike the work in [91], a
consistent weight vector W is learned for all the modality combinations, and the different
weight vector for each task (recall that a task is a prediction using samples from a certain
modality combination) is achieved by thecombination of W and α i . For instance, for a
least square loss, we have L(·) = yi − Sj=1 α i,j Xi,j Wj 2F , where Xi,j denotes the j -
th modality data of the i-th modality combination, α i,j denotes the modality source weight
for this part of the data (if that data source is missing, then its value is zero, otherwise,
the weights are learned in formulation (26)), while Wj is the weight vector for the j -th
modality. On the other hand, Rβ and Rα are the regularizers for the weight vector W and
α i , respectively.

Fig. 2 Left: Original classification problem with incomplete multimodal data. Right: Multi-task learning for
incomplete multimodal data [91]. (X: Input data; Y: Target data; M1, M2, M3: Multimodal data; White:
Missing data; Grey: Available data, Green: Learned weights)
29718 Multimed Tools Appl (2018) 77:29705–29725

4 MTL for deep learning

The formulations we discussed so far use hand crafted features and assume there are lin-
ear relationships between the data and target labels. However, in many applications where a
complex data-to-target relationship exists [73], this too conservative assumption may limit
the performance of prediction models. Recently, thanks to its capability in learning latent
representation of the data without any explicit hand-crafted formulation, deep learning tech-
nique or deep neural network structure has been adopted for multi-task learning [68]. Most
of the applications utilizes either the hard (i.e., sharing the hidden layers between all tasks)
or soft (i.e., each task has its own model (hidden layers) with its own parameters) parameter
sharing.
Figure 3 shows a typical example of multi-task deep learning used in the literature [8, 65,
70, 86, 92, 96, 103], in which we assume there is only one input data with multiple output
targets. As shown in Fig. 3, there are generally two types of hidden layers in the model, i.e.,
the shared layers and task-specific layers. The shared layers learn the intrinsic low-level
representation of the data, which are general among all the tasks, while the task-specific
layer learns the classification network parameters that map the learned latent representations
from the previous shared layers to the task-specific output layers.
The combination of multi-task learning and deep learning has been employed for com-
puter vision [20, 21, 42, 65]. As mentioned previously, the multi-task model performs well
only if the jointly learned tasks are related among each other. However, there is no adequate
definition of task relevance or task-relatedness, leading to the difficulty of grouping the
related tasks for joint learning. To overcome this issue, Fang et al. [21] proposed a dynamic
multi-task convolutional neural network (DMT CNN) with each task holds a subnet and
the degree of information shared among subnets (or tasks) is flexible. This is achieved
by including dynamic connections called task transfer connections among subnets that are
learned dynamically to measure the impact of supervisory signals (e.g., the class label) from
the higher layers on a certain lower layer, leading to a automatic grouping of tasks during
training based on the degree of relevance among tasks. One additional advantage of this
model is its incremental learning capability where subnet of a new task can be augmented
to the existing learned model without the need of updating the already learned parameters.

Fig. 3 Multi-task learning for deep learning


Multimed Tools Appl (2018) 77:29705–29725 29719

Ranjan et al. [65] proposed a deep multi-task learning framework that learns face detec-
tion, landmark localization, pose estimation and gender recognition simultaneously from
images. In this framework, intermediate layers of a deep CNN are fused using a separate
CNN followed by a multi-task learning that operates on the fused features which are based
on the synergy among tasks [65].
On the other hand, Fan et al. [20] proposed a deep multi-task learning algorithm for
large-scale visual recognition that can recognize more than ten thousands of object classes.
Multiple sets of deep features are first extracted from the different layers of a deep CNN
before the visually-similar object classes are assigned to the same group. Based on the inter-
task relatedness or similarities, more discriminative group-specific deep representations can
be learned jointly with a more discriminative tree classifier. This algorithm is able to learn
new object class when given a new training image through a incremental deep learning
approach.
Besides applications in the more conventional computer vision field, the deep multi-
task learning is also applied in medical image analysis recently [2, 9, 23, 50, 57, 69, 80,
89, 97, 100]. Wachinger et al. [80] designed a 3D CNN based segmentation algorithm to
segment neuroanatomy from T1 -weighted magnetic resonance images by jointly learning
an abstract feature representation and a multi-class classification. The algorithm, named as
DeepNAT, predicts not only the center voxel of the patch but also its neighbors via multi-
task learning. In another medical image analysis study, Moeskops et al. [57] proposed a
deep learning algorithm that based on CNN for different segmentation tasks including brain
MRI, breast MRI and cardiac CT angiography (CTA). This deep multi-task algorithm is
able to identify the imaging modality, the visualized anatomical structures, and the tis-
sue classes simultaneously. It also demonstrates potential application of a single system in
clinical practice to automatically perform diverse segmentation tasks without task-specific
training [57].
There is a special type of deep neural network where no shared layers are used, i.e., the
network parameters are not shared among different tasks directly [15]. Rather, the informa-
tion among different tasks is shared by imposing regularization (e.g., look-up table) between
network parameters of different tasks.

4.1 Resources and tools

The MTL algorithms have been implemented in various platforms. For Matlab user, the
available MTL tools include Multi-Task Learning via StructurAl Regularization (MAL-
SAR) [106], Matlab MTL [6, 13, 14, 33, 98, 102], Argyrious’s implementation of MTL
[4, 5, 7], etc.
Online resources for neural network based MTL include [25], which describes how to
use tensorflow to train a multi-task learning neural network, [68] which gives a very com-
prehensive overview of MTL in deep learning, [16], which releases python code for neural
network MTL, [54], which describes how to implement MTL in Keras, etc.

5 Conclusion

In this paper, we have briefly discussed the motivation of multi-task learning, review some
of the current multi-task learning (MTL) methods, and compare their formulations. In addi-
tion, we also discuss how MTL can be used for incomplete multi-modality or longitudinal
data (i.e., some modalities or data at certain time points are missing), how neural network
29720 Multimed Tools Appl (2018) 77:29705–29725

or deep learning can be used to solve complicated multi-task learning problems, and some
of the useful resources and tools that can be used to help us implement codes to solve MTL
problem. We hope that this review can be served as guidance for general readers that are
new in this area to have a quick understanding of the key idea in multi-task learning, and as
a brief refresh for the more advanced users on this topic.

Publisher’s Note Springer Nature remains neutral with regard to jurisdictional claims in published maps
and institutional affiliations.

References

1. Agarwal A, Gerber S, Daume H (2010) Learning multiple tasks using manifold regularization. In:
Advances in neural information processing systems. pp 46–54
2. Ahmed B, Thesen T, Blackmon K, Kuzniecky R, Devinsky O, Dy J, Brodley C (2016) Multi-task
learning with weak class labels: leveraging ieeg to detect cortical lesions in cryptogenic epilepsy. In:
Machine learning for healthcare conference. pp 115–133
3. Ando RK, Zhang T (2005) A framework for learning predictive structures from multiple tasks and
unlabeled data. J Mach Learn Res 6(Nov):1817–1853
4. Argyriou A (2015) Machine learning software. http://ttic.uchicago.edu/∼argyriou/code/
5. Argyriou A, Evgeniou T, Pontil M (2007) Multi-task feature learning. In: Advances in neural
information processing systems. vol 19, pp 41–48. MIT press
6. Argyriou A, Evgeniou T, Pontil M (2008) Convex multi-task feature learning. Mach Learn 73(3):243–
272
7. Argyriou A, Micchelli CA, Pontil M, Ying Y (2008) A spectral regularization framework for multi-task
structure learning, nips 20 Journal Publications on Mathematics (Harmonic Analysis)
8. Caruana R (1998) Multitask learning. In: Learning to learn, pp 95–133. Springer
9. Chaichulee S, Villarroel M, Jorge J, Arteta C, Green G, McCormick K, Zisserman A, Tarassenko L
(2017) Multi-task convolutional neural network for patient detection and skin segmentation in continu-
ous non-contact vital sign monitoring. In: 2017 12th IEEE International conference on automatic face
& gesture recognition (FG 2017). p 5110
10. Chen J, Liu J, Ye J (2012) Learning incoherent sparse and low-rank patterns from multiple tasks. ACM
Trans Knowl Discov Data 5(4):22:1–22
11. Chen J, Tang L, Liu J, Ye J (2009) A convex formulation for learning shared structures from multiple
tasks. In: Proceedings of the 26th Annual International Conference on Machine Learning. pp 137–144.
ACM
12. Chen J, Zhou J, Ye J (2011) Integrating low-rank and group-sparse structures for robust multi-task
learning. In: Proceedings of the 17th ACM SIGKDD international conference on knowledge discovery
and data mining. pp 42–50. ACM
13. Ciliberto C (2017) Matmtl. https://github.com/cciliber/matMTL
14. Ciliberto C, Mroueh Y, Poggio T (2015) Convex learning of multiple tasks and their structure. In:
International conference on machine learning (ICML)
15. Collobert R, Weston J (2008) A unified architecture for natural language processing: Deep neural
networks with multitask learning. In: Proceedings of the 25th international conference on machine
learning. pp 160–167. ACM
16. Crichton G, Pyysalo S (2017) Code supporting: a neural network multi- task learning approach to
biomedical named entity recognition. software, https://doi.org/10.17863/CAM.12584
17. Elgammal A, Lee CS (2004) Separating style and content on a nonlinear manifold. In: Computer
Vision and Pattern Recognition, 2004. CVPR 2004. Proceedings of the 2004 IEEE Computer Society
Conference on. vol 1, pp I–I. IEEE
18. Evgeniou T, Micchelli CA, Pontil M (2005) Learning multiple tasks with kernel methods. J Mach Learn
Res 6(Apr):615–637
19. Evgeniou T, Pontil M (2004) Regularized multi–task learning. In: Proceedings of the tenth
ACM SIGKDD international conference on Knowledge discovery and data mining. pp 109–117.
ACM
20. Fan J, Zhao T, Kuang Z, Zheng Y, Zhang J, Yu J, Peng J (2017) HD-MTL: hierarchical deep multi-task
learning for large-scale visual recognition. IEEE Trans Image Process 26(4):1923–1938
Multimed Tools Appl (2018) 77:29705–29725 29721

21. Fang Y, Ma Z, Zhang Z, Zhang XY, Bai X (2017) Dynamic multi-task learning with convolutional
neural network. In: Proceedings of the Twenty-Sixth International Joint Conference on Artificial
Intelligence, IJCAI-17. pp 1668–1674. https://doi.org/10.24963/ijcai.2017/231
22. Fazel M (2002) Matrix rank minimization with applications. Ph.D. thesis, Department of Electrical
Engineering Stanford University
23. Ghafoorian M, Karssemeijer N, Heskes T, van Uden IWM, Sanchez CI, Litjens G, de Leeuw
FE, van Ginneken B, Marchiori E, Platel B (2017) Location sensitive deep convolutional neu-
ral networks for segmentation of white matter hyperintensities. Scientific Reports 7(1):5110.
https://doi.org/10.1038/s41598-017-05300-5
24. Girshick R (2015) Fast r-cnn. In: IEEE International conference on computer vision. pp 1440–1448
25. Godwin J (2018) Multi-task learning in tensorflow: Part 1. https://www.kdnuggets.com/2016/07/
multi-task-learning-tensorflow-part-1.html
26. Gong P, Ye J, Zhang Cs (2012) Multi-stage multi-task feature learning. In: Advances in neural
information processing systems. pp 1988–1996
27. Gong P, Ye J, Zhang C (2012) Robust multi-task feature learning. In: Proceedings of the 18th ACM
SIGKDD international conference on Knowledge discovery and data mining. pp 895–903. ACM
28. Gong P, Zhou J, Fan W, Ye J (2014) Efficient multi-task feature learning with calibration. In: Proceed-
ings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining. pp
761–770. ACM
29. Han L, Zhang Y (2015) Learning tree structure in multi-task learning. In: Proceedings of the 21th ACM
SIGKDD International Conference on Knowledge Discovery and Data Mining. pp 397–406. ACM
30. Han L, Zhang Y (2016) Multi-stage multi-task learning with reduced rank. In: AAAI. pp 1638–1644
31. Han L, Zhang Y, Song G, Xie K (2014) Encoding tree sparsity in multi-task learning: a probabilistic
framework. In: AAAI. pp 1854–1860
32. Hu R, Zhu X, Cheng D, He W, Yan Y, Song J, Zhang S (2017) Graph self-representation method for
unsupervised feature selection. Neurocomputing 220:130–137
33. Jacob L, Vert Jp, Bach FR (2009) Clustered multi-task learning: A convex formulation. In: Advances
in neural information processing systems. pp 745–752
34. Jalali A, Ravikumar P, Sanghavi S (2013) A dirty model for multiple sparse regression. IEEE Trans Inf
Theory 59(12):7947–7968
35. Jalali A, Sanghavi S, Ruan C, Ravikumar PK (2010) A dirty model for multi-task learning. In: Laf-
ferty JD, Williams CKI, Shawe-Taylor J, Zemel R. S, Culotta A (eds) Advances in neural information
processing systems 23, pp 964-972. Curran Associates, Inc
36. Jebara T (2004) Multi-task feature and kernel selection for svms. In: Proceedings of the twenty-first
international conference on Machine learning. p 55. ACM
37. Jebara T (2011) Multitask sparsity via maximum entropy discrimination. J Mach Learn Res 12(Jan):75–
110
38. Kim S, Xing EP (2010) Tree-guided group lasso for multi-task regression with structured sparsity. In:
International conference on international conference on machine learning. pp. 543–550
39. Lee H, Battle A, Raina R, Ng AY (2007) Efficient sparse coding algorithms. In: Advances in neural
information processing systems. pp 801–808
40. Lee S, Zhu J, Xing EP (2010) Adaptive multi-task lasso: with application to eqtl detection. In: Advances
in neural information processing systems. pp 1306–1314
41. Li C, Gupta S, Rana S, Nguyen V, Venkatesh S, Ashley D, Livingston T (2016) Multiple adverse effects
prediction in longitudinal cancer treatment. In: Pattern recognition (ICPR), 2016 23rd international
conference on. pp 3156–3161. IEEE
42. Li X, Zhao L, Wei L, Yang MH, Wu F, Zhuang Y, Ling H, Wang J (2016) Deepsaliency: Multi-task
deep neural network model for salient object detection. IEEE Trans Image Process 25(8):3919–3930
43. Liu F, Wee CY, Chen H, Shen D (2014) Inter-modality relationship constrained multi-modality multi-
task feature selection for alzheimer’s disease and mild cognitive impairment identification. NeuroImage
84:466–475
44. Liu G, Yan Y, Song J, Sebe N (2014) Minimizing dataset bias: Discriminative multi-task sparse coding
through shared subspace learning for image classification. In: Image processing (ICIP), 2014 IEEE
international conference on. pp 2869–2873. IEEE
45. Liu H, Palatucci M, Zhang J (2009) Blockwise coordinate descent procedures for the multi-task lasso,
with applications to neural semantic basis discovery. In: Proceedings of the 26th Annual International
Conference on Machine Learning. pp 649–656. ACM
46. Liu J et al (2009) SLEP: Sparse Learning with efficient projections arizona state university
47. Liu J, Ji S, Ye J (2009) Multi-task feature learning via efficient l 2, 1-norm minimization. In: Proceed-
ings of the twenty-fifth conference on uncertainty in artificial intelligence. pp 339–348. AUAI Press
29722 Multimed Tools Appl (2018) 77:29705–29725

48. Liu J, Ye J (2009) Efficient euclidean projections in linear time. In: Proceedings of the 26th Annual
International Conference on Machine Learning. pp 657–664. ACM
49. Liu J, Ye J (2010) Moreau-yosida regularization for grouped tree structure learning. In: Advances in
neural information processing systems. pp 1459–1467
50. Liu M, Zhang J, Adeli E, Shen D (2017) Deep multi-task multi-channel learning for joint classification
and regression of brain status. In: International conference on medical image computing and computer-
assisted intervention. pp 3–11. Springer
51. Lounici K, Pontil M, Tsybakov AB, Van De Geer S (2009)
52. Lozano AC, Swirszcz G (2012) Multi-level lasso for sparse multi-task regression. In: Proceedings
of the 29th International Coference on International Conference on Machine Learning. pp 595–602.
Omnipress
53. Mairal J, Bach F, Ponce J, Sapiro G (2009) Online dictionary learning for sparse coding. In: Proceedings
of the 26th annual international conference on machine learning. pp 689–696. ACM
54. Mandal MK (2018) Multi-task learning in keras — implementation of multi-task classification loss.
https://blog.manash.me/multi-task-learning-in-keras-implementation-of-multi-task-classification-loss-
f1d42da5c3f6
55. Maurer A, Pontil M, Romera-Paredes B (2013) Sparse coding for multitask and transfer learning. In:
International conference on machine learning. pp 343–351
56. McDonald AM, Pontil M, Stamos D (2014) Spectral k-support norm regularization. In: Advances in
neural information processing systems. pp 3644–3652
57. Moeskops P, Wolterink JM, van der Velden BHM, Gilhuijs KGA, Leiner T, Viergever MA, Isgum
I (2017) Deep learning for multi-task medical image segmentation in multiple modalities. CoRR
arXiv:1704.03379
58. Negahban S, Wainwright MJ (2008) Joint support recovery under high-dimensional scaling: Bene-
fits and perils of 1,∞ -regularization. In: Proceedings of the 21st International Conference on Neural
Information Processing Systems. pp 1161–1168. Curran Associates Inc
59. Ng A (2018) Multi-task learning. https://www.coursera.org/learn/machine-learning-projects/lecture/
l9zia/multi-task-learning
60. Obozinski G, Taskar B, Jordan M (2006) Multi-task feature selection. Statistics Department UC
Berkeley Tech Rep2
61. Obozinski G, Taskar B, Jordan MI (2010) Joint covariate selection and joint subspace selection for
multiple classification problems. Stat Comput 20(2):231–252
62. Olshausen BA, Field DJ (1997) Sparse coding with an overcomplete basis set: a strategy employed by
v1? Vis Res 37(23):3311–3325
63. Pan SJ, Yang Q (2010) A survey on transfer learning. IEEE Trans Knowl Data Eng 22(10):1345–1359
64. Pong TK, Tseng P, Ji S, Ye J (2010) Trace norm regularization: reformulations, algorithms, and multi-
task learning. SIAM J Optim 20(6):3465–3489
65. Ranjan R, Patel VM, Chellappa R (2017) Hyperface:A deep multi-task learning framework for face
detection, landmark localization, pose estimation, and gender recognition. IEEE Transactions on Pattern
Analysis and Machine Intelligence
66. Rao N, Cox C, Nowak R, Rogers TT (2013) Sparse overlapping sets lasso for multitask learning and
its application to fmri analysis. In: Advances in neural information processing systems. pp 2202–2210
67. Romera-Paredes B, Argyriou A, Berthouze N, Pontil M (2012) Exploiting unrelated tasks in multi-task
learning. In: International conference on artificial intelligence and statistics. pp 951–959
68. Ruder S (2017) An overview of multi-task learning in deep neural networks. arXiv:1706.05098
69. Samala RK, Chan HP, Hadjiiski L, Helvie MA, Richter C, Cha K (2018) Cross-domain and multi-
task transfer learning of deep convolutional neural network for breast cancer diagnosis in digital breast
tomosynthesis. In: MICCAI. vol 10575. https://doi.org/10.1117/12.2293412
70. Seltzer ML, Droppo J (2013) Multi-task learning in deep neural networks for improved phoneme recog-
nition. In: Acoustics, speech and signal processing (ICASSP), 2013 IEEE international conference on.
pp 6965–6969. IEEE
71. Seraj RM (2014) Multi-task learning Internet: https://www.cs.ubc.ca/∼schmidtm/MLRG/multi-task
%20learning.pdf
72. Suo Y, Dao M, Tran T, Mousavi H, Srinivas U, Monga V (2014) Group structured dirty dictionary
learning for classification. In: Image processing (ICIP), 2014 IEEE international conference on. pp
150–154. IEEE
73. Thung KH et al (2014) Neurodegenerative disease diagnosis using incomplete multi-modality data via
matrix shrinkage and completion. Neuroimage 91:386–400
74. Tibshirani R (1996) Regression shrinkage and selection via the lasso. Journal of the Royal Statistical
Society. Series B (Methodological) 58(1):267–288
Multimed Tools Appl (2018) 77:29705–29725 29723

75. Titsias MK, Lázaro-Gredilla M (2011) Spike and slab variational inference for multi-task and multiple
kernel learning. In: Advances in neural information processing systems. pp 2339–2347
76. Turlach BA, Venables WN, Wright SJ (2005) Simultaneous variable selection. Technometrics
47(3):349–363
77. Vasilescu MAO, Terzopoulos D (2002) Multilinear image analysis for facial recognition. In: Pattern
recognition, 2002. Proceedings. 16th international conference on. vol 2, pp 511–514. IEEE
78. Vogt J, Roth V (2012) A complete analysis of the l 1, p group-lasso. arXiv:1206.4632
79. Vounou M, Nichols TE, Montana G, Initiative ADN et al (2010) Discovering genetic associations with
high-dimensional neuroimaging phenotypes: a sparse reduced-rank regression approach. Neuroimage
53(3):1147–1159
80. Wachinger C, Reuter M, Klein T (2018) Deepnat: Deep convolutional neural network for seg-
menting neuroanatomy. NeuroImage 170:434–445. http://www.sciencedirect.com/science/article/pii/
S1053811917301465
81. Wang H et al (2003) Facial expression decomposition. In: Computer vision, 2003. Proceedings. Ninth
IEEE international conference on. pp 958–965. IEEE
82. Wang H, Nie F, Huang H, Yan J, Kim S, Risacher S, Saykin A, Shen L (2012) High-order multi-
task feature learning to identify longitudinal phenotypic markers for alzheimer’s disease progression
prediction. In: Advances in neural information processing systems. pp 1277–1285
83. Wang J, Ye J (2015) Safe screening for multi-task feature learning with multiple data matrices. In:
International conference on machine learning. pp 1747–1756
84. Wang Z, Zhu X, Adeli E, Zhu Y, Nie F, Munsell B, Wu G (2017) Multi-modal classification
of neurodegenerative disease by progressive graph-based transductive learning. Med Image Anal
39:218–230
85. Weiss K, Khoshgoftaar TM, Wang D (2016) A survey of transfer learning. Journal of Big Data 3(1):9
86. Wu Z, Valentini-Botinhao C, Watts O, King S (2015) Deep neural networks employing multi-task learn-
ing and stacked bottleneck features for speech synthesis. In: Acoustics, speech and signal processing
(ICASSP), 2015 IEEE international conference on. pp 4460–4464. IEEE
87. Xiang S, Yuan L, Fan W, Wang Y, Thompson PM, Ye J, Initiative ADN et al (2014) Bi-level multi-
source learning for heterogeneous block-wise missing data. NeuroImage 102:192–206
88. Xin B, Kawahara Y, Wang Y, Hu L, Gao W (2016) Efficient generalized fused lasso and its
applications. ACM Transactions on Intelligent Systems and Technology (TIST) 7(4):60
89. Xue W, Brahm G, Pandey S, Leung S, Li S (2018) Full left ventricle quantification via deep multitask
relationships learning. Med Image Anal 43:54–65. https://doi.org/10.1016/j.media.2017.09.005
90. Yan K, Zhang D, Xu Y (2017) Correcting instrumental variation and time-varying drift using parallel
and serial multitask learning. IEEE Trans Instrum Meas 66(9):2306–2316
91. Yuan L et al (2012) Multi-source feature learning for joint analysis of incomplete multiple heteroge-
neous neuroimaging data. NeuroImage 61(3):622–632
92. Zhang C, Zhang Z (2014) Improving multiview face detection with multi-task deep convolutional neural
networks. In: Applications of computer vision (WACV), 2014 IEEE winter conference on. pp 1036–
1041. IEEE
93. Zhang D et al (2012) Multi-modal multi-task learning for joint prediction of multiple regression and
classification variables in Alzheimer’s disease. Neuroimage 59(2):895–907
94. Zhang J, Ghahramani Z, Yang Y (2006) Learning multiple related tasks using latent independent
component analysis. In: Advances in neural information processing systems. pp 1585–1592
95. Zhang J, Ghahramani Z, Yang Y (2008) Flexible latent variable models for multi-task learning. Mach
Learn 73(3):221–242
96. Zhang J, Liang J, Hu H (2017) Multi-view texture classification using hierarchical synthetic images.
Multimedia Tools and Applications 76(16):17511–17523
97. Zhang J, Liu M, Shen D (2017) Detecting anatomical landmarks from limited medical imaging
data using two-stage task-oriented deep neural networks. IEEE Trans Image Process 26(10):4753–
4764
98. Zhang J, Liu M, Wang L, Chen S, Yuan P, Li J, Shen SGF, Tang Z, Chen KC, Xia JJ et al (2017)
Joint craniomaxillofacial bone segmentation and landmark digitization by context-guided fully con-
volutional networks. In: International conference on medical image computing and computer-assisted
intervention. pp 720–728. Springer
99. Zhang S, Li X, Zong M, Zhu X, Wang R (2017) Efficient knn classification with different numbers of
nearest neighbors IEEE transactions on neural networks and learning systems
100. Zhang W, Li R, Zeng T, Sun Q, Kumar S, Ye J, Ji S (2015) Deep model based transfer and multi-task
learning for biological image analysis. https://doi.org/10.1145/2783258.2783304
101. Zhang Y, Yang Q (2017) A survey on multi-task learning. arXiv:1707.08114
29724 Multimed Tools Appl (2018) 77:29705–29725

102. Zhang Y, Yeung DY (2012) A convex formulation for learning task relationships in multi-task learning.
arXiv:1203.3536
103. Zhang Z, Luo P, Loy CC, Tang X (2014) Facial landmark detection by deep multi-task learning. In:
European conference on computer vision. pp 94–108. Springer
104. Zheng J, Ni LM (2013) Time-dependent trajectory regression on road networks via multi-task learning.
In: AAAI
105. Zheng W, Zhu X, Zhu Y, Hu R, Lei C (2017) Dynamic graph learning for spectral feature selection.
Multimedia Tools and Applications, pp 1–17
106. Zhou J, Chen J, Ye J (2011) Malsar: Multi-task learning via structural regularization. Arizona State
University 21
107. Zhou J, Liu J, Narayan VA, Ye J (2012) Modeling disease progression via fused sparse group lasso.
In: Proceedings of the 18th ACM SIGKDD international conference on Knowledge discovery and data
mining. pp 1095–1103. ACM
108. Zhou J, Yuan L, Liu J, Ye J (2011) A multi-task learning formulation for predicting disease progression.
In: Proceedings of the 17th ACM SIGKDD international conference on Knowledge discovery and data
mining. pp 814–822. ACM
109. Zhou Y, Jin R, Hoi SCH (2010) Exclusive lasso for multi-task feature selection. In: Proceedings of the
Thirteenth International Conference on Artificial Intelligence and Statistics. pp 988–995
110. Zhu X, Li X, Zhang S, Ju C, Wu X (2017) Robust joint graph sparse coding for unsupervised spectral
feature selection. IEEE transactions on neural networks and learning systems 28(6):1263–1275
111. Zhu X, Li X, Zhang S, Xu Z, Yu L, Wang C (2017) Graph pca hashing for similarity search. IEEE
Transactions on Multimedia 19(9):2033–2044
112. Zhu X, Suk HI, Huang H, Shen D (2016) Structured sparse low-rank regression model for brain-
wide and genome-wide associations. In: International conference on medical image computing and
computer-assisted intervention. pp 344–352. Springer
113. Zhu X, Suk HI, Huang H, Shen D (2017) Low-rank graph-regularized structured sparse regression for
identifying genetic biomarkers. IEEE Transactions on Big Data 3(4):405–414
114. Zhu X, Suk HI, Lee SW, Shen D (2016) Subspace regularized sparse multitask learning for multiclass
neurodegenerative disease identification. IEEE Trans Biomed Eng 63(3):607–618
115. Zhu X, Zhang S, Hu R, Zhu Y et al (2017) Local and global structure preservation for robust
unsupervised spectral feature selection IEEE Transactions on Knowledge and Data Engineering
116. Zhu Y, Kim M, Zhu X, Yan J, Kaufer D, Wu G (2017) Personalized diagnosis for alzheimers disease. In:
International conference on medical image computing and computer-assisted intervention. pp 205–213.
Springer
117. Zhu Y, Zhu X, Zhang H, Gao W, Shen D, Wu G (2016) Reveal consistent spatial-temporal patterns
from dynamic functional connectivity for autism spectrum disorder identification. In: International
conference on medical image computing and computer-assisted intervention. pp 106–114. Springer

Kim-Han Thung received his M.Eng.Sc. and Ph.D. degree from the University of Malaya in 2005 and 2012,
respectively. He was a software engineer of Motorola Penang in 2006-2007, and is currently a postdoc in
University of North Carolina at Chapel Hill, Chapel Hill, U.S.A. since 2012. His research interests include
biomedical data processing and analysis, machine learning, multi-task learning, matrix completion, and deep
learning.
Multimed Tools Appl (2018) 77:29705–29725 29725

Chong-Yaw Wee received the Ph.D. degree in electrical engineering from the University of Malaya,
Malaysia, in 2007. He was a Postdoctoral Research Associate at the Biomedical Research Imaging Center,
University of North Carolina at Chapel Hill from 2010 to 2015. He is currently a Senior Research Postdoc at
National University of Singapore. His current research interests include machine learning, neurodegenerative
and neurodevelopment disorders, and multimodal multivariate neuroimaging classification.

You might also like