SPICE a Sparse Covariance-Based Estimation Method for Array Processing
SPICE a Sparse Covariance-Based Estimation Method for Array Processing
Abstract—This paper presents a novel SParse Iterative Co- a small number of sources exists and therefore that only a few
variance-based Estimation approach, abbreviated as SPICE, to rows of the signal matrix
array processing. The proposed approach is obtained by the
minimization of a covariance matrix fitting criterion and is par-
ticularly useful in many-snapshot cases but can be used even in .. .. .. .. (2)
. . . .
single-snapshot situations. SPICE has several unique features not
shared by other sparse estimation methods: it has a simple and
sound statistical foundation, it takes account of the noise in the
data in a natural manner, it does not require the user to make are different from zero. The estimation problem is then to de-
any difficult selection of hyperparameters, and yet it has global cide from the data which rows of the above matrix are
convergence properties. nonzero. Indeed, once this is done, the solution to the location
Index Terms—Array processing, covariance fitting, direction-of- estimation problem, which is usually the main goal of array pro-
arrival (DOA) estimation, sparse parameter estimation. cessing, is immediate : if the row (let us say) of (2) is deemed
to be different from zero then we can infer that there is a corre-
sponding source at an estimated location equal to .
I. INTRODUCTION AND PRELIMINARIES The previous formulation of the location problem begs for the
use of basic ideas from the area of sparse parameter estimation,
(depending on the desired resolution and the dimension of This assumption on the noise term in (1) is reasonable in most
(1D, 2D, 3-D etc.)) can be rather large. For such large dimen- applications. Let us also assume that the signals and the
sional problems the currently available SOCP software (see, noise are uncorrelated with each other for any and that
e.g., [5]) is too slow to use. In an attempt to overcome this com-
putational problem, [3] suggested a way to reduce the number (9)
of columns of and in (6) to manageable values by means
of a singular value decomposition operation. Then the data snapshots are uncorrelated
The approaches that rely on (6), such as the one in [3], suffer with one another and have the following covariance matrix:
from a number of problems. First, the motivation of (6) is
more clearly established in the noise-free case than in the more
practical noisy data case. Second, and likely related to the first
problem, there exist no clear-cut guidelines for the selection of
in (6) (see, e.g., [6] for a critical discussion on this aspect). .. .. .. ..
. . . .
Finally, solving (6) as an SOCP may be too time consuming
for some applications in which , and especially take
on large values, and the available techniques for reducing the
dimensions in (6) (such as that suggested in [3]) require the .. .. .. .. .. ..
choice of further hyperparameters, besides , and even so they . . . . . . ..
.
are only approximate.
In this paper we present a new method of sparse parameter
.. .. .. .. .. .. ..
estimation in models of the type of (1). This method, which is . . . . . . .
called SPICE (SParse Iterative Covariance-based Estimation),
is obtained using a novel covariance-based fitting approach that (10)
was recently introduced in [7] where the focus was on time-
series data (the single-snapshot case) as opposed to the array where
data (the multi-snapshot scenario) considered here. SPICE has a
number of useful features that are hardly shared by other sparse
estimation methods. In particular, SPICE does not suffer from
(11)
the problems described in the previous paragraph. Indeed, as
will be shown in the next section, SPICE has a sound (covari-
ance-based) statistical motivation which makes it possible to use .. .. .. .. .. ..
the method in noisy data scenarios without the need for choosing . . . . . .
any hyperparmeters. Additionally, the SPICE algorithm has a
simple form, and yet it enjoys global convergence properties.
We will show that the covariance fitting problem that SPICE .. .. .. .. .. .. ..
. . . . . . .
solves can be reformulated as an SOCP of the form of (6) but
with , with and replaced by other matrices with
only columns (typically ), and with row-augmented
matrices and that, unlike and in (6), take account of .. .. .. .. .. ..
the noise term in the data (1). This SOCP formulation of the . . . . . .
SPICE estimation criterion shows that the -norm minimiza- (12)
tion problem in (6) can be given a simple statistical motivation
based on covariance fitting, provided that the matrices in (6) are .. .. .. .. .. .. ..
. . . . . . .
suitably defined, see the rest of this paper for details.
II. SPICE ESTIMATION CRITERION The assumption in (9) that the source signals are spatially un-
correlated, which led to the above expression for the covariance
Let us assume that matrix , does not always hold: in some applications the signals
can be correlated or even coherent. However the SPICE method
proposed in this paper is robust to this assumption—we refer to
.. .. .. .. (7) [7] (see also [1]) for a theoretical explanation of this robustness
. . . . property, which will be illustrated numerically in Section V.
We will consider the following covariance fitting criterion for
the purpose of parameter estimation (see, e.g., [8] and the refer-
where stands for the expectation operator, and ences therein):
if
(8) (13)
elsewhere.
Authorized licensed use limited to: UNIVERSITY OF MANITOBA. Downloaded on August 04,2024 at 03:22:44 UTC from IEEE Xplore. Restrictions apply.
STOICA et al.: SPICE: A SPARSE COVARIANCE-BASED ESTIMATION METHOD FOR ARRAY PROCESSING 631
where denotes the positive definite square-root of Appendix A, and is therefore convex (see, e.g., [9]). Solving (18)
as an SDP, however, is not recommended due to the fact that the
(14) available SDP solvers are too computationally intensive for the
values of , , and encountered in many array processing
and the inverses of and are assumed to exist. Note that applications. Consequently we adopt a different line of attack
exists under weak conditions: indeed, while the “true” values of that consists of replacing the problem of minimizing (18) by a
many in may be zero, one typically has related problem, as described next.
which renders nonsingular. Regarding , the in- It follows from (17) that a consistent (in ) estimate of the
verse of this matrix exists with probability one as long as right-hand side of this equation is given by . Therefore, we
. However, for the sample covariance matrix is sin- can think of reformulating the problem of minimizing as the
gular and therefore (13) cannot be used. In the latter case, one following constrained minimization:
can estimate the parameters by minimizing the following
criterion, instead of (13): (20)
(15)
where
The above criterion is the one used in [7] where the focus was
on the time series case with . In the present array pro-
cessing scenario, however, we prefer to consider (13) because (21)
usually the condition is satisfied and because (13) has
a statistically stronger motivation than (15): indeed, under cer- Interestingly, the problems (18) and (20) are not only asymptoti-
tain conditions specified in [8], it can be shown that the param- cally equivalent (as increases, and under the condition that
eter estimates minimizing (13) are statistically asymptotically in (10) can represent the true covariance matrix) but they are ex-
(in ) efficient, whereas the estimates obtained from (15) are actly equivalent (in general) in the sense that their solutions are
suboptimal. Nevertheless, we should note that what we do in scaled versions of each other (note that a scaling of has no
the following for (13) applies with minor modifications to (15) effect on source location estimation). This equivalence property
as well, see Remark 1 below—this observation can be useful in is proved in Appendix B where we also show that the problem
those array processing applications in which and there- obtained from (18) by constraining the first term to (a con-
fore in which (15) should be used in lieu of (13). straint suggested, once again, by asymptotic considerations), in-
A simple calculation shows that stead of constraining the second term as in (20), is equivalent to
(18) and (20) as well.
The problem (20) is also an SDP, and therefore convex.
Furthermore, note that the linear constraint in (20) is of the
(weighted) -norm type, and thus it can be expected to be
sparsity inducing for the solution to (20). Apparently, the crite-
(16)
rion in (20) was never considered in the previous literature on
where sparse parameter estimation. On the other hand, interestingly
enough, this type of criterion occurs frequently in the seemingly
unrelated literature on optimal experiment design (see, e.g.,
(17)
[10] and the references of that paper). The latter literature (in
particular [10]) has served as a source of inspiration for [7] and,
It follows from (16) and (17) that the minimization of is equiv- to some extent, for this paper as well.
alent to the minimization of the function
III. SPICE UPDATING FORMULAS
(18) Let and consider the following problem:
(23)
(19) and the corresponding minimum value of the function in (22) is
Hermitian matrices of appropriate dimensions, means that the From this, it follows that the minimization of the objective in
difference matrix is positive semi-definite): (22) with respect to (s.t. and )
gives (for fixed ):
(25) (31)
(33)
(27)
(29) (36)
(30) (37)
Authorized licensed use limited to: UNIVERSITY OF MANITOBA. Downloaded on August 04,2024 at 03:22:44 UTC from IEEE Xplore. Restrictions apply.
STOICA et al.: SPICE: A SPARSE COVARIANCE-BASED ESTIMATION METHOD FOR ARRAY PROCESSING 633
This function is to be minimized with respect to and Because SPICE monotonically decreases the objective
, subject to function (due to its cyclic operation) and as the minimization
problem it solves is convex, we can expect that the algorithm
(38) has global convergence properties. That this is indeed the case
follows from the general analysis in [10] where it was proved
where that, under weak conditions (essentially requiring that
and that the matrix stays positive definite as the iteration
proceeds), the limit points of the SPICE iterative process are
(39)
global solutions to (20).
By the same argument as above, see (30)–(32), the solution to IV. SOCP FORMULATION OF SPICE
this optimization problem is
SPICE takes account of the noise without introducing any hy-
(40) perparameters to be selected by the user. Indeed, SPICE is a
fully data adaptive approach that estimates both and
from the available data snapshots. As already mentioned, to our
knowledge, the SPICE criterion was never used in the previous
literature on sparse parameter estimation. It is therefore some-
(41) what unexpected that SPICE can be related to (an extended ver-
sion of) the classical -norm minimization approach in (6), as
where we explain in the rest of this section.
As indicated in Section II the SPICE estimation problem,
(42) (20), can be cast and solved as an SDP. However it can also
be reformulated as an SOCP which is a special case of SDP. It
is this reformulation of (20) as an SOCP that reveals the connec-
and the corresponding minimum function (for fixed ) is tion between SPICE and a certain extended -norm minimiza-
tion problem. To reformulate SPICE as an SOCP we proceed
in the following way. Instead of minimizing (22) cyclically (or
(43) alternately) with respect to and , we minimize this function
with respect to for arbitrary . The result of this minimiza-
tion operation is already available from the calculations in the
Inserting the expression (23) for in (40) and (41), we obtain previous section where it was shown that the minimizing powers
the following modified SPICE algorithm for the case of (36): are given by (31) and the corresponding minimum function by
(32). It remains to minimize (32) with respect to (under the
(44) constraint in (22)):
(48)
(45)
SNR dB
Fig. 3. Tracking performance of PER, SPICE and SPICE+ (the NULA case). The solid lines denote the trajectories of the true DOAs: (a) PER; (b) SPICE with
five iterations; (c) SPICE+ with five iterations; and (d) SPICE+ with 20 iterations.
(57) (64)
and
Let be auxiliary variables satisfying ,
or equivalently (65)
(66)
(70)
for some ; note that the scaling factor in (70) must nec-
(62) essarily be positive, which is why we wrote it as ; below we
will let denote the square root of . We will show in
Using we can restate (53) in the following form: the following that the previous assumption leads to a contradic-
tion : cannot be the solution to if . To do so let
and observe that, at , the two terms in the ob-
jective of are identical:
(71)
Making use of (70) and (71) along with the assumption that
.. is the solution to , we obtain the following inequality:
.
(72)
which cannot hold, and hence the solution to must satisfy [12] T. Yardibi, J. Li, P. Stoica, M. Xue, and A. B. Baggeroer, “Source
(70) with . Using this fact we can reformulate as localization and sensing: A nonparametric iterative adaptive approach
based on weighted least squares,” IEEE Trans. Aerosp. Electron. Syst.,
follows: vol. 46, no. , pp. 425–443, Jan. 2010.
and
(74) Petre Stoica (SM’91–F’94) is a Professor of systems modeling in the Depart-
or, equivalently (as ), ment of Information Technology, Uppsala University, Uppsala, Sweden.
More details about him can be found at http://www.it.uu.se/katalog/ps.
and (75)
where , as before, and is a new auxiliary Prabhu Babu received the M.Sc. degree in electrical engineering from the In-
dian Institute of Technology, Delhi, India, in 2007.
variable. Comparing (75) and (67) concludes the proof of the He is currently working towards the Ph.D. degree in electrical engineering
equivalence of and . with applications in signal processing at the Department of Information Tech-
nology, Uppsala University, Uppsala, Sweden.
REFERENCES
[1] P. Stoica and R. Moses, Spectral Analysis of Signals. Upper Saddle
River, NJ: Prentice-Hall, 2005. Jian Li (S’88–M’90–SM’97–F’05) received the M.Sc. and Ph.D. degrees in
[2] R. Tibshirani, “Regression shrinkage and selection via the lasso,” J. R. electrical engineering from The Ohio State University (OSU), Columbus, in
Stat. Soc. .B, Methodol., vol. 58, no. 1, pp. 267–288, 1996. 1987 and 1991, respectively.
[3] D. Malioutov, M. Cetin, and A. Willsky, “A sparse signal reconstruc- From April to June 1991, she was an Adjunct Assistant Professor with the
tion perspective for source localization with sensor arrays,” IEEE Department of Electrical Engineering, OSU. From July 1991 to June 1993,
Trans. Signal Process., vol. 53, no. 8, pp. 3010–3022, Aug. 2005. she was an Assistant Professor with the Department of Electrical Engineering,
[4] M. Lobo, L. Vandenberghe, S. Boyd, and H. Lebret, “Applications of University of Kentucky, Lexington. Since August 1993, she has been with the
second-order cone programming,” Linear Algebra Appl., vol. 284, no. Department of Electrical and Computer Engineering, University of Florida,
1-3, pp. 193–228, 1998. Gainesville, where she is currently a Professor. In fall 2007, she was on
[5] J. Sturm, “Using SeDuMi 1.02, a MATLAB toolbox for optimization sabbatical leave at the Massachusetts Institute of Technology, Cambridge.
over symmetric cones,” Optim. Methods Softw., vol. 11, no. 1, pp. Her current research interests include spectral estimation, statistical and array
625–653, 1999. signal processing, and their applications.
[6] A. Maleki and D. Donoho, “Optimally tuned iterative reconstruction al- Dr. Li was a member of the Editorial Board of Signal Processing from
gorithms for compressed sensing,” IEEE J. Sel. Topics Signal Process., 2005 to 2007. She has been a member of the Editorial Board of Digital Signal
vol. 4, no. 2, pp. 330–341, 2010. Processing—A Review Journal since 2006. She was an Executive Committee
Member of the 2002 International Conference on Acoustics, Speech, and
[7] P. Stoica, P. Babu, and J. Li, “New method of sparse parameter estima-
Signal Processing, Orlando, FL, in May 2002. She was an Associate Editor of
tion in separable models and its use for spectral analysis of irregularly
the IEEE TRANSACTIONS ON SIGNAL PROCESSING from 1999 to 2005 and an
sampled data,” IEEE Trans. Signal Process., vol. 59, no. 1, pp. 35–47, Associate Editor of the IEEE Signal Processing Magazine from 2003 to 2005.
Jan. 2011. She is presently a member of the Sensor Array and Multichannel Technical
[8] B. Ottersten, P. Stoica, and R. Roy, “Covariance matching estima- Committee, IEEE Signal Processing Society. She received the 1994 National
tion techniques for array signal processing applications,” Digital Signal Science Foundation Young Investigator Award and the 1996 Office of Naval
Process., vol. 8, no. 3, pp. 185–210, 1998. Research Young Investigator Award. She is a coauthor of papers that received
[9] Y. Nesterov and A. Nemirovskii, Interior-Point Polynomial Algorithms the First and Second Place Best Student Paper Awards, respectively, at the 2005
in Convex Programming. Philadelphia, PA: SIAM, 1994. and 2007 Annual Asilomar Conferences on Signals, Systems, and Computers,
[10] Y. Yu, “Monotonic convergence of a general algorithm for computing Pacific Grove, CA. She is also a coauthor of the paper that received the M. Barry
optimal designs,” Ann. Stat., vol. 38, no. 3, pp. 1593–1606, 2010. Carlton Award for the best paper published in the IEEE TRANSACTIONS ON
[11] T. Söderström and P. Stoica, System Identification. Englewood AEROSPACE AND ELECTRONIC SYSTEMS in 2005. She is a Fellow of IET and a
Cliffs, NJ: Prentice-Hall, 1989. member of Sigma Xi and Phi Kappa Phi.
Authorized licensed use limited to: UNIVERSITY OF MANITOBA. Downloaded on August 04,2024 at 03:22:44 UTC from IEEE Xplore. Restrictions apply.