Detection and Estimation Theory
Detection and Estimation Theory
net/publication/44199094
CITATIONS READS
18 969
5 authors, including:
Arthur Baggeroer
Massachusetts Institute of Technology
263 PUBLICATIONS 5,332 CITATIONS
SEE PROFILE
All content following this page was uploaded by Arthur Baggeroer on 02 January 2017.
Graduate Students
The work of this group may be divided into four major areas.
1. Sonar
2. Communications
a. Digital Systems
The optimum receiver for the detection of Gaussian signals in Gaussian noise is well
known. Except for limiting cases it is difficult to evaluate the error behavior. Work con-
tinues on developing performance measures for evaluating the performance, emphasizing
techniques that are computationally tractable and, at the same time, give a good meas-
ure of the system performance. Both tight upper bounds and computational algorithms
have been developed for the probability of error which emphasize the fundamental role
of optimum linear systems in detection problems. Future work includes the application
of these techniques to the analysis and design of radar, sonar, and communication sys-
tems.
The availability of a noiseless feedback channel from receiver-to-transmitter enables
a significant increase in performance. By utilizing the continuous feedback signal at the
modulator, the behavior system performance at the end of the transmission interval is
greatly improved. The feedback link could be used to obtain the same performance over
a shorter transmission interval. The actual structure of the system is very flexible and
simple.
Noise in the feedback channel degrades the achievable system performance with the
simple feedback system. Improvement over the no-feedback system is obtained, but it
is not as dramatic as when noiseless feedback is available.
b. Analog Systems
(i) Investigations of the performance of analog modulation systems operating in addi-
3
tive noise channels are essentially completed.
(ii) When a noiseless feedback link is available from receiver-to-transmitter simple
modulation schemes can be developed which achieve the rate-distortion bound. Realizible
feedback systems perform very close to the rate-distortion bound.
The effects of additive noise in the feedback link depend on the relative noise levels
in the two channels. For relatively small feedback channel noise, the system performance
is close to the rate-distortion bound. For large feedback noise the availability of a feed-
back link does not significantly improve the system performance.
(iii) A new approach has been developed for estimating continuous waveforms in real
time. The approach is formulated with continuous Markov processes and use is made of
state-variable concepts. The approach has been applied successfully to the problem of
estimating continuous stochastic messages transmitted by various linear and nonlinear
modulation techniques over continuous random channels; particular emphasis has been
given to phase and frequency modulation. An advantage of this approach over alternative
schemes is that it leads automatically to physically realizable demodulators that can be
4
readily implemented.
3. Seismic
The random process(es) is represented as the output of a linear dynamic system that
is described by a state equation. The Fredholm integral equation is reduced to a vector
differential equation that is directly related to the state equation of the dynamic system.
From this equation, a determinant is found which must vanish in order that an eigenvalue
exist. Once the eigenvalue is found, the eigenfunction follows from the transition matrix
of the vector differential equation.
The technique is general enough to handle a large class of problems. Constant-
parameter (possibly nonstationary) dynamic systems for both scalar and vector processes
can be handled in a straightforward and analytic manner. Time-varying systems can
5
also be treated by computational techniques.
(ii) The problem of formulating a state-variable model for random channels encoun-
tered in practice is being investigated. A particular class of channels of interest are
those exhibiting frequency-selective fading.
(iii) The system identification problem is being studied. Applications include meas-
urement of noise fields, random process statistics, and linear system functions.
b. Detection Techniques
References
1. Introduction
00
where Tb is the baud duration, and k contains the information transmitted on the
th
k baud. Our problem is to decide between the hypothesis H that to = +1 and hypothesis
H 1 that to = -1. We make the following assumptions.
(i) The ek are independent.
(ii) Ho and H1 are equally likely.
(Fig. XXVII-4); then we shall optimize its parameters for the binary communication
problem of interest (Eqs. 5 and 6). Finally, we shall work a simple example demon-
strating that our new decision-feedback equalizer is capable of rendering far better per-
formance than the conventional MF-TDL receiver.
We want to determine the optimum receiver structure for the problem and
assumptions that have been introduced, but first we introduce some notation and
definitions that will prove useful in the derivation that is to follow.
Definitions: 1. -_ = {( kk<0}
2. + = k jk>0}
Under our assumptions, - is known correctly via decision feedback, and with this
taken into account, the optimum receiver calculates the likelihood ratio
prr(t) (,H,A =
1exp r(b kh(t-Tb- oh
h(t)- h(t- T b t
k<O f>0 o1
and since
is independent of + and o , it will factor out of the integral and cancel with the same
0
term arising in the denominator of A. Thus the terms of importance remaining in the
integrand of the numerator are
(t ) dt
N0 oh + C
> h(t- fTb)
j>0
0o =1
2
oo 2(o 1 bj
f>0 j>0 j>0
I =
where we have defined Qj= 2 6j. Thus, by factoring out of the integral those terms
S 2o e
that are independent of , the numerator of A becomes proportional to
+2> 0
-z
k<O
b Ik-flk
- d o
exp 2 o a - 2 bk -
2
bo +
1I0 a - b j-k Ik ob P af
k<0 j>0o f>0 ( k< 0
- b Im I - ob
m< 0 ) =
0o 1
-1
where the Pj are elements of the matrix P defined byP = (B+Q) (for j, f >0), and
Qj, are as defined above. This same expression evaluated at o = -1 gives the denom-
inator of A. It thus follows that the optimum receiver computes
A = exp 4a - 4 bk k -4 bf a j bIkjk j
k<0 j>0 fj>0 k<O
gj - Pjfb
fk bk + gblj-k'
j>
The receiver structure may now be found from this decision rule. From definition 3,
it is seen that the sufficient statistics a. may be generated by using a TDL having taps
spaced at the baud duration Tb, as shown in Fig. XXVII-1. Moreover, since the
weightings on the a. may be placed before the integration and multiplication operations,
and the multipliers and integrators are common to all taps of Fig. XXVII-1, then
clearly we can generate
jaj
j> 0
as shown in Fig. XXVII-2, in which we define go = 1.
r(t)
a.I a a0
RECEIVER
INPUT 1
L-DECISION
FEEDBACK
a-k
k .... (f~
1 SUMMING BUS k k0
k<O
We now want to adopt the equalizer structure derived above (Fig. XXVII-4) and deter-
mine the forward and feedback TDL tap gains that minimize the sum of the signal side-
lobes and noise power at the receiver output in the absence of decision errors. We first
introduce some additional notation and definitions that will prove useful in the following
discussion.
SAMPLE AT
BAUD RATE
INPUT MATCHED "FORWARD" TDL
FILTER FORWARD TDL
aj I a o
gj .... 91 go
SUMMING BUS
+
H0
H
1
SUMMING BUS
k$-
-k -1 DECISION
" FEEDBACK" TDL
FEEDBACK
(a)
q_1 I q
9-3
(b)
N more samples occurring before the main sample (which is denoted sample number 0)
than after it, where the MF output has 2N + 1 nonzero samples. This is in contrast with
the typical output from the conventional MF-TDL equalizer, which is seen in
Fig. XXVII-5b to always exhibit symmetry about the main sample.
Before we can proceed to determine the optimum choices of g and f under our
minimum-output-sample-variance criterion, we must first understand the effect of the
decision feedback on the output distortion. Consider the signal component out of the
forward TDL at the first sample time:
j 0
gj[aj+ 1 ]
signal
= k
j >0
g[
0
{
k
kkh(t-kTb{)h(t-(j+1)Tb)}dtj.
The contribution to this component, which is due to the bauds for which decisions have
already been made (that is, on all the k up to and including o ), is then
Next, consider the output of the feedback TDL at this same first sample time. With
the use of Eq. 3 for the fk, this becomes
and if we let k = k + 1,
"
k> O I I1 1-k*
j>0 k +j+
j20k <0
Here, we have used b * = b* and our earlier definition, g 0 1. Thus we see that
k -1 l-k*
in the absence of decision errors the feedback-TDL output is exactly the same as the
contribution to the forward-TDL output at the first sample time, attributable to past bauds,
and hence there is no net contribution to the distortion from those bauds upon which
decisions have already been made. We now see the three important advantages that the
decision-feedback equalizer enjoys over the conventional equalizer.
(i) The conventional equalizer cannot completely eliminate interference due to past
bauds, due to noise-enhancement considerations, as well as the practical constraint of
a finite TDL length. The decision-feedback equalizer, in the absence of decision errors,
completely eliminates the intersymbol interference due to past bauds.
(ii) For the decision-feedback equalizer, the forward-TDL gain-vector g may be
optimized without consideration of the q, for I > 0, since these are eliminated by the
decision-feedback, while the conventional equalizer must be designed to simultaneously
suppress all of the q, for I * 0. This additional freedom enables the decision-feedback
equalizer to achieve much better suppression of the intersymbol interference due to
future bauds.
(iii) Since the intersymbol interference due to past bauds is suppressed through the
noiseless feedback-TDL rather than through using additional taps on the forward-TDL
as in the conventional equalizer, then clearly the output noise power is significantly
reduced.
Each of these advantages contributes to a much better performance for the decision-
feedback equalizer compared with that of its conventional counterpart, as we shall show
in the example. In view of the conclusions, stated above, it is clear that the output
sample variance is given by
Under the constraint that the main sample be unity, we may include qo in the summation
to find that
qf= gjXjkgk = g X
g,
Q 0 j >0 k>0O
o oT
2 gjjkk Yg.
j O k0O
Thus under the constraint that qgo = gT = 1, we want to minimize the quantity
J= gT + oY g + k(1-gT
over the choice of the forward-TDL gain vector g. The unique solution is given by
g -j (5)
g= -1
,TX+ oJ
2 - ]-
This result appears formally the same as that for the optimum tap-gain vector in the
7
conventional MF-TDL equalizer. The difference is that for the conventional equalizer
the summation of definition 8 is taken over all 1, while the vector of definition 9 has ele-
ments for all i. Thus X matrix for the conventional equalizer becomes symmetrical and
Toeplitz, while the c vector becomes symmetrical, and these properties do not hold for
the decision-feedback equalizer.
The proper choice of the f vector now follows directly from Eq. 3, except that we
have dropped the 1/N 0 factor common to all terms of Eq. 4 (that is, the aj contained this
factor, while it is not included in the MF output in the present discussion) and thus
replace the bk there by k:
fk =
gjlj-kI
If we define a matrix Y with elements Y jk =j-kj-k for j > 0 and k < 0 (note that it is the
range of k which distinguishes this matrix Y from the matrix Y of definition 7), then
the feedback-TDL tap-gain vector may be conveniently written as
f = Yg (6)
Thus, once the sampled channel autocorrelation function and additive noise level have
been specified, one can use Eqs. 5 and 6 to determine the parameters of the minimum-
variance decision-feedback equalizer. This is illustrated by the example that follows.
We shall now work out an example, applying the derived decision-feedback structure
to the equalization of the channel whose sampled autocorrelation function is shown in
Fig. XXVII-6a. This particular example was chosen, not because it necessarily pro-
vides a realistic model of channels of practical importance, but because it illustrates
the important advantages that the decision-feedback equalizer has over the conventional
MF-TDL equalizer. More complex examples could have been chosen, for they pose no
additional difficulties.
We thus want to find the conventional equalizer and decision-feedback equalizer
TDL
C1
d d
+
(a) ( H0
0
H
1
FROM MF
TDL
DECISION
STDL FEEDBACK
H0 (a)
< 0 1 + cd
SUMMING BUS
c+d d
(b)H1
1+ 2 cd
Icd (b)
1 + cd
c+d c+d
cd (cd cd
(C) d (c)
structures as a function of the noise level, No/2, and the sidelobe level, d, of
Fig. XXVII-6a.
Using the 3-tap conventional equalizer shown in Fig. XXVII-6b, one can solve Eq. 5
(with appropriate X and _, and normalizing the result so that go = 1) to find that
3
c 2dd 3N - d
(7)
2 )2
1- d (1-2d
2
minimizes the output-sample variance, which results in the response to a single trans-
mitted to = +1 baud shown in Fig. XXVII-6c. The noise present at the output has vari-
ance given by
N2
2 (1+4dc+Zc 2).
(8)
Given -2 and an arbitrary set of sidelobes, we can determine the probability of error
very efficiently, using an "error tree" algorithm developed by the author. 7 This
algorithm was applied to the present problem for d = . 4, d = . 48, and d = . 50, with the
resulting performance curves for the conventional MF-TDL equalizer shown in
Figs. XXVII-8, XXVII-9, and XXVII-10. These curves are discussed further below,
when we compare them with the corresponding curves of the decision-feedback equalizer.
For the channel of Fig. XXVII-6a, the decision-feedback equalizer is as shown in
Fig. XXVII-7a, where we consider it to be the counterpart of Fig. XXVII-6b, since the
delay and tap-gain requirements are the same with both equalizers. We should note at
this point that although the forward TDL of the decision-feedback equalizer appears to be
half of the conventional-equalizer TDL, this is not the case, in general. In general, if
the sampled-channel autocorrelation has M sidelobes, then the feedback TDL requires
M taps, with the remaining in the forward TDL. Thus, for example, if one had a con-
ventional equalizer of 55 taps to equalize a channel having 10 autocorrelation sidelobes,
then the hardware-equivalent decision-feedback equalizer would have 45 forward-TDL
gains and 10 feedback-TDL gains.
Solving Eq. 5 and normalizing so that go = 1, one finds that
d -d
c= N (9)
1+ 2)
2o (1-d
N
- - 2 (1+2dc+c ). (10)
As discussed further below, the output noise is smaller than that appearing at the output
of the conventional equalizer after appropriate normalizations have been made. Note that
the parameter c of Eqs. 9 and 10 is numerically different from that of Eqs. 7 and 8.
The performance of the decision-feedback equalizer was determined through digital
computer simulations, with the results shown in Figs. XXVII-8, XXVII-9 and XXVII-10.
The signal-to-noise ratio in these figures is given by
SNR = 10 log 1 0 ( 2
since we have assumed unit signal energy on each baud. Also, to place the performance
curves of the conventional and decision-feedback equalizers in better perspective,
Figs. XXVII-8, XXVII-9, and XXVII-10 show the performance curves of an unequalized
receiver (matched filter only) and of the ideal receiver (that obtained when transmitting
0-:
Fig. XXVII-8. Performance for the channel indicated Fig. XXVII-9. Performance for the channel indicated
in Fig. XXVII-6a, with d = . 40. in Fig. XXVII-6a, with d = . 48.
(XXVII. DETECTION AND ESTIMATION THEORY)
100
m 10-2
O
-16 -10 -4 2 8 14 20 26
SNR (db)
As the sidelobe energy is increased (Figs. XXVII-9 and XXVII-10), the decision-feedback
equalizer becomes 3-5 db off the ideal, and approximately 12 db better than the conven-
tional equalizer at d = . 48. For d =. 50, the conventional equalizer is seen in
Fig. XXVII-10 to approach a limiting performance with increasing SNR, while the
decision-feedback equalizer continues to improve rapidly beyond ~5 db. The behavior of
the conventional equalizer here is due to the fact that d = . 50 renders an input distortion
32 -
28 -
24
U 20
CONVENTIONAL
EQUALIZER
Z
16 -
U
z
Z 12
O
Z
-16 -10 -4 2 8 14 20 26
SNR (db)
of unity (with a sum of the side-lobe magnitudes used as measure; thus the "eye" was
closed even in the absence of noise, for those familiar with "eye diagrams"), and a
(2M+1)-tap TDL conventional equalizer exhibits a limiting performance of 2 - 2M - 3
with
increasing SNR.
If one normalizes the main output samples to unity, and correspondingly normalizes
the output noise variances, the noise enhancement is considerably more with the conven-
tional equalizer than with the decision-feedback equalizer in this example, as shown in
Fig. XXVII-11 for d = .40. This illustrates advantage (iii), which we listed for the
decision-feedback equalizer.
Note that at low SNR the performance of the two equalizers coincides in
Figs. XXVII-8, XXVII-9, and XXVII-10. This is not perhaps what one might have
expected from heuristic arguments, which point out that when an error is made by the
decision-feedback equalizer the feedback-TDL contribution enhances rather than
eliminates the q 1 sample (see Fig. XXVII-7b and 7-c), thereby resulting in this example
in an additional equivalent interfering sample whose magnitude exceeds that of the main
sample by cd . This gives a large probability of error on the next decision (=1/3 for
intermediate SNR, calculated under the assumption of uncorrelated distortion), and thus
it appears quite possible that "one bad decision will lead to another," and cause the per-
formance at low SNR to become eventually worse than that of the conventional equalizer.
Such behavior was not observed in this example, however, and the decision-feedback
equalizer appears at least as good as the conventional equalizer at all SNR.
For more complex channels that we have studied, in which the energy in the side
lobes is much greater than in the simple case considered here, the anticipated thresh-
olding effect has been noted, but only at high conventional equalizer error rates, and
the decision-feedback equalizer still exhibits far better performance at all SNR of prac-
tical importance.
M. E. Austin
References
1. Introduction
2. Signal Model
Let us now discuss the equations which describe the system of interest (see
AMPLITUDE
MODULATION
OF THE N+ 1
1 INPUTS
DELAY
al I
DELAY x(t-a2
2
Co (t) x(t) +
N
x= Fx + G ( x(t-aN-1 I C (t) x (t-ai) r()
MESSAGE (t) DELAY +
SOURCEN-
OBSERVATION
INTERVAL
x ( t - N O < t < Tf
-a N < t <n Av
Fig. XXVII-12). We assume that the dynamics of the message process are determined
by a linear state equation
dx(t)
dt = F(t) x(t) + G(t) v(t)
over the time interval -aN < t < Tf, where F(t), G(t) are matrices determining the
dynamics of the state equation, and v(t) is a white Gaussian source noise with
In order to completely specify this random process, we need to make some assump-
tions about the conditions of the initial state. We assume that the initial state is a
Gaussian random vector with
E[x(-aN)] = (-aN)
EL (x(-aN)-(-aN))((-aN)-(-aN))T] = P(-a N)
We shall now discuss the observation process. We assume that we observe N dif-
ferent delayed versions of an amplitude modulation of the state vector in the presence
of a white Gaussian observation noise. Our observation is, therefore,
where
0 <t <Tf
T
E[w(t)_w (T)] = R6(t-T). (7)
Notice that our observation equation is defined over a different time interval from the
state equation. For convenience, let us define
(8)
Yi(t) = x(t- a).
Aside from the delay terms that enter, our assumptions do not differ from the usual
ones made in state-variable estimation procedures.
It can be shown that the problem of maximizing the a posteriori density is equivalent
to minimizing the following quadratic functional:
+ f Lx~t 1- dt N (9)
+ T v(t) -1 dt (9)
aN Q
Subject to the constraint of the state equation (1) and the delay operations specified by
Eq. 6 (xA = xTAx).
Let us first consider the state-equation constraint. We can introduce this by using
a Lagrangian multiplier. This is done by adding to the quadratic functional the term
T fT dx(t)
L = p (t) dt - F(t) x(t) -G(t) v(t)) dt. (10)
aN
It will be useful to integrate the first term of the integrand by parts. Doing this, we
have
_i(t, T) a_ i(t, T)
+ =0, (12)
at aT
where
_i(t, 0) = x(t).
As a result, we are able to impose the constraint of each delay term by using a
Lagrangian multiplier that is a function of two variables, that is, we want to add to the
quadratic functional terms of the form
ai
1 0 Tf F t T 'T)
(t (t'T)
L. = , + dtdT, (16)
in which we have
L. = f ( -
(t, a.).(t, a.)
-i i 1
(t, 0) (t, 0) dt
a i T T
+ _ (Tf, T) i(Tf, T)-.L (0, T) .(0 , T) dT
T
a ia (t,)
+ .((t,
T)
- 0 Tf (t,T) dtdr. (18)
at IT ---.)
When we add the Lagrangian multiplier constraints Lo, L1, ... L N imposed by the state
equation and the delay terms to the quadratic functional described by Eq. 9, we have
J(v(t),x(-aN)) = x(-aN)-P(-aN)
-(-aN
T
r(t) - C (t) x(t) C.(t) y.(t)
1 -1
i=1 R-1
T
v(t) 1-1 dt +p x(T) - p T(T ) x(To)
Tf)(Tf)
aN
T fdp (tM T T
-dt x(t) + p (t) F(t) x(t) + p (t) G(t) v(t)I
aN
+ i=l i T( t ,
ai )Yi(t)-T(t ,
0)x(t)
Tf 0 0aI
/ T
(t
t T
+
T IT7
i(t.T , 7) dtd .
(19)
d62(t)
dt= F(t) 8x(t) + G(t) 6v(t). (23)
dt
The variation by the quadratic functional which results from these perturbations is given
by
+ C (t) 8y (t)
j= 1
Tf ^ T T (
6
+ v(t) Q-16(t) dt + p (T ) 6 x(Tf) - p (T 0 ) x(-aN)
aN
Tf dp (t)
5x(t) + p T (t) F(t) 6x(t)
x
+ p (t) G(t) 6v(t)f dt
dt
N
N
STf T(t, a )6y (t)- LT (t, 0) 5x() d
0
i=l
T T dT]
i=l [ (TfV)6x(T T-)-1 . (0,T)5x( T)
N T T A
a1 j (t, T)
T a.
+ i at + 87
T 6c.(t,
1
7) dtdT} + 0(E ). (24)
00
i= 1
We now want to combine the various variations into common factors. We have
J(y(t),x(-aN)) = J(v(t),x(-aN)) +
x- N ) IT P
1(-aN)-pT(-aN)) bx(-aN)
E 1x( aN)-5(-aN)]
N 7T
+Yff
+ ,~T
N
dp T(t)
dt p T(t) F(t) - t,
S (t,0) 6X(t)
i=1
Tf T -1 T ^
((V(t) Q p (t)G(t))5v(t)) dt
aN
Tf (
- r(t) - Co(t) x(t)- C (t) j(t R -1 C.(t)
(t)
i+ 0 1
i= 1 0
+ (t, a (t) dt
1 i -
+
0 -
dt
-p~t
p t) F(t)
N
6x(t) dt
N1 j= 1
N- 1 N
dp T (t)
dt
(t) F(t)- k -J3
T (0, -t) 6x(t) d
i=l j=i+l
N a. N Op.. (t, 7)
Lj(Tf, T) 6x(Tf-_) ( at )
j=1
0
j=1
KTf
YO\ at
T
+ . (t 7) t,
+ OT /6c.(t T) dtdT . (25)
We shall now make a series of arguments to cause the E variation of the functional
to vanish. We shall require that the Lagrangian multiplier functions satisfy some equa-
tions so that the coefficients of some of the variations vanish.
First we require the delay constraints to have the functional form
so that the last term in the equation vanishes identically ([ii is the adjoint function for the
delay equation). Furthermore, we require that
(t- ) = 0 (27)
This completes our restrictions of the Lagrangian multipliers for the delay con-
straints. Notice that the restrictions may be made independently of all the other con-
straints, including the state-equation constraint.
Now we shall impose some restrictions on the Lagrangian multiplier for the state-
equation constraint. First we impose the restriction, 0 < t < Tf,
N
dt F C (t) R(t)yj(t)(t) - Co(t ) (30)
3
j=1 j=1
Finally, we impose restrictions for the time before the observation interval. Within this
t
and, for -ai+ < < a., i = 12,... N- 1,
N
N
dp(t)
dt
dt
-F T(t) p(t) - L 1 .
(t) (32)
j=i+ 1
As a result, we have defined p(t) over the entire interval of the process. Therefore we
have
1 (34)
p(-aN) = P (-aN)(((-aN))-_(-aN))
-1^ T
Q v(t) = G (t) p(t). (35)
Equation 34 imposes an initial boundary condition, and Eq. 35 relates the terms v(t) and
p(t). By using Eq. 35, our state equation becomes
dx(t)
dt = F(t) x(t) + G(t) QG(t) p(t). (36)
We shall now summarize our equations determining the MAP interval estimate of the
process x(t). For convenience, we shall define
a =0
y (t) = x(t).
d^(t) T
dt F(t) x(t) + G(t) QG (t) p(t).
By using the definitions stated above, we may write Eqs. 33 and 35 in the same manner.
for -ai+ < t <-a. these equations become
T N
dp(t) (t )
dt -F (t) p(t)- o ,
dt 0
j=i+l
i = 0, 1, ... N-1.
N N
dp(t) o1 R- 1 (r(t) -
dt = -F(t) p(t) - Co(t) Cj(t) yj(t
j=0 j= 1
The functions i. are defined to be, for 0 < T < ai,
Soi(T f -T)
-o
= 0,
References
1. A. C. Bryson and M. Frazier, "Smoothing for Linear and Nonlinear Dynamic Sys-
tems," Proc. Optimum Systems Conference, Wright-Patterson Air Force Base, Ohio,
September, 1962.
2. R. E. Kalman and R. S. Bucy, "New Results in Linear Filtering and Prediction
Theory," ASME Paper 60-JAC, May 31, 1960.
3. M. Athans and P. Falb, Optimal Control (McGraw-Hill Book Company, Inc.,
New York, 1966).
4. A. Baggeroer, "Maximum A Posteriori Interval Estimation," WESCON/66 Technical
Papers, Session 7, Paper 7/3.