0% found this document useful (0 votes)
23 views15 pages

Error Analysis in Homography Estimation by First Order

Uploaded by

fibec66073
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
23 views15 pages

Error Analysis in Homography Estimation by First Order

Uploaded by

fibec66073
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 15

J Math Imaging Vis (2009) 33: 281–295

DOI 10.1007/s10851-008-0113-2

Error Analysis in Homography Estimation by First Order


Approximation Tools: A General Technique
Pei Chen · David Suter

Published online: 5 September 2008


© Springer Science+Business Media, LLC 2008

Abstract This paper shows how to analytically calculate space based refinement stage. Comparison with the straight-
the statistical properties of the errors in estimated parame- forward subspace refinement approach (without taking into
ters. The basic tools to achieve this aim include first order account the statistical properties of the homography coef-
approximation/perturbation techniques, such as matrix per- ficients) shows that our statistical characterization of these
turbation theory and Taylor Series. This analysis applies for errors is both correct and useful.
a general class of parameter estimation problems that can be
abstracted as a linear (or linearized) homogeneous equation. Keywords Error analysis · Matrix perturbation theory ·
Of course there may be many reasons why one might Singular value decomposition · Low rank matrix
which to have such estimates. Here, we concentrate on the approximation · Homography · First order approximation ·
situation where one might use the estimated parameters to Mahalanobis distance
carry out some further statistical fitting or (optimal) refine-
ment. In order to make the problem concrete, we take ho-
mography estimation as a specific problem. In particular, we 1 Introduction
show how the derived statistical errors in the homography
coefficients, allow improved approaches to refining these Parameter estimation is a common problem in engineering.
coefficients through subspace constrained homography es- In some applications, parameter estimation is the ultimate
timation (Chen and Suter in Int. J. Comput. Vis. 2008). goal, while in others, estimated parameters are further fed
Indeed, having derived the statistical properties of the er- to follow-up procedures. In the latter case, knowledge of the
rors in the homography coefficients, before subspace con- statistical properties of the errors in estimated parameters
strained refinement, we do two things: we verify the cor- (such as knowing the covariance matrix) helps one “intelli-
rectness through statistical simulations but we also show gently” design algorithms that use these parameters for fur-
how to use the knowledge of the errors to improve the sub- ther calculations.
Suppose the parameter θ = f(x) estimation problem
(given data x) can be abstracted as:
P. Chen ()
School of Information Science and Technology, Sun Yat-sen F (θ, x) = 0 (1)
University, Guangzhou, China
e-mail: [email protected] We are specifically interested in the special case

P. Chen Aθ = 0 (2)
Shenzhen Institute of Advanced Integration Technology,
CAS/CUHK, Shenzhen, China where A is a linear operator (matrix) formed by the data
D. Suter
values.
ARC Centre for Perceptive and Intelligent Machines in Complex If we had the more simple (to analyze) case
Environments, Department of Electrical and Computer Systems
Engineering, Monash University, Melbourne, Australia θ = f(x) (3)
282 J Math Imaging Vis (2009) 33: 281–295

then conceptually, the computation of the covariance matrix tion problems that can be abstracted as solving a (homoge-
of the errors in θ can be expressed as: neous) linear equation (2), where the singular vector associ-
 T ated with the least singular value is the estimate of the pa-
∂θ ∂θ rameters of interest (essentially the Direct Linear Transform
θ = Cx (4)
∂x ∂x or DLT algorithm—see below). In particular, we focus on
homography estimation. This is because if one derives a se-
where Cx is the covariance matrix of x.
ries of such homographies, from the same pair of images; in
Although (4) looks direct and simple, it is not always the
case that there are direct ways available to compute the par- principle, one can improve the accuracy of the estimated ho-
tial differentiation in (4). In particular, we do not have such mographies by further statistical refinement (by exploiting a
an explicit expression when calculating the singular vector rank constraint). However, such a refinement stage requires
of a matrix associated with its least singular value. estimates of the error correlations between the homography
In related work [13, 14], Haralick proposed to calculate coefficients—which is thus a useful example to illustrate our
the propagations of perturbations in observed variables x, in analysis.
the minimization problem of In Sect. 2, we first review the normalized Direct Linear
Transformation (DLT) algorithm [15] and the subspace con-
min F (θ, x) (5) strained homography estimation [4]. In Sect. 3, we present
how to analytically compute the statistical property of the er-
The perturbation of x results in an error in θ , θ , as: rors in estimated homography parameters, and generally in
 −1 other linearized parameter estimation problems. In Sect. 4,
∂ 2F ∂ 2F
θ = − x (6) we present simulations that fit the analytically calculated
∂θ 2 ∂θ ∂x statistics very well. In Sect. 5, the usefulness of this statis-
Then, the covariance matrix of θ can be calculated as tical analysis is demonstrated in the subspace constrained
approach to homography estimation.
 −1  2 T  2 −1
∂ 2F ∂ 2F ∂ F ∂ F
Cθ = Cx (7)
∂θ 2 ∂θ ∂x ∂θ ∂x ∂θ 2
2 Normalized DLT Algorithm and Subspace
However, in many cases, two difficulties prevent us from
Constrained Homography Estimation
directly employing (6) and (7) to calculate Cθ . First, as men-
tioned above, there is no explicit formula for the partials
in (6), for example, when calculating the singular vectors. 2.1 Normalized DLT
2
Second, ∂∂θF2 is not always invertible, as in homography esti-
mation and other parameter estimation problems where the In this section, we review the normalized direct linear trans-
degree of freedom of the parameters is less than the parame- formation (DLT) algorithm [15] in general (and for homog-
ter number. raphy estimation in particular).
We note that there is much work that concentrates on The DLT approach is essentially to take the singular sub-
how to estimate the optimal parameters: including Taubin space as the solution to (2) (and this is invariably by SVD).
method [31], the renormalization method [17–19], the HEIV In what follows, we concentrate on homography estimation
method [22], the FNS method and its variants [7–9], and the but, as far as the DLT goes, the methodology of this paper
equilibration method [24–27]. In [21], it is reported that a applies to any setting (e.g. estimation of fundamental matri-
more accurate estimate can be obtained by taking into ac- ces) that leads to the same form (homogeneous linear equa-
count higher-order error. In [6, 20, 21], a rigorous KCR tion as in (2))—it is just that the particular instantiation of A
(Kanatani-Cramer-Rao) bound for the uncertainty in the es- will be different.
timated parameters is given. For comprehensive reviews on It is well known that the direct approach (via SVD) does
parameter estimation and its applications in computer vi- not always produce good results and that a “normalizing”
sion, see [5, 10, 11, 16, 20, 21, 28, 29, 35–37]. However, step generally improves the results—see the end of this sub-
though many of these target the same general forms as above section (which is explained in the case of homography esti-
(e.g. (2)), they do not generally characterize the error in mation but easily generalizes to other cases).
those estimated parameters. For a homography
In contrast, in this paper our primary focus is to ana- ⎡ ⎤
lytically characterize the uncertainty of the estimated pa- h1 h2 h3
⎢ ⎥
rameters. As said earlier, this is useful when those esti- H=⎢
⎣ h4 h5 h6 ⎥

mated parameters are further fed to follow-up statistical fit-
ting/refinements. We consider the class of parameter estima- h7 h8 h9
J Math Imaging Vis (2009) 33: 281–295 283

which maps x = [x1 x2 1]T on the first view as x = It is well known that one can collect the coefficients of the
[x1 x2 1]T on the second view: x = λHx. homographies between two views into a large, rank four ma-
From x × Hx = 0, each pair of the matches, {xi , xi }, pro- trix H. A brief review of this can be found in Appendix A.
duces a 3 × 9 matrix: In this section, we review how to calculate the homogra-
⎡  xT ⎤
phy embedded in a dimension four subspace [4]. Suppose
0 −xTi x2,i i (just for now) the dimension four subspace basis U ∈ R 9,4
⎢ ⎥
Ai = ⎢ ⎣ x T
i 0 −x  xT ⎥
1,i i ⎦ (8) is known and the linearization matrix is A as in (9). The
subspace constrained DLT solution is as follows: First, cal-
 xT
−x2,i  xT
x1,i 0
i i culate the solution of AUx = 0 as x̂ (standard smallest sin-
gular value way). Second, take the Ux̂ as the solution of
which satisfies Ai h = 0, with h = [h1 h2 . . . h9 ]T . Stack the homography, which is obviously embedded in the sub-
{Ai } as space U.
T As in the normalized DLT, we also use a normalization
A = AT1 ... ATn (9) step in this dimension-four constrained homography estima-
tion. Suppose n (n ≥ 4) planes are available. The subspace
Ah = 0 holds. The solution of h is the right singular vector constrained algorithm is:
of A, associated with the least singular value. This is the
DLT algorithm [15] for homography estimation. 1. Taking all the feature points in the n planes as a whole
In [15], a normalization step is recommended. It con- set, calculate the normalization transforms T and T , for
sists of a translation and a scaling, so that the centroid of the first view and the second view respectively.
the transformed points is the 2. For each normalized plane, calculate its homography.
√ origin (0, 0) and the average
distance from the origin is 2. Suppose the centroid of the 3. Calculate the dimension four subspace U of these homo-
original points is (c1 , c2 ) and the average distance to this graphies.
centroid is l. The normalization transform T is 4. For each normalized plane, calculate its subspace-U con-
⎡1 ⎤ strained homography.
l 0 − cl1 5. Calculate the denormalized homographies for all the
⎢ ⎥ planes, as in the denormalization step of the normalized
⎢ 0 1 − c2 ⎥ (10)
⎣ l l ⎦ DLT.
0 0 1
There are two approaches1 to calculate the dimension-
Similarly, there exists a normalization transform for the sec- four subspace U, in step 3 above. One obvious approach is
ond view, T . to employ the SVD [12] to calculate the dimension four sub-
The normalized DLT algorithm takes the DLT algorithm space2 U of the rank-four matrix H [33, 34]. We refer to
as its core algorithm. First, calculate the transformed points this approach as the SVD-based subspace constrained ap-
for each view and their associated normalization transforms proach, or SVD-Sub-Cnstr, if the SVD is employed in step 3.
T and T . Second, using DLT, calculate the homography H However, the errors in estimated parameters produced by the
from the normalized matches. Last, in the denormalization DLT (step 2), can not be modeled as independent (much less
step, set as i.i.d. Gaussian). Thus, although the estimated subspace
by the SVD method is the “best” in terms of the Frobenius-
H = T−1 HT (11) norm distance, it is generally not optimal.
In the other approach, the statistical properties of the er-
as the homography in the original views. rors in estimated parameters (homographies from step 2) are
utilized to more optimally calculate the subspace U. We re-
2.2 Homography Estimation Embedded in a Dimension fer to this approach as the statistical subspace constrained
Four Subspace approach or Sta-Sub-Cnstr. More formally, with the covari-
ance matrix of the error, we first employ the bilinear ap-
proach [1, 3] or the alternating projection approach [23] to
Thus far, we are not saying anything new. The above proce-
dure can now be considered routine in the computer vision
1 For a practical approach other than Sta-Sub-Cnstr, refer to the algo-
community. However, the settings we are really concerned
with are ones where the output of the DLT is to be used rithm in our companion paper [4], where more constraints are utilized
to produce a more accurate estimate. The algorithm in [4] can be ap-
in further estimations/refinements. Here, one does generally plied, even to the case of as few as three planes.
need to take account of the correlations in the outputs of the 2 We remind the reader that in such a scheme there are now two stages
DLT (and hence our focus on calculating those correlations). of SVD calculation. First in the DLT for individual homographies
We illustrate with homography refinement. (step 2) and here for step 3.
284 J Math Imaging Vis (2009) 33: 281–295

calculate weighted rank-four approximation matrix of H: of xi and xi is corrupted with the noise of (εi,1 , εi,2 ), and
H4 (see Appendix B). Then, the subspace of H4 can rea-  , ε  ), respectively.3
(εi,1 i,2
sonably be taken as a solution of U. The essence of the analysis below is to represent the er-
rors in the estimated homography in terms of the random
 , ε  } for (1 ≤ i ≤ n). Here, we use
variables {εi,1 , εi,2 , εi,1 i,2
3 A Statistical Analysis of the Errors in Estimated the second subscript in εi,• (or εi,•  ) to denote the x or y

Homography coordinates in 2d images.


Using the SVD [12], A is decomposed as:
In this section, we present a statistical analysis of the er-
rors in estimated homography parameters. The covariance A = USVT (12)
matrix of the errors in nine parameters is analytically com-
puted. First, we show how to calculate the covariance matrix where U ∈ R3n,3n , V ∈ R9,9 , UUT = I3n , VVT = I9 ,
of the errors for the DLT algorithm. Then, we extend this and S = diag{s1 , s2 , . . . , s8 , 0} ∈ R 3n,9 . The noise-free ho-
to the normalized DLT algorithm. Finally, we estimate the mography vector is the 9th column of V: v9 . Due to
the noise of {εi,1 , εi,2 } and {εi,1  , ε  }, in x and x re-
noise level. We assume a small noise level so that first order i,2 i i
expansion/approximation techniques can be used. spectively, the error Ei in the ith block of A, Ai , is:

3.1 The Case of the DLT Algorithm

Suppose the matrix A in (9) is obtained from n noise free


feature matches and that the ith noise free feature match

⎡  ⎤
0 0 0 −εi,1 −εi,2 0 Ei,{1,7} Ei,{1,8} εi,2
⎢ ⎥
Ei = ⎢
⎣ εi,1 εi,2 0 0 0 0 Ei,{2,7} Ei,{2,8}  ⎥
−εi,1 ⎦ (13)
Ei,{3,1} Ei,{3,2} 
−εi,2 Ei,{3,4} Ei,{3,5} 
εi,1 0 0 0

where Ei,{1,7} = xi,2  ε 


i,1 + xi,1 εi,2 = −Ei,{3,1} , Ei,{1,8} = 1 ≤ i ≤ n; and the second term in (15) is also a random
   ε −x ε  =
xi,2 εi,2 +xi,2 εi,2 = −Ei,{3,2} , Ei,{2,7} =−xi,1 i,1 i,1 i,1 vector: each entry of which is a linear combination of the

−Ei,{3,4} , and Ei,{2,8} = −xi,1 εi,2 − xi,2 εi,1  = −E
i,{3,5} . 4n random variables. Thus, we can express the errors in the
 in E have been dropped.
Quadratic terms of εi,• εi,◦ i 9 parameters using a 9 × 4n matrix: h ∈ R 9,4n , as is our
Define C as the transformed error matrix: staring point of this analysis.
In order to analytically do this, we represent the error ma-
C = UT EV (14)
trix of E as a linear combination of 4n matrices: a stack
From matrix perturbation theory [30, 32], the first order per- of 4n 3n × 9 matrices, each of which represents the error
 , ε  } for
component in one of 4n “directions”: {εi,1 , εi,2 , εi,1
turbed solution for the DLT algorithm is i,2
1 ≤ i ≤ n.
8
ci,9
ṽ9 = v9 − vi (15) n  
si  
i=1 E= εi,1 E4i−3 + εi,2 E4i−2 + εi,1 E4i−1 + εi,2 E4i (16)
i=1
The second term in (15) is the errors in the estimated para-
meters.
For example, the (4(i − 1) + j )th matrix E4(i−1)+j , for 1 ≤
The entries in Ei are random variables, and so are ci,j .
i ≤ n and 1 ≤ j ≤ 4, is
Consequently, each entry of E, ci,j , is a linear combina-
 , ε  } for
tion of the 4n random variables: {εi,1 , εi,2 , εi,1 i,2 4(i−1)+j T T
0T ... 0T E 0T ... 0T (17)

3 Note that x in Sect. 2.1 is used for the homogeneous representation where 0 is a 3 × 9 zero matrix and only the ith block
of a feature point. By a slight abuse of notation, we will also use x to 4(i−1)+j 4(i−1)+j
represent the feature points in non-homogeneous form: with x and y E is nonzero. From (13), the 3 × 9 matrix E
coordinates as its two entries. can be calculated: see Appendix C.
J Math Imaging Vis (2009) 33: 281–295 285

Then, for each 3n × 9 matrix Ei , calculate Ci as 3.1.1 Replacing Noise Free Data

Ci = UT Ei V (18) When presenting a statistical analysis of the errors in esti-


mated homography above, we assumed that noise free fea-
Substituting (18) and (16) into (14), the transformed error ture points are available. We now examine this assumption.
matrix C is represented as a 3n × 9 random matrix:4 From (15) and (20), each column ξ is a linear combi-
nation of {vi |i < 9}. This means that the matrix h lies in
n   the subspace spanned by these 8 vectors. Consequently, the
 
C= εi,1 C4i−3 + εi,2 C4i−2 + εi,1 C4i−1 + εi,2 C4i
covariance matrix Ch in (23) and (22) has a zero singular
i=1
(19) value and the associated singular vector is the ground truth
homography.
Consequently, the second term in (15) can be computed. In practice, we do not have this knowledge of the ground
Specifically, take the 9 × 1 vector ξj truth data. A practical solution, as adopted in this paper, is
to use noisy data (actually observed) instead. In assessing
8 j
ci,9 the impact of this approximation, we use the following mea-
ξj = − vi (20)
si sures to describe the differences:
i=1
|(h̃ − h)T ui − (h̃ − h)T ũi |
as the j th column of h : h = [ξ1 ξ2 . . . ξ4n ]. (24)
The error in estimated homography is represented in |(h̃ − h)T ui |
terms of the random variables of {εi,1 , εi,2 , εi,1 , ε  } for
i,2 where h̃ and h denote the homographies, calculated from
(1 ≤ i ≤ n): noisy data and noise free data, respectively, and ũi and ui
are the singular vectors of the covariance matrices Ch̃ and
(h) = h e (21) Ch also from noisy data and noise free data, respectively.
    ]T Equation (24) measures the differences of the errors’ pro-
where e = [ε1,1 ε1,2 ε1,1 ε1,2 . . . εn,1 εn,2 εn,1 εn,2 4n×1 . jections upon the directions ũi and ui , i.e., the effect of the
(•) here is used to denote a random vector: the errors in
replacement of noisy data for noise free data. Experiments
the vector of •. Similar usages will appear in the following:
in Sect. 4 show that, for i < 9, the above measure is less
not only for a vector but also for a matrix or a scalar. The than 0.01. This means that the difference introduced by this
general rule is: The symbol of (•) denotes for a general replacement of noisy data for noise free data can be over-
random variable, which can be characterized by a stack of looked.
4n quantities (scalars, vectors or matrices).
From (21), the error covariance matrix is 3.1.2 Out of First-order Perturbation: Second-order Effect
on the 9th Vector u9
Ch = h Cx Th (22)
It is quite another matter when one considers the 9th direc-
where Cx is the 4n × 4n covariance matrix for the noise e tion u9 of Ch . From the calculations and analysis above, it
in the image points. In the special case, where i.i.d. 0-mean- can be seen that, even with noisy data, Ch̃ still has a rank
σ 2 -variance Gaussian noise (in feature point) is assumed, of 8, and its null vector ũ9 is the calculated homography.
the error covariance matrix in the homography is This means that, in the direction of ũ9 , there is no error,
even in noisy data cases. This is obviously not the truth. The
Ch = σ 2 h Th (23) reason for this can be ascribed to the first-order perturba-
tion technique, employed above. Such an effect needs to be
Note that, although the analysis above looks complicated, characterized.
the computation can be greatly reduced by taking into con- Supposing ground truth feature points are available, ũ9
sideration these two facts: From (20), only the first 8 entries can be expressed as, in terms of first order perturbation:
of the 9th column of Cj are needed; and each Ej has only
8
one nonzero 3 × 9 block in (17).
u9 + ςi ui (25)
i=1
4 In (14), E is a random matrix, which can be represented by a series of
matrices {Ei } in (16). The operation on this random matrix E, such as
where ςi  0 for 1 ≤ i ≤ 8. In practice, because of
matrix computations in (14), is decomposed into the same operation on ũ9 F = 1,
the series of matrices {Ei }, as in (18). The transformed random matrix  
C in (14) is a linear combination of transformed Ci , as in (19). This, or u9 + 8i=1 ςi ui u9 + 8i=1 ςi ui
similar operations apply to random variables (vector, scalar, or matrix),
ũ9 =  =  
u9 + 8i=1 ςi ui F 1 + 8i=1 ςi2
in the following of this paper.
286 J Math Imaging Vis (2009) 33: 281–295
 8
 8

1 an error in “raw” data. As done in the homography example,
≈ 1− ςi2 u9 + ςi ui (26) the covariance matrix of the error in estimated parameter θ̂
2
i=1 i=1 is analytically calculated in the following steps:
Thus, the projection of the errors upon the ũ9 direction is
of A:i A = USV , as in (12);
T
1. Calculate the SVD factors
 T 2. Represent the error E = εi E , as in (16);
ũ9 − u9 ũ9 = γ − 3γ 2 + 2γ 3 (27) 3. Calculate Ci = UT Ei V, as in (18);
 j
ci,r−1
where 4. Calculate ξj = − r−1 i si vi , as in (20); and make θ
as: θ = [ξ1 ξ2 . . . ξn ].
8
1 5. Calculate the error covariance matrix by (22) or by (23).
γ= ςi2 (28)
2
i=1
3.2 Extension to the Normalized DLT Algorithm
The projection of the errors upon ui , for i < 9, is:
 T The above analysis ignored the normalization step that is
ũ9 − u9 ui = ςi (1 − γ ) ≈ ςi (29) almost always recommended. We now address this version
of DLT. In the normalized DLT algorithm (11), two factors
It is clear that, from (29), ςi is the errors’ projection upon have to be considered: First, T and T depend on the mea-
ui , i.e., approximately the errors’ projection upon ũi , be- surements and are random matrices; Second, the error in the
cause the difference between these two directions, measured normalized matches will not be i.i.d. Gaussian noise.
by (24), can be ignored. (Note that h̃ and h are ũ9 and u9 , From (11),
respectively).
From (29), the first order perturbation of (25) suffices to (H) = (T−1 )HT + T−1 (H)T
characterize the errors in the directions of ũi , for i < 9.
From (27), the error in the direction of ũ9 is zero in terms of + T−1 H(T) = H e (32)
first order perturbation; however actually, it is not zero, but
In (32), (•) denotes a 3 × 3 random matrix, which can
a second order error, as γ . (Since ςi  1, terms of γ 2 and
be represented by a stack of 4n 3 × 3 matrices. The oper-
γ 3 in (27) can also be ignored.)
ations on the random matrices in (32) are similar to those
To first-order perturbation, ςi is indeed a 0-mean Gaus-
of (14), which are related by (16), (18) and (19).
sian random variable, with its variance as the ith largest sin-
gular value of Ch : λi . Thus, γ in (28) is a chi-square-like
3.2.1 Calculation of (H)
random variable: Its expectation is
8 The critical step to calculate (H) in (32) is to analyze the
1
E(γ ) = λi (30) random variable of the inverse of the scale, as 1l in (10);
2  n n
i=1 (xi,1 −c1 ) +(xi,2 −c2 )
2 2 x
i=1
where l = 2n , c1 = i=1n i,1 and c2 =
n
and its variance is i=1 xi,2
n . Define x i,• as the centered coordinates: x i,1 =
8 xi,1 − c1 and x i,2 = xi,2 − c2 . The error in the centered co-
1
var(γ ) = λ2i (31) ordinates is
2  
i=1
 x i,1 = [−1/n . . . − 1/n (n − 1)/n − 1/n . . . − 1/n]
In order to account for the error γ in the ũ9 direction, we
scale the normalized homography up to a factor of 1 − E(γ ) × [ε1,1 . . . εn,1 ]T
 
and set the 9th singular value of Ch̃ as var(γ ).  x i,2 = [−1/n . . . − 1/n (n − 1)/n − 1/n . . . − 1/n]

3.1.3 Generalization to Other Linearized Systems × [ε1,2 . . . εn,2 ]T

The technique above can be generalized to other parame- where the (n − 1)/n is the ith components. Thus, the error
ter estimation problem, which is abstracted as solving a lin- in the inverse of l is:
ear or linearized system. Suppose the data matrix A ∈ R m,r   n
1 1    
and the parameter θ ∈ R r satisfy the constraint of Aθ = 0,  =− 3 x i,1  x i,1 + x i,2  x i,2 (33)
l 2nl
which approximately holds in practice because the data ma- i=1
trix A is generally corrupted with an error of E. By analyz- x i,•
ing the The normalized image feature is l . The error in it is
 linearization process, E is generally represented as x (x i,• )
E = i εi Ei , where εi is a random variable, representing ( i,•
l ) = x i,• ( 1l ) + l , which can be expressed as
J Math Imaging Vis (2009) 33: 281–295 287

pTi,• e. Similarly, the error in the second normalized view is In matrix terms, (38) can be expressed as:
pT T
i,• e. We stack the vectors p and p as
T
 
h (h) hhT (h)
 = −
P = [pi,1 pi,2 p1,1 p1,2 ... pn,1 pn,2 pn,1 pn,1 ]T h F h F h 3F
 
Pe are the errors in the normalized coordinates. Accord- (h) hhT
= I9 − (39)
ing to (21), h F h 2F

(H) = Pe = H e (34) where (I9 − hhT


) is a projection matrix upon the subspace,
h 2F
where  is calculated as h in (21), however,  is arranged which is orthogonal to h.
as a stack of 4n 3 × 3 matrices. It should be emphasized that the errors in the NON-
normalized h from (32) are not orthogonal to h, in contrast
to the analysis in Sect. 3.1.1. However, from (39), the error
3.2.2 Calculation of (T)
in the normalized h is still orthogonal to h, due to the effect
of the projection matrix in the normalization step (39). This
Another quantity that will be used in calculating (T) and
can be interpreted from the meaning of “normalization”: Be-
(T−1 ) is the error in the centroid of the original feature
cause the normalization seeks to essentially make the Frobe-
points, (ci ):
nius norm of the homography set to 1, there is no random-
n ness in this direction.
1
 (c• ) = εi,• (35) In practice, there is also error in the direction of esti-
n
i=1 mated homography, because we calculate the homography
from noisy data. This effect of the noisy data is same as that
From (10),
in Sect. 3.1.1. And, the analysis concerning the second or-
⎡ (c1 ) ⎤ der effect, in Sect. 3.1.2, also applies for the normalized DLT
( 1l ) 0 −c1 ( 1l ) − l
⎢ ⎥ homography estimation.
(T) = ⎢
⎣ 0 ( 1l ) −c2 ( 1l ) − (c2 )
l

⎦ (36)
0 0 0 3.3 Noise Level Estimation

Substituting (33) and (35) into (36), we calculate T.


From first order approximation, (T + T)−1 = T−1 − Now, we have represented the errors in homography para-
T−1 TT−1 meters in terms of random variables, i.e. the noise in im-
age feature points. For i.i.d. Gaussian noise or general noise
(T−1 ) = −T−1 (T )T−1 (37) in feature points, the covariance matrix of the errors in the
parameters, can be obtained by (23) or (22). To do so, we
where (T ) can be calculated as in (36). need to know some statistical properties of the noise in im-
Substituting (34), (36) and (37) into (32), we obtain age points. Here, we consider the simplest case, where the
(H) and rearrange it as (h). noise in image points is i.i.d. 0-mean-σ 2 -variance Gaussian
noise. In this case, the noise level σ needs to be estimated.
3.2.3 The Effect of the Normalization Step The major tool, as in the sections above, is also the first order
approximation.
In this subsection, we analyze the effect of the normalization Due to the noise in image points, there exists a difference
step on the calculated homography. Here, different from that between the projection Hx and x . In the following, we will
in the normalized DLT algorithm, the normalization step is show that the projection error can also be represented as a
to scale the homography so that its Frobenius norm is 1. random variable, which depends on the noise in the images:
 , ε  } for (1 ≤ i ≤ n).
{εi,1 , εi,2 , εi,1
    i,2
h (h) 1 Suppose that the noise free homography is
 = + h (38)
h F h F h F
⎛ ⎞
9 h1 h2 h3
where (h) is defined in (21), ( 1
)=− 1
i=1 hi × ⎜ ⎟
h F h 3F H=⎜
⎝ h4 h4 h5 ⎟

(hi ), hi is the ith component of h and (hi ) is the ith row
of (h). h7 h8 h9
288 J Math Imaging Vis (2009) 33: 281–295

H projects each point {xi,1 , xi,2 } of the first view on the sec- Due to the noise, the projection upon the second view is
 as an example:
ond view as, by taking xi,1

 h1 ∗ xi,1 + h2 ∗ xi,2 + h3
xi,1 = (40)
h7 ∗ xi,1 + h8 ∗ xi,2 + h9

[h1 + (h1 )] ∗ (xi,1 + εi,1 ) + [h2 + (h2 )] ∗ (xi,2 + εi,2 ) + h3 + (h3 )


(41)
[h7 + (h7 )] ∗ (xi,1 + εi,1 ) + [h8 + (h8 )] ∗ (xi,2 + εi,2 ) + h9 + (h9 )

b+b = b +
According to first order approximation, a+a −
a a
b can be analytically calculated while calculating the homog-
ab raphy using the normalized DLT algorithm.
b2
approximately holds. From this, (41) equals to
We compare this theoretical result with simulations. Of
 A BD course we know the “ground truth data” in simulations.
xi,1 + − 2 (42)
E E Thus, first from noise free data, we calculate its homogra-
phy and the covariance matrix for the errors in estimated
where A = h1 εi,1 +h2 εi,2 +xi,1 (h1 )+xi,2 (h2 )+(h3 ),
homography from (23). After adding i.i.d. Gaussian noise to
D = h1 ∗ xi,1 + h2 ∗ xi,2 + h3 , B = h7 εi,1 + h8 εi,2 +
feature points, we similarly calculate its estimated homogra-
xi,1 (h7 ) + xi,2 (h8 ) + (h9 ), and E = h7 ∗ xi,1 + h8 ∗
phy and the covariance matrix from noisy data. This process
xi,2 + h9 . Note that second order terms, like (h• )εi,◦ , have
 , the with noisy data repeats 20,000 times to obtain enough data
been dropped. Including the noise in the observed xi,1
 . It can be repre-
for statistical properties. Note that the ground truth data is
projection error is actually E A
− BD
E2
− εi,1 the same for these 20,000 times and each time random noise
sented as qTi,1 e. Similarly, from the projection of the second is added to the feature points.
coordinate, qTi,2 e can be obtained. Stack qTi,• as Suppose, by (23), the covariance matrices are C and C (C
is different every time), calculated from noise free feature
Q = [q1,1 q1,2 ... qn,1 qn,2 ]T (43) points and noisy ones respectively:

In practice, the projection error is actually available, C = U diag{λ1 λ2 ... λ8 0}UT and
as μ. Then, Qe = μ approximately holds. Because the i.i.d.
Gaussian noise is assumed here, qi,• 2F σ 2 = μ22(i−1)+• . C = U diag{λ̃1 λ̃2 ... λ̃8 0}UT
Then, the noise level is estimated as:
We calculate three types of indexes, for 1 ≤ i ≤ 8:
μ F
σ̂ = (44) (ũ9 − u9 )T ũi
Q F ρ̃i =  (45)
λ̃i
Note, the noise levels in two views can be assumed as dif-
ferent, up to a known scale. Suppose σ1 and σ2 are the noise (ũ9 − u9 )T ui
ρi = √ (46)
levels in the first and second views, respectively; however, λi
unknown. Suppose, further, σ1 = κσ2 . Then, by multiply- (ũ9 − u9 )T ui − (ũ9 − u9 )T ũi
ing the (4 • +1)th and (4 • +2)th columns of Q by a factor τi = (47)
|(ũ9 − u9 )T ui |
of κ, we can calculate σ̂2 , according to (44). Consequently,
σ̂1 = κ σ̂2 . Note that ũ9 and u9 are the estimated homography from
noisy data and the ground truth homography, respectively,
from the analysis in Sect. 3.1.1. The numerator parts of ρ̃i
4 Simulations of the Errors in the Homography and ρi are the errors projected upon the directions of ui and
Coefficients ũi , respectively. Because C and C are the covariance matri-
ces of the errors of ũ9 − u9 , ρ̃i and ρi should be of 0-mean-
The above concludes our methodology. Our purpose is now 1-variance Gaussian distribution. τi quantifies the difference
to confirm the validity of the correlation information the caused by the replacement of noisy data for noise free data.
above methodology provides, and then demonstrate the ef- The simulations in Fig. 1 and Fig. 2 show that both ρ̃i and
fectiveness of the information. ρi obey the 0-mean-1-variance Gaussian distribution. From
In this section, we carry out simulations to confirm the the fact that ρ̃i in Fig. 1 is a 0-mean-1-variance Gaussian
statistical analysis in Sect. 3. From the theory in Sect. 3, the variable, we can draw a conclusion that the replacement of
statistical properties of the errors in estimated homography noisy data for noise free data can be overlooked. This is also
J Math Imaging Vis (2009) 33: 281–295 289

Fig. 1 Simulations of ρ̃i for 1 ≤ i ≤ 8

 T
confirmed by the simulations in Fig. 3, where we can see the ˜ = ũ9 − u9 ũ9 − γ (49)
magnitude of τi exceeds 0.01 in very few cases.
Consider now the 9th direction. As discussed in Note that in practice, we can only calculate E(γ ). We cal-
Sect. 3.1.2, we scale the normalized homography up to a culate ˜ only for the purpose of validating (28).
factor of 1 − E(γ ) and set the 9th singular value of C as From the analysis in Sect. 3.1.2, ˜ should be almost 0,
var(γ ), as in (30) and (31). Here, we use simulations to val- compared with , because from (27) ˜ only has the 2nd and
idate this approach. We only simulate the projection of the 3rd order terms of γ ; and  is a Chi-square like random vari-
errors upon the direction ũ9 (because of (ũ9 − u9 )T ũ9 = able. From the simulations of  and ˜ in Fig. 4, the error of
−(ũ9 − u9 )T u9 ). We know its expectation E(γ ) and in sim- ˜ can be totally overlooked, compared with  (with a scale
ulations, we can furthermore calculate γ from (29) and (28). up to 10−5 ). This means that the error of (ũ9 − u9 )T ũ9 is al-
Thus, we calculate the following indexes: most modeled by γ in (28). The simulation of  shows that
 T  is a Chi-square like random variable, also confirming the
 = ũ9 − u9 ũ9 − E(γ ) (48) rationality of (30) and (31).
290 J Math Imaging Vis (2009) 33: 281–295

Fig. 2 Simulations of ρi for 1 ≤ i ≤ 8

5 Simulation Result of Subspace Constrained be improved by employing the statistical properties of the
Homography Estimation homography coefficients. Because we need the ground truth
data in the comparison, we also resort to simulations.
It has been shown [2] that, for the case of > 4 planes over First, we compare the subspace constrained homography
2 views, the accuracy of the homographies can be improved
estimation in two cases. One is to use the SVD [12] to cal-
by utilizing the rank 4 constraint. However, the experimen-
culate the rank 4 subspace from more than 4 homographies,
tal setting in [2] was impractical so that we could avoid the
complications that the current paper now addresses (SVD then use the subspace constrained method to refine each ho-
being sub-optimal in the presence of non-i.i.d.-Gaussian mography. We refer to this method as SVD-Sub-Cnstr. An-
noise). In this section, we will show that the mapping ac- other is same except we use the correlation information, de-
curacy of the homographies, in more practical setting, can rived in this paper, in calculating the rank 4 subspace. More
J Math Imaging Vis (2009) 33: 281–295 291

Fig. 3 Simulations of τi for 1 ≤ i ≤ 8

formally, we use the Bilinear approach5 in [1, 3] to calcu- The next experiment shows how the statistical properties
late the rank 4 weighted approximation matrix: H4 . Then can be “intelligently” employed in calculating the rank 4
the subspace spanning H4 is taken as the solution of the subspace. In the simulations above, we add the same level
subspace. We refer to this second method as Sta-Sub-Cnstr. of noise to feature points in all planes. In the following ex-
From Fig. 5, we can see that the general SVD based method periment, we add equal-level noise to the first n − 1 planes,
SVD-Sub-Cnstr even increases the mapping error, compared
and add a much stronger noise in the last plane. By instinct,
with the normalized DLT algorithm. The superiority of the
in the SVD based method SVD-Sub-Cnstr, other planes with
Sta-Sub-Cnstr can be easily seen in Fig. 5.
weak noise will be affected by the plane that is severely pol-
luted by noise. What will happen in the Sta-Sub-Cnstr? Note
5 The alternate projection (AP) approach in [23] achieves the same aim.
that the plane with stronger noise is treated the same as oth-
However, the Bilinear approach in [1, 3] is preferred here because the
errors in each homography can be reasonably assumed to independent ers when using subspace constrained methods, in the SVD-
to those in another homography. Sub-Cnstr or the Sta-Sub-Cnstr; i.e., we do not know the
292 J Math Imaging Vis (2009) 33: 281–295

Fig. 4 Simulations of  and ˜

difference of the last plane from other planes in calculating


the subspace.
Here we consider the cases of n ≥ 6 planes, where the
nth plane is severely corrupted (in this example, five times
noise level as that in other n − 1 planes). We calculate the
homographies in two rounds. First, calculate all the n ho-
mographies, by the normalized DLT algorithm, the SVD-
Sub-Cnstr and the Sta-Sub-Cnstr. In this round, we have to
consider two different mapping errors: one for the mapping
error for the first n − 1 planes (They are denoted as “SVD-
Sub-Cnstr, 1st round”, “Sta-Sub-Cnstr, 1st round” and “Nor-
malized DLT” in Fig. 6a) and the other for the nth plane (in
Fig. 6b). In another round, we calculate the first n − 1 ho-
mographies, by the three methods (discarding the nth plane.)
They are also shown in Fig. 6a, denoted as “SVD-Sub-Cnstr,
2nd round” and “Sta-Sub-Cnstr, 2nd round”. For the normal-
ized DLT algorithm, the results are the same in two rounds. Fig. 5 Simulations of mapping errors, compared with the normalized
In total, we list 8 curves in Fig. 6. DLT algorithm
It should be noted that in Fig. 6a, for example, although
the abscissa is 6 (i.e. n), only the first 5 (i.e., n − 1) planes
are used to calculate the subspace constrained homographies
in the second round. Though, in the first round, we jointly those in the second round. For the SVD based method SVD-
calculate the 6 (i.e. n) homographies, we only calculate the Sub-Cnstr, it is quite another matter: the badly corrupted
performance of the first 5 (i.e. n − 1) homographies. Thus, plane deteriorates the accuracy of the other planes. On the
we can compare the mapping accuracy of the same 5 (i.e.,
first n − 1 planes, the mapping errors in the first round are
n − 1) homographies in these two rounds.
larger than those in the second round (let alone their accu-
The advantage of using the statistical properties in the
Sta-Sub-Cnstr can be found from two aspects. First, the first racy, compared with the normalized DLT or the Sta-Sub-
n − 1 planes are NOT affected by the nth severely corrupted Cnstr).
plane; instead, this badly corrupted plane helps to improve Second, the badly corrupted plane is also improved by
the accuracy of the other n − 1 planes: On the first n − 1 utilizing the statistical properties in the Sta-Sub-Cnstr, as
planes, the mapping errors in the first round are smaller than can be seen from Fig. 6b.
J Math Imaging Vis (2009) 33: 281–295 293

Fig. 6 Simulations of mapping errors, compared with the normalized DLT algorithm, when one plane is severely corrupted with noise

6 Conclusion Note that this is a particular representation (we call it the


canonical representation): all matrices related to this matrix
In this paper, we show how to analytically compute the sta- by a scale are also representations of the same homography.
tistical properties (specifically correlations) of the errors in In some applications, we need to relate the matrix ho-
estimated parameters from the (normalized) DLT. We spe- mography Hi to its vector form hi . Suppose
cialize this to homography parameters. Simulations confirm
⎡ ⎤
the results. To illustrate the usefulness of being able to de- h1,i h2,i h3,i
rive such information, we consider the subspace constrained ⎢ ⎥
method for estimating a collection of homographies. The re- Hi = ⎢
⎣ h4,i h5,i h6,i ⎥

sults of these simulations not only confirm the usefulness but h7,i h8,i h9,i
provide interesting illustrations of precisely how excessive
noise in part of that data influences the outcomes in terms of hi is defined to be hi = [h1,i h2,i . . . h9,i ]T . The matrix
the individual refined parameters. H = [h1 h2 . . . hn t]9,n . whose columns are homographies
Though we have focussed on homography estimation, in canonical form, can be expressed as the following, in
our work is potentially useful in many problems, where the terms of R, t, and {vi }:
estimated parameters are used as the input for further analy-
sis. A direct application may be to employ the same tech- H = vec(RT ) 1 1 . . . 1 1,n
niques in the calculation of the induced dimension-4 homol-
ogy subspace in the cases of two-plane-over-multiple-view − U t v1 v2 ... vn 3,n
(52)
or multiple-plane-over-multiple-view [33, 34].
where
⎡ ⎤
Appendix A: Rank-4 Constraint t1 I3
⎢ ⎥
Ut = ⎢ ⎥
⎣ t2 I3 ⎦
First, we cite the Result 12.1 on p. 312 of [15], which de-
scribes the relationship between a homography and the pro- t3 I3 9,3
jection matrices. Given the projection matrices for 2 views
From (52), the homography matrix H=[h1 h2 . . . hn ]9,n ,
P = [I|0] P = [R|t] (50) has a rank of 4. In addition, the homography matrix H in
and the ith plane defined by πiT X = 0 with πi = [vTi 1]T , (52) has special structure that can be employed to produce
the homography induced by the plane is x = Hi x with a more constraints so that even 2 homographies suffice to cal-
matrix representation: culate the dimension 4 subspace [4]. However, in this pa-
per, we are only interested in the rank 4 constraint, implied
Hi = R − tvTi (51) by (52).
294 J Math Imaging Vis (2009) 33: 281–295

Appendix B: Definition of Weighted Rank-r 8. Chojnacki, W., Brooks, M.J., van den Hengel, A., Gawley, D.:
Approximation Matrix A new approach to constrained parameter estimation applicable to
some computer vision problems. Image Vis. Comput. 22(2), 85–
91 (2004)
We suppose 0-mean noise in the entries of the matrix M ∈ 9. Chojnacki, W., Brooks, M.J., van den Hengel, A., Gawley, D.:
R m,n but we do not assume row or column independence. From fns to heiv: a link between two vision parameter estimation
In order to characterize the noise in M, we first rearrange M methods. IEEE Trans. Pattern Anal. Mach. Intell. 26(2), 264–268
as a vector vec(M) ∈ R mn,1 . Suppose the covariance matrix (2004)
10. Chum, O., Pajdla, T., Sturm, P.: The geometric error for homogra-
for the noise in vec(M) is C. The weighted rank-r approxi- phies. Comput. Vis. Image Underst. 97(1), 86–102 (2005)
mation matrix of M is defined to be Mr that has properties: 11. Chum, O., Werner, T., Matas, J.: Two-view geometry estimation
rank(Mr ) = r and Mr minimizes the objective function of unaffected by a dominant plane. In: Proc. Conf. Computer Vision
(vec(M − X))T C− vec(M − X). Methods for finding rank r and Pattern Recognition (1), pp. 772–779 (2005)
12. Golub, G.H., Loan, C.F.V.: Matrix Computations, 3nd edn. Johns
approximation matrix can be found, as the Bilinear approach Hopkins Press, Baltimore (1996)
in [1, 3] and the alternate projection (AP) approach in [23]. 13. Haralick, R.M.: Propagating covariance in computer vision. In:
Proc. of 12th ICPR, pp. 493–498 (1994)
14. Haralick, R.M.: Propagating covariance in computer vision. Int. J.
4(i−1)+k Pattern Recogn. Artif. Intell. 10(5), 561–572 (1996)
Appendix C: Definition of E in (17) in Sect. 3.1 15. Hartley, R.I., Zisserman, A.: Multiple View Geometry in Com-
puter Vision, 2nd edn. Cambridge Univ Press, Cambridge (2003)
16. Jain, A.K., Mao, J., Duin, R.: Statistical pattern recognition: A re-
⎡  ⎤ view. IEEE Trans. Pattern Anal. Mach. Intell. 22(1), 4–37 (2000)
0 0 0 −1 0 0 xi,2 0 0 17. Kanatani, K.: Unbiased estimation and statistical analysis of 3-d
⎢ ⎥
=⎢ 0⎥
4(i−1)+1  rigid motion from two views. IEEE Trans. Pattern Anal. Mach.
E ⎣ 1 0 0 0 0 0 −xi,1 0 ⎦ Intell. 15(1), 37–50 (1993)

−xi,2 0 
0 xi,1 0 0 0 0 0 18. Kanatani, K.: Statistical bias of conic fitting and renormalization.
⎡  ⎤ IEEE Trans. Pattern Anal. Mach. Intell. 16(3), 320–326 (1994)
0 0 0 0 −1 0 0 xi,2 0 19. Kanatani, K.: Statistical Optimization for Geometric Computa-
⎢ ⎥
=⎢ 0⎥
4(i−1)+2  tion: Theory and Practice. Elsevier, Amsterdam (1996)
E ⎣0 1 0 0 0 0 0 −xi,1 ⎦ 20. Kanatani, K.: Uncertainty modeling and model selection for geo-

0 −xi,2 0 0 
xi,1 0 0 0 0 metric inference. IEEE Trans. Pattern Anal. Mach. Intell. 26(10),
⎡ ⎤ 1307–1319 (2004)
0 0 0 0 0 0 0 0 0 21. Kanatani, K.: Statistical optimization for geometric fitting: Theo-
⎢ ⎥
=⎢ −1 ⎥
4(i−1)+3 retical accuracy bound and high order error analysis. Int. J. Com-
E ⎣0 0 0 0 0 0 −xi,1 −xi,2 ⎦ put. Vis. (2008, in print)
0 0 0 xi,1 xi,2 1 0 0 0 22. Leedan, Y., Meer, P.: Heteroscedastic regression in computer vi-
⎡ ⎤ sion: Problems with bilinear constraint. Int. J. Comput. Vis. 37(2),
0 0 0 0 0 0 xi,1 xi,2 1 127–150 (2000)
⎢ ⎥
=⎢ 0⎥
4(i−1)+4 23. Manton, J.H., Mahony, R., Hua, Y.: The geometry of weighted
E ⎣ 0 0 0 0 0 0 0 0 ⎦ low-rank approximations. IEEE Trans. Signal Process. 51(2),
−xi,1 −xi,2 −1 0 0 0 0 0 0 500–514 (2003)
24. Mühlich, M., Mester, R.: A considerable improvement in non-
iterative homography estimation using tls and equilibration. Pat-
tern Recogn. Lett. 22(11), 1181–1189 (2001)
25. Mulich, M., Mester, R.: The role of total least squares in motion
References analysis. In: ECCV, pp. 305–321 (1998)
26. Mulich, M., Mester, R.: Subspace methods and equilibration in
1. Chen, P.: An investigation of statistical aspects of linear subspace computer vision. In: Scandinavian Conference on Image Analysis
analysis for computer vision applications. Ph.D. Thesis, Monash (2001)
University (2004) 27. Mulich, M., Mester, R.: Unbiased errors-in-variables estimation
2. Chen, P., Suter, D.: An analysis of linear subspace approaches using generalized eigensystem analysis. In: ECCV Workshop
for computer vision and pattern recognition. Int. J. Comput. Vis. SMVP, pp. 38–49 (2004)
68(1), 83–106 (2006) 28. Nadabar, S.G., Jain, A.K.: Parameter estimation in Markov ran-
3. Chen, P., Suter, D.: A bilinear approach to the parameter estima- dom field contextual models using geometric models of objects.
tion of a general heteroscedastic linear system, with application to IEEE Trans. Pattern Anal. Mach. Intell. 18(3), 326–329 (1996)
conic fitting. J. Math. Imaging Vis. 28(3), 191–208 (2007) 29. Nayak, A., Trucco, E., Thacker, N.A.: When are simple ls estima-
4. Chen, P., Suter, D.: Rank constraints for homographies over two tors enough? An empirical study of ls, tls, and gtls. Int. J. Comput.
views: Revisiting the rank four constraint. Int J. Comput. Vis. Vis. 68(2), 203–216 (2006)
(2008, to appear) 30. Stewart, G.W., Sun, J.G.: Matrix Perturbation Theory. Academic
5. Chernov, N.: On the convergence of fitting algorithms in computer Press, San Diego (1990)
vision. J. Math. Imaging Vis. 27(3), 231–239 (2007) 31. Taubin, G.: Estimation of planar curves, surfaces, and nonplanar
6. Chernov, N., Lesort, C.: Statistical efficiency of curve fitting algo- space curves defined by implicit equations with applications to
rithms. Comput. Stat. Data Anal. 47(4), 713–G728 (2004) edge and range image segmentation. IEEE Trans. Pattern Anal.
7. Chojnacki, W., Brooks, M.J., van den Hengel, A., Gawley, D.: On Mach. Intell. 13(11), 1115–1138 (1991)
the fitting of surfaces to data with covariances. IEEE Trans. Pattern 32. Wilkinson, J.H.: The Algebraic Eigenvalue Problem. Clarendon,
Anal. Mach. Intell. 22(11), 1294–1303 (2000) Oxford (1965)
J Math Imaging Vis (2009) 33: 281–295 295

33. Zelnik-Manor, L., Irani, M.: Multi-view subspace constraints on research interest includes subspace analysis in computer vision, struc-
homographies. In: Proc. Int’l Conf. Computer Vision, pp. 710–715 ture from motion, and wavelet application in image processing.
(1999)
34. Zelnik-Manor, L., Irani, M.: Multi-view subspace constraints on David Suter holds the position
homographies. IEEE Trans. Pattern Anal. Mach. Intell. 24(2), of Professor of Computer Science
214–223 (2002) in the School od Computer Sci-
35. Zhang, Z.: Parameter estimation techniques: A tutorial with appli- ence at The University of Adelaide,
cation to conic fitting. Image Vis. Comput. 15, 59–76 (1997) South Australia. He was previously
36. Zhang, Z.: Determining the epipolar geometry and its uncertainty:
Professor of Computer Systems in
A review. Int. J. Comput. Vis. 27(2), 161–195 (1998)
the Department of Electrical and
37. Zhang, Z.: On the optimization criteria used in two-view motion
Computer Systems Engineering at
analysis. IEEE Trans. Pattern Anal. Mach. Intell. 20(7), 717–729
Monash University. During 2008-
(1998)
2010, Professor Suter will also serve
as a member of the Australian Re-
search Council College of Experts.
Pei Chen received two Ph.D. de-
He is a Senior Member of the IEEE.
grees, on wavelets and computer vi-
His main research interests are Im-
sion, respectively, from Shanghai
age Processing, Computer Vision,
Jiaotong University in 2001, and
Video Compression, Computer Graphics and Visualization, Data Min-
from Monash University in 2004.
ing and Artificial Intelligence. He currently serves on the editorial
He worked as a postdoctoral re-
board of four international journals: Journal of Mathematical Imag-
searcher with Monash University;
ing and Vision; Machine Vision and Applications, IPSJ Transactions
as a Senior Research Engineer with
on Computer Vision and Applications, and the International Journal of
Motorola Labs; then as a Research
Computer Vision. He was previously a member of the editorial board
Professor with Shenzhen Institute of
of the International Journal of Image and Graphics.
Advanced Integration Technology,
CAS/CUHK, China. He is currently
a Professor with School of Informa-
tion Science and Technology, Sun
Yat-sen University, China. His main

You might also like