0% found this document useful (0 votes)
81 views18 pages

Extrinsic Parameters Calibration Method of Cameras PDF

Uploaded by

ghaieth
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
81 views18 pages

Extrinsic Parameters Calibration Method of Cameras PDF

Uploaded by

ghaieth
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 18

Article

Extrinsic Parameters Calibration Method of Cameras


with Non-Overlapping Fields of View in Airborne
Remote Sensing
Lei Yin1,2, Xiangjun Wang1,2,*,Yubo Ni2, Kai Zhou 2 and Jilong Zhang 2
1 State Key Laboratory of Precision Measurement Technology and Instruments, Tianjin University, No. 92
Weijin Road, Nankai District, Tianjin 300072, China; [email protected]
2 MOEMS Education Ministry Key Laboratory, Tianjin University, No. 92 Weijin Road, Nankai District,

Tianjin 300072, China; [email protected](Y.N.); [email protected](K.Z.); [email protected] (J.Z.)


* Correspondence: [email protected]; Tel.: +86-022-2740-3395

Received: 02 July 2018; Accepted: 13 August 2018; Published: 16 August 2018

Abstract: Multi-camera systems are widely used in the fields of airborne remote sensing and
unmanned aerial vehicle imaging. The measurement precision of these systems depends on the
accuracy of the extrinsic parameters. Therefore, it is important to accurately calibrate the extrinsic
parameters between the onboard cameras. Unlike conventional multi-camera calibration methods
with a common field of view (FOV), multi-camera calibration without overlapping FOVs has certain
difficulties. In this paper, we propose a calibration method for a multi-camera system without
common FOVs, which is used on aero photogrammetry. First, the extrinsic parameters of any two
cameras in a multi-camera system is calibrated, and the extrinsic matrix is optimized by the re-
projection error. Then, the extrinsic parameters of each camera are unified to the system reference
coordinate system by using the global optimization method. A simulation experiment and a
physical verification experiment are designed for the theoretical arithmetic. The experimental
results show that this method is operable. The rotation error angle of the camera’s extrinsic
parameters is less than 0.001rad and the translation error is less than 0.08 mm.

Keywords: stereo vision; camera calibration; extrinsic parameter; non-overlapping field of view;
airborne remote sensing

1. Introduction
Photogrammetry is widely used in aerial remote sensing [1–3]. More and more airborne remote
sensing platforms are equipped with multi-camera systems to complete tasks such as image mosaic
[4], terrain reconstruction [5,6] and monitoring [7–9]. In some multi-camera remote sensing
photography systems, the angle between the optical axes of the cameras is large enough to obtain a
large field of view (FOV), resulting in the non-overlapping of the FOV, and increasing the calibration
difficulty of extrinsic parameters. However, the quality of terrain reconstruction and panoramic
mosaic tasks depend on the accuracy of the camera extrinsic parameters. To solve this problem, we
introduce a universal multi-camera extrinsic calibration method without common FOVs. The
commonly used method for calibrating the extrinsic parameters is based on a two-dimensional target
proposed by Zhang Z [10]. This method calculates the extrinsic matrix by placing the target in the
public FOV of the cameras. In addition, there are also some calibration methods based on other
targets such as: 1D target [11], plane-based targets [12,13], stereo targets [14,15], spheres target [16,17]
and so on. When calibrating multiple cameras, the above method requires that the different cameras
must have overlapping FOV, and the traditional calibration method is no longer applicable when the
cameras have no common FOV, because the same target cannot be observed by different cameras.
Remote Sens.2018, 10, 1298;doi:10.3390/rs10081298 www.mdpi.com/journal/remotesensing
Remote Sens. 2018, 10, 1298 2 of 18

To solve this problem, Gong et al. [18] and Dong et al. [19] used the three-dimensional
reconstruction method to calculate the extrinsic matrix of cameras without a common FOV. These
authors reconstruct the coding landmarks on the calibration plan by 3D scanner or visual method.
Cameras without a common field of view capture different section of the plane, the non-overlapping
FOVs are connected by the coding marks, and then the extrinsic parameter matrix can be solved. This
method requires reconstruction of the large calibration plane in advance. The reconstruction process
is complex and the calibration accuracy depends on the reconstruction accuracy. Moreover, it is
difficult to ensure the reconstruction accuracy for the large-sized target planes. Xia et al. [20] used
two fixed targets connected by metal rods to calibrate the non-overlapping-field camera. First, two
targets are captured by a large FOV camera, and the pose relationship between the two fixed targets
is computed by the pose recovery method. Then, the transformation matrix between targets is used
to solve the extrinsic parameters between the cameras. This method requires a special target and is
only suitable for binocular camera calibration. When the angle between the optical axes of the two
cameras is large, it is difficult to create the target. Lu et al. [21] used a theodolite pair called a
theodolite coordinate measurement system (TCMS) to calibrate the extrinsic parameters of the
cameras. A calibration block was placed in the camera’s FOV by using the theodolite to measure the
spatial coordinates of the marker points on the target. The transform matrix between the target
coordinate system and the global coordinate system can be obtained, and then the transform matrix
between the camera coordinate system and the global coordinate system can be obtained.
Lébraly et al. [22] introduced a calibration method by using plane mirrors. The key of this
method is to build a virtual camera through the reflection of the flat mirror and to use the virtual
camera to associate real cameras. There are some marks on the plane mirror, and the pose of the
virtual camera relative to the real camera can be calculated by these marks. However, when the
number of cameras is large, it is difficult to control the reflection scope of the flat mirror. Kumar et al.
[23] also calibrated multiple cameras using flat mirrors. The method removes the need for cameras
to observe the calibration object directly. The advantage of this method is that there is no need to
place markers on the mirror and through a novel set of constraints the mirror pose is not needed for
calibration.
Pagel et al. [24] introduced an extrinsic calibration method of multiple cameras with non-
overlapping views in vehicles. In this paper, the calibration problem is divided into the problems of
visual odometry and Hand-Eye Calibration. Visual odometry is calculated by the structure from the
motion method, which is based on the corresponding image features of each camera, and then the
visual odometry is optimized by sparse bundle adjustment (SBA). An Iterated Extended Kalman
Filter (IEKF) algorithm is used to calculate the uncertainties of the transformation, and merge the
redundant information to improve the calibration accuracy. By using the IEKF algorithm the extrinsic
parameters can be continuously estimated and the system is able to perform a self-calibration. Liu et
al. [25] proposed a calibration method for vision sensors with non-overlapping FOVs using a spot
laser. By placing the laser in suitable positions, the laser line passes through the FOVs of multiple
vision sensors. Utilizing the plane target to obtain the spatial coordinates of the intersection point of
the laser line and the target in each camera coordinate system, the equation is established by using
collinear conditions of the intersection points. By moving the laser line at least two positions,
sufficient constraint equations can be established to solve the extrinsic parameter matrix.
In this paper, we propose a calibration method of multi-cameras without a common field of view
based on multiple chessboard targets. The difference from this method and the traditional non-
common-field calibration method is that the target does not need to be rigidly connected through a
mechanical structure or with the aid of other auxiliary equipment, and the configuration is more
flexible. This method has high accuracy and is suitable for the on-site calibration of the remote sensing
imaging system. The number of chessboard targets is consistent with the number of cameras, so that
each camera corresponds to a target. We keep the target stationary during the calibration process,
and make the multi-camera system move multiple positions and shoot their corresponding targets.
Sufficient equations can be obtained to solve the extrinsic parameters by moving at least two
locations.
Remote Sens. 2018, 10, 1298 3 of 18

2. Materials and Methods

2.1. Calibration Principle


When calibrating the multi-camera system used in airborne remote sensing, we select one
camera as the reference coordinate system. The calibration purpose is to calculate the extrinsic matrix
of the remaining cameras relative to the reference coordinate system. To reduce the accidental error
introduced by single measurement and obtain robust measurement results, the method is divided
into two steps: the relative extrinsic parameters between any two cameras are calculated first, and
then the extrinsic parameters of each camera relative to the reference coordinate system are obtained
using the redundant data based on the two camera calibration result.

2.1.1. Introduction of Two Cameras Calibration Method


The calibration principle of the extrinsic parameters is shown in Figure 1. First, we use the Zhang
Z. [10] method to calibrate the intrinsic parameters of each camera. Without loss of generality, this
article uses a 5-camera system as an example to introduce a multi-camera system calibration method
without the common field of view used in airborne remote sensing.

Rm,tm
Target1 Target2

RTC1(1 i ) tTC1(1 i ) RCT2(2 i) tCT2(2 i)


R c ,tc
Camera1 Camera2

Camera5 Camera3

Target5 Camera4
Target3

Target4

Figure 1. Extrinsic parameters calibration schematic.

As shown in Figure 1, the FOVs of the five cameras do not overlap with each other, and the 5
cameras correspond to 5 checkerboard targets respectively. We first select any two cameras (such as
camera1 and camera2) to introduce the calibration method between two cameras in the system. In
the figure, Rc and tc are the rotation matrix and translation matrix between the two cameras, and
Rm and tm are the rotation and translation matrix between the targets. RTC1(1 i ) , tTC1(1 i ) , RCT 2(
2 T2
i ) , tC 2(i ) are

the rotation and translation matrix between the cameras and their corresponding target at the ith
position of the multi-camera system. PTT22 is the coordinates of the corner points of target 2 in the
target 2 coordinate system; PTT12 is the coordinates of the corner points of target 2 in the target 1
coordinate system. It can be seen from the figure that the relationship between PTT22 and PTT12 can be
transformed by two different paths.
Remote Sens. 2018, 10, 1298 4 of 18

R tm  T 2
PTT12 =  m  PT 2
1 
(1)
0

 R C1 tTC1(1 i )   Rc tc   RCT 22(i ) tCT 2(


2
 T2
PTT12 =  T 1(i )  
i)
  PT 2
1   0
(2)
 0 1  0 1 

A set of equations for the unknown Rc , t c , Rm , tm can be obtained from Equations (1) and(2):

 RTC1(1 i ) tTC1(1 i )   Rc tc   RCT 22( i ) tCT 2(2 i )   Rm tm 


   =
1   0 1 
(3)
 0 1  0 1  0

The rotation and translation matrix RTC1(1 i ) , tTC1(1 i ) , RCT 2(


2 T2
i ) , tC 2(i ) between cameras and the
corresponding targets can be solved by the EPNP method [26]. The 4 pending matrices Rc tc Rm tm
contain 24 unknown parameters, sorting Equation (3) can obtain 12 linear independent equations
with unknown quantity. Enough constraint equations can be obtained by moving the multi-camera
system at least twice, and we use the linear method to solve the Rc , tc , Rm , tm matrix. It should be
noted that when moving the system, we should avoid pure translational motions, which will result
in the independence of Equation (3), and makes the solution results unstable.
 Rc tc 
Matrix  T
1 
is the extrinsic matrix between the two cameras.
0

2.1.2. Optimization Method of Extrinsic Parameters Based on Reprojection Error


In the above linear solution process, the orthogonal constraint properties of the rotation matrix
are not considered, and the solution results are susceptible to noise. Therefore, we use the linear
solution result as the initial value, and an objective function is established by using the reprojection
error. Then, we optimize the extrinsic parameter matrix between two cameras by using the
Levenberg–Marquardt algorithm [27]. In the optimization process, each rotation matrix is
represented by Euler angles to ensure that the solution results meet the orthogonal constraints. The
principle of the reprojection error between the two cameras is shown in Figure 2.

R m , tm P2
P1

Target1 Path1 Target2

RTC11 tTC11 RCT 22 tCT 22


path2

Camera1 Camera2
Rc , tc

Figure 2. Reprojection error optimization schematic.

The coordinate of the jth corner point on target 1 is transformed to the camera 1 coordinate
system by path 2, and we can obtain the re-projected image coordinate Pcm1,1 j of the corner point from
Remote Sens. 2018, 10, 1298 5 of 18

the intrinsic parameter matrix of camera 1. The coordinate of the corner point on target 2 is
transformed to the camera 2 coordinate system by path 1, and we can obtain the re-projected image
coordinate Pcm2,2 j of the corner point from the intrinsic parameter matrix of camera 2. According to
the camera model, we have:
−1
R tc   RCT 22 tCT 22   Rm tm 
Pcm1 1 = A1  I 0  c  
1   0 1 
P1,j (4)
0 1  0
−1 −1
R tc   RTC11 tTC11   Rm tm 
Pcm2 2 = A2  I 0  c    (5)
1   0 1 
P2,j
0 1 0

The reprojection error equation is:

(
f ( Rc,Tc ) = min( j =1 d Pcm1,1 j , Pˆcm1,1 j ) + ( )
2 2
d Pcm2,2 j , Pˆcm2,2 j )
n n
j =1
(6)

where A1 and A2 denote the intrinsic parameters matrix of the two cameras, P1,j and P2,j denote the
coordinates of the jth corner point on the two targets in the target coordinate system, I denotes a 3 ×
3 unit matrix. Pˆ m1 and Pˆ m2 denote the image coordinates, which remove the lens distortion. By
c1,j c 2, j

using Levenberg–Marquardt method [27] to solve Equation (6), the optimized extrinsic parameter
matrix can be obtained.

2.2. The Global Optimal Method of Multi-cameras Calibration


In this section, we introduce the global calibration method of multi-cameras. Without loss of
generality, we select the camera coordinate system of camera 1 in Figure 1 as the reference coordinate
system of the multi-view system. We can calculate the extrinsic parameter matrix of an arbitrary
camera relative to camera 1 by using the methods mentioned in Section 2.1. In this way, the extrinsic
matrix of the cameras relative to the reference coordinate system can be obtained. However, this way
cannot make full use of redundant motion information, and any accidental calibration noise will
affect the calibration result. To solve this problem, a robust calibration method is proposed in this
paper. First, the extrinsic matrix between any two cameras is calculated using the method described
in Section 2.1, and then the redundant extrinsic parameter data are used to optimize the extrinsic
matrix of each camera relative to the reference coordinate system by the global optimization method.
In this way, for a multi-camera system, the effect of single calibration error on the final calibration
results can be effectively reduced, and the calibration accuracy can be improved.

2.2.1. Global Optimization Method of Rotation Matrices in Extrinsic Parameters.


To obtain accurate global calibration data, the rotation matrix and the translation matrix are
optimized separately. The optimization of rotation matrix is inspired by refer [28], in which the
relative pose relationship between sequential images is improved and this method is used for 3D
reconstruction. We refer to the idea of this method and optimize the extrinsic parameters of the
cameras relative to the reference coordinate system. Ri denotes the rotation matrix of the camera i
relative to the reference coordinate system. Rij denotes the rotation matrix between camera i and
camera j. By using the method in the Section 2.1, we can obtain the rotation matrix Rij between any
two cameras. And then, we have:
R j = Rij Ri (7)

The above formula can be divided into 3 parts:


Remote Sens. 2018, 10, 1298 6 of 18

r1 j − Rij r1i = 031


r2j − Rij r2i = 031 (8)
r3 − R r = 031
j
ij 3
i

where r1i , r2i , r3i denote columns of Ri ,where Ri =  r1i r2i r3i  , Equation (8) can be arranged into
a form of linear equations:

 r1i 
 i
 r2   0 
 ri  31

 - Rij I 33   3j  =  031  (9)


 r1   0 
 r j   31 
 2j 
 r3 

The rotation matrices between any two cameras can be organized into the equation as shown in
Equation (9). Taking a 5-camera system as an example, C (5, 2) =10 equations as shown in Equation
(9) can be obtained. The 10 sets of equations can be combined and written into an over-determined
linear equation like Ax = 0 . Where A is a matrix of 30 × 15 dimensions that is made by stacking all
relative rotation matrices Rij and x is made by stacking all columns of camera extrinsic parameters Ri .
By solving the linear equation, the rotation matrix of each camera relative to the reference coordinate
system can be obtained. The orthonormality constraints are forced by projecting the approximate
rotation to the closest rotation in the Frobenius norm using SVD. In this case, the global rotation
parameters of each camera are obtained by 10 sets of relative rotation, which can effectively reduce
the influence of single measurement noise on global calibration accuracy because of the use of the
redundancy rotation matrix.

2.2.2. Global Optimization Method of Translation Vector in Extrinsic Parameters.


Based on the global rotation extrinsic parameters, in this section the optimization method of
global translation extrinsic parameters is introduced. Using the method mentioned in Section 2.1, we
can obtain the translation vector between any two cameras. Where tij denotes the translation vector
between camera i and camera j. c i denotes the translation vector of camera i in the reference
coordinate system. Then, we have [29]:
R j (ci − c j )=tij (10)

ci − c j = R j −1tij (11)

Equation (11) can be organized as

 ci1 
 2
ci   t 
 c3  1

 I33 | -I33   i1  = t2  (12)


c j   
c 2  t3 
 3j 
 c j 

where ci = ci1 ci2 ci3  , c j = c1j c 2j c 3j  , R j −1tij = t1 t2 t3  . The translation vector between any two
T T T

cameras can be organized into the equation as shown in Equation (12). Taking a 5-camera system as
an example, C (5, 2) =10 equations as shown in Equation (12) can be obtained. The 10 sets of equations
can be combined and written into an over-determined linear equation like Ax = b . Where A is a
matrix of 30 × 15 dimensions that is made by stacking identity matrix, and x is made by stacking
Remote Sens. 2018, 10, 1298 7 of 18

all camera translation parameters ci . b is a vector of 30 × 1 dimensions that is made by the elements
in vector R j −1tij . By solving the linear equation, we can obtain the translation vectors ci , (i = 1, 2  5)
of each camera relative to the reference coordinate system. In this case the influence of single
measurement noise can effectively be reduced by using the redundant translation vector.
After obtaining the matrix Ri and vector ci through the above two subsections, we can obtain
 Ri ci 
the extrinsic matrix of each camera relative to the reference coordinate system 
1 
.
0

3. Results

3.1. A Simulation Experiment of Two Cameras


In this paper, the above theory and calibration accuracy are analyzed by simulation experiment
and physical experiments respectively. First, a two-camera calibration model is simulated and tested.
For all experiments, we use an internally calibrated camera with a focal length of 16 mm and pixel
size of 4.8 um×4.8 um. The image resolution is 1280 × 1024, the principal point is in the ideal optical
center. The target contains 12 × 12 grid of feature points and the interval between points is 30 mm in
both directions. We compare the accuracy of the linear solution and the optimization of re-projection
error algorithm solution. In the calculation process, the different levels of noise are added to pixel
coordinates of the corner points, which are extracted from images. The noise is divided into 10 levels
from 0 pixels to 2 pixels, to verify the robustness of the proposed algorithm. The noise values are
Gaussian distributed with a standard deviation of 0–2 pixels in 10 steps.
In the simulation experiment, the true value of the three relative rotation angles between the two
cameras are R = [0.5233 rad, 0.6977 rad, 0.6977 rad], the true value of the translational vector is T =
[106 mm, −5 mm, 2 mm]. During the simulation the cameras moved ten times. The pose relationship
between the camera and target is calculated using the EPNP [26] algorithm. For each noise level, 100
tests are carried out. The RMS errors of the three rotation angles and the three translations at different
noise levels is shown in Figure 3. The ordering of the Euler angles in experiment is Roll, Yaw, and
Pitch.
Experimental results show that the linear algorithm is easily affected by noise. With the increase
in noise level, the error of rotation angle increases exponentially, and the error of translation oscillates
greatly. When the re-projection error optimization method and the orthogonal constraint are taken
into account, the rotation angle error is obviously suppressed, and the improvement of the yaw angle
accuracy is most obvious. At the same time, the translation error has also been suppressed to a certain
extent.

(a) (b)
Remote Sens. 2018, 10, 1298 8 of 18

(c) (d)
Figure 3. The RMS error curve of the calibration result affected by noise. The graph (a) shows the
three rotation angle error curves obtained with the linear solution method with ten noise levels. The
graph (b) shows the translation error curve obtained with the linear solution method with ten noise
levels. The graph (c) is the rotation angle error curve after optimization of the re-projection error. The
graph (d) is the translation error curve after optimization of the re-projection error.

3.2. Simulation Experiment of Multi-Camera Extrinsic Parameters Calibration with the Global Optimization
Method
After calculating the extrinsic parameter matrix between any two cameras, the global
optimization method is used to calculate the extrinsic parameter matrices of all cameras relative to
the reference coordinate system. We analyzed the error variation of the extrinsic parameters before
and after global optimization.
The simulation experiment takes a five-camera system as an example. Using the method in
Section 2.1 to calculate the extrinsic matrix between any two cameras, and we can obtain 10 sets of
extrinsic parameters. Ten levels of noise are added to the 10 sets of extrinsic parameters. The noise
added to the rotation angles range from 0.001rad to 0.01rad, and the translation noise ranges from
0.05 mm to 0.5 mm. For each noise level, 100 tests are carried out. The extrinsic parameters of the five
cameras relative to the reference coordinate system are shown in Table 1. Without loss of generality,
we take camera 3 as an example, before and after optimization, the changes in the RMS errors of the
extrinsic parameters are shown in Figure 4.

Table1. The extrinsic of each camera relative to the reference coordinate system.

Extrinsic Parameters
Camera Roll Yew Pitching X-Axis Y-Axis Y-Axis
Number Angle Angle Angle Translation Translation Translation
(rad) (rad) (rad) (mm) (mm) (mm)
1 0 0 0 0 0 0
2 0.698 0.175 0.157 500 −100 10
3 1.484 0.262 0.367 600 300 −50
4 2.426 0.559 0.716 −10 600 20
5 −1.047 −0.070 0.401 −590 400 20
Remote Sens. 2018, 10, 1298 9 of 18

(a) (b)

(c) (d)
Figure 4. The extrinsic parameters’ RMS error curve of camera 3 before and after global optimization.
The graph (a) shows the three rotation angle error curves obtained before optimization with ten noise
levels. The graph (b) shows the translation error curve obtained before optimization with ten noise
levels. The graph (c) is the rotation angle error curve after optimization with ten noise levels. The
graph (d) is the translation error curve after optimization with ten noise levels.

The experimental results show that the global optimization method can effectively suppress the
random noise of single calibration. Because the noise of this simulation experiment is directly added
to three Euler angles and three translations of the extrinsic parameters, so the error has a linear
relationship with the noise level before optimization. It can be seen from the graph that the rotation
angle error is obviously suppressed after optimization. From the figure we see within a certain noise
level, the relationship between the rotation angle RMS error and noise level is linear. From Section
2.2.2 we can see that since the global optimization of the translation vector depends on the rotation
angle, the noise of the rotation angle has a great influence on the translation vector. The experimental
results show that when the noise level is higher and the rotation angle error increases, the error of
the translation vector increases more intense.
To illustrate the applicability of the algorithm to all cameras, we give the extrinsic parameters
errors of the remaining cameras relative to the reference coordinate system (camera coordinate
system of the camera 1) at different noise levels (level 2 and level 5 which are consistent with Figure
4). The errors before and after optimization are shown in Table 2.
From the Table 2, we can see that the optimization results of the remaining cameras are
consistent with the results of the camera 3 as described in Figure 4. The calibration accuracy of all the
cameras have improved after global optimization.
Remote Sens. 2018, 10, 1298 10 of 18

Table 2. The errors before and after optimization of the remaining cameras.
Camera 2 Camera 4 Camera 5
Noise Extrinsic
Before After Before After Before After
Level Parameters
Optimization Optimization Optimization Optimization Optimization Optimization
a (rad) 0.00194 0.00135 0.00208 0.00165 0.00183 0.00122
b (rad) 0.00192 0.00114 0.00210 0.00194 0.00174 0.00116
Level c (rad) 0.00203 0.00156 0.00199 0.00118 0.00207 0.00095
2 x (mm) 0.103 0.064 0.121 0.071 0.104 0.063
y (mm) 0.089 0.077 0.009 0.069 0.009 0.054
z (mm) 0.101 0.068 0.110 0.078 0.100 0.077
a (rad) 0.00506 0.00398 0.00400 0.00334 0.00512 0.00265
b (rad) 0.00513 0.00351 0.00523 0.00416 0.00501 0.00296
Level c (rad) 0.00498 0.00324 0.00491 0.00359 0.00523 0.00312
5 x (mm) 0.251 0.172 0.254 0.192 0.249 0.156
y (mm) 0.262 0.150 0.251 0.209 0.250 0.165
z (mm) 0.249 0.149 0.236 0.186 0.251 0.227

3.3. Two Camera Experiment With Real Image


The above simulation experiment tests the robustness of the algorithm. In order to verify the
validity of the algorithm, the following physical experiments are designed to verify the algorithm
described in the article. First, we verify the calibration method of the binocular system. The method
described herein can also be applicable to calibration cameras with a common field of view. In order
to obtain the true value of the extrinsic parameters, a binocular camera system with partially
overlapping FOVs is used to verify the method. The intrinsic parameters of the two cameras are
shown in Table 3. Because the FOVs of the cameras are partially overlap, the Calibration Toolbox for
Matlab [30] method is used to calibrate the extrinsic parameters of the two cameras as true values.
The true value of the rotation angles is [−0.2024 rad, 0.1583 rad, 0.0191 rad], and the true value of the
translation is [−215.709 mm, −1.769 mm, 13.586 mm]. To obtain accurate results in the experiment, we
have evaluated the accuracy of the Calibration Toolbox, the reprojection error of the checkerboard
corners is less than 0.3 pixels. At the same time, the Calibration Toolbox can calculate the confidence
interval for the calibration results. The confidence interval of the rotation angles is ±[0.00067 rad,
0.00086 rad, 0.00010 rad], the confidence interval of the translation vector is ±[0.008215 mm, 0.01422
mm, 0.04773 mm]. The above results show that the calibration results of the toolbox have high
precision and the results can be used as a standard to test the proposed algorithm. Then, 15
experiments are performed using the method proposed in this article, as shown in Figure 5. The error
between 15 experimental values and true values is shown in Figure 6.

Table 3. Intrinsic parameters of the two cameras.

Parameters Left Camera Right Camera


Focal length/pixel [2564.60, 2564.09] [2571.31, 2570.61]
Principal point/pixel [599.75, 492.33] [625.87, 503.24]
Image Distortion coefficients [−0.436, −0.347, 0.00069, 0.0018, 0] [−0.456, −0.109, −0.00023, −0.00044, 0]

The experimental setup is shown in Figure 5, we keep the target position stationary and the
camera is moved 10 positions, and then the extrinsic matrix between the two cameras can be solved.
Remote Sens. 2018, 10, 1298 11 of 18

Figure 5. The two cameras experimental setup.

(a) (b)
Figure 6. The error curves of the calibration results for the two-camera experiment. (a) the three
rotation angle error curves; (b) the translation error curves.

The experimental results show that the error of the three rotation angles is within 0.002 rad,
where the RMS error of the roll angle is 0.0041, the RMS error of the yaw angle is 0.0039, and the RMS
error of the pitch angle is 0.0056. The difference between the mean value and true value of the roll
angle is 9.023 × 10−4, the difference between the mean value and the true value of the yaw angle is
9.417 × 10−4, and the difference between the mean value and the true value of the pitch angle is 0.0014.
The error of the three translational is within 0.08 mm, where the RMS error in the x-direction
displacement is 0.1816, the RMS error in the y-direction displacement is 0.1102, and the RMS error in
the z-direction is 0.2206. The difference between the mean value and the true value in the x direction
is 0.0424, the difference between the mean value and the true value in the y direction is 0.0073, and
the difference between the mean value and the true value in the z direction is 0.0531.

3.4. Real Data Experiment of Global Optimization Calibration for Multi-cameras


The previous section has verified the two camera calibration method, in order to verify the global
optimization process of multi-camera systems, a 4-camera system without common field of view is
built to verify the accuracy of the calibration algorithm based on the reprojection error and global
Remote Sens. 2018, 10, 1298 12 of 18

optimization. To verify the accuracy of the calibration algorithm under real conditions, cameras in
this experiment have no overlapping FOVs. Since the extrinsic parameters’ true value of cameras
without a common field of view cannot be directly calculated, we verify the accuracy of extrinsic
parameters by indirect methods with a coordinate measuring machine. The experimental device is
shown in Figure 7.

Coordinate measuring Measurement


machine point

Camera 2
Target

Camera 4

Camera 1
Camera 3

Figure 7. The left figure shows the experimental setup, and the right figure shows the measurement
points marked by the red box on the target.

In this experiment, 4 checkerboard targets are used to calibrate 4 cameras. First, using the
method described in Section 2.1 to calculate the extrinsic parameter matrix between any two cameras,
a total of 6 combinations can be obtained. Then, using the global optimization method to calculate
the extrinsic parameters of each camera relative to the reference coordinate system through 6 sets of
extrinsic parameter data. Since the position of the camera relative to the corresponding checkerboard
can be calculated by EPNP algorithm [26], we can use the extrinsic parameters between the
checkerboards to obtain the extrinsic parameters between the corresponding cameras, and we used
it as ground truth.
Without loss of generality, we use the extrinsic parameters of camera 1 and camera 2 as an
example to verify the algorithm. The extrinsic parameters between the corresponding checkerboard
targets are obtained using a coordinate measuring machine. The measurement method of the
extrinsic parameters between the checkerboard targets is as follows:
(1) Select 8 corner points on target 1, and measure the coordinates of the 8 points using the
coordinate measuring machine. The position of the eight points on target 1 is shown in Figure 7.
(2) Measure the 8 points on target 2 at the same position as target 1.
(3) Establish constraint equations with Absolute Directional Algorithm [31], then we can obtain
the extrinsic parameters between the two targets using the coordinates of the 8 points. The equation
is shown in Equation (13):
n
min  Rpi + t − qi , subject to RT R = I
2

R ,t
(13)
i =1

where pi denotes the coordinates of corner point on target 1 and qi denotes the coordinates of corner
point on target 2.
The spatial coordinates of the measured points on target 1 and target 2 in the coordinate system
of the three coordinates are shown in Figure 8 and the coordinate values of the 8 points are shown in
Table 4.
Remote Sens. 2018, 10, 1298 13 of 18

Table 4. Coordinate values of measuring points.

Points Target1 Target2


Index X/mm Y/mm Z/mm X/mm Y/mm Z/mm
1 −308.693 374.455 −19.030 −298.538 −31.703 −13.293
2 −334.978 592.731 −27.055 −235.764 179.053 −7.990
3 −319.212 435.772 38.660 −279.511 23.678 48.073
4 −331.265 534.998 35.012 −250.941 119.475 50.494
5 −324.809 438.777 138.458 −276.212 20.179 147.967
6 −336.763 537.993 134.810 −247.634 115.978 150.378
7 −320.996 381.044 200.526 −291.380 −39.398 206.452
8 −347.291 599.319 192.504 −228.510 172.059 211.754

The coordinate measuring machine we used is the Romer model of the Hexagon company. The
measurement accuracy of the coordinate measuring machine is 5μm. The accuracy of the coordinate
measuring machine is much higher than the vision measurement system we used. Therefore, we can
use the coordinate measuring machine to evaluate the accuracy of the visual system in the
experiment. We have evaluated the measurement results of the coordinate measuring machine, and
calculated the residuals of Equation (13) using the measured data. Using the data in Table 4, the
residual of Equation (13) is 0.2365 mm (the sum of the residuals of the 8 corner points). The smaller
residual value indicates that the coordinate measuring machine has higher measurement accuracy
and can be used to verify the accuracy of the proposed algorithm.

Target2
Target 1

Figure 8. The spatial coordinates of the measured points on target 3 and target 4.The red circle
represents the corner point of the targets, and the blue crosses represent the measured points.

A total of 15 experiments were carried out, the experiment results are shown in Table 5, and the
errors between calculated results and the ground truth data measured using a coordinate measuring
machine are shown in Figure 9.
Remote Sens. 2018, 10, 1298 14 of 18

Table 5. The experiment results and the truth value measured by a coordinate measuring machine.

Rotate Translation
Index a (rad) b (rad) c (rad) x (mm) y (mm) z (mm)
1 0.412535 −0.073078 −0.084716 −165.3740 −497.73 −43.4709
2 0.413592 −0.072140 −0.084732 −165.3776 −497.759 −43.5122
3 0.414233 −0.072867 −0.084769 −165.4139 −497.785 −43.4713
4 0.413808 −0.072550 −0.085103 −165.4118 −497.771 −43.4689
5 0.412850 −0.072154 −0.084903 −165.3968 −497.78 −43.5357
6 0.413207 −0.072816 −0.085246 −165.3841 −497.77 −43.5158
7 0.413391 −0.072739 −0.085004 −165.3767 −497.757 −43.5191
8 0.414433 −0.072085 −0.085072 −165.3915 −497.75 −43.4718
9 0.412782 −0.072789 −0.085163 −165.3668 −497.756 −43.4529
10 0.4141816 −0.071671 −0.08515 −165.3728 −497.723 −43.5063
11 0.413759 −0.071354 −0.085006 −165.3578 −497.751 −43.4597
12 0.413222 −0.071859 −0.085203 −165.3841 −497.717 −43.471
13 0.412851 −0.072632 −0.084900 −165.3964 −497.742 −43.5423
14 0.413326 −0.072151 −0.084977 −165.4096 −497.716 −43.4788
15 0.413434 −0.073104 −0.084842 −165.3793 −497.773 −43.5017
truth values 0.41347 −0.07232 −0.08496 −165.391 −497.746 −43.524
std 0.000546 0.000527 0.000172 0.01679 0.02255 0.02860

(a) (b)
Figure 9. Errors between the calculated results and the ground truth data. (a) The figure on the left
shows three rotation angle errors; (b) the figure on the right shows three translation errors.

It can be seen from the above experimental results that the three rotation angle errors are within
0.001rad, the RMS error of the roll angle is 0.0021, the RMS error of the yaw angle is 0.0020, and the
RMS error of the pitch angle is 0.00065. The difference between the mean value and the true value is
−2.323 × 10−5, the difference between the mean value and the true value of the yaw angle is −7.779 ×
10−5, and the difference between the mean value and the true value of the pitch angle is −2.632 × 10−5.
The errors of the three translational values are within 0.08 mm, the RMS value in the x direction is
0.0655, the RMS value in the y direction is 0.0877, and the RMS value in the z direction is 0.1611. The
difference between the average value and the true value of the x-direction displacement is 4.734 ×
10−3. The difference between the mean value and the true value of the y-direction displacement is
−6.176 × 10−3, and the difference between average value and the true value the z-direction
displacement is 3.111 × 10−2.
To illustrate the applicability of the algorithm to all cameras, we present the extrinsic parameters
errors of the remaining cameras relative to the reference coordinate system (camera coordinate
system of the camera 1). The maximum values of the errors in 15 experiments are shown in Table 6.
Remote Sens. 2018, 10, 1298 15 of 18

Table 6. The maximum values of the errors in 15 experiments.

The Maximum Value of the Errors


Camera Number
a(rad) b(rad) c(rad) x(mm) y(mm) z(mm)
Camera 3 0.000921 0.000894 0.000565 0.066 0.059 0.078
Camera 4 0.000945 0.000818 0.000632 0.071 0.062 0.077

From the Table 6, we can see that the optimization results of the remaining cameras are
consistent with the results of camera 2 as shown in Figure 9. The three rotation angle errors of these
cameras are within 0.001rad, and the three translational errors of these cameras are within 0.08 mm.
The accuracy of the extrinsic parameters is improved after optimization. Compared with the two-
camera experiment in Section 3.3, after global optimization, the calibration accuracy has been
obviously improved. The main performance is: the three angles of the extrinsic parameters and the
average of the three translations are all close to the true value, and the RMS error has also been
suppressed to a certain extent.

4. Discussion
In this paper, two steps of the optimization method are used to improve the calibration accuracy.
In the first step, the re-projection error optimization method is used to optimize the extrinsic
parameters of any two cameras in the system. In the second step, the global optimization method is
used to calculate the extrinsic parameters of each camera. The extrinsic parameters are relative to the
reference coordinate system and are optimized by redundant information. From the experimental
results in Section 3, it can be found that after the two optimization processes, the noise has been
effectively suppressed. It should be noted that when the multi-camera system is moved, if the camera
pose changes with a large range, the time of iterative calculation in the re-projection error
optimization process will be reduced. As the number of movements increase, the initial value of the
linear solution is closer to the true value. When the number of the movement is greater than 8, the
initial value is more stable. We use the Euler angle to calculate the extrinsic parameters, which will
result in gimbal lock when the angle changes greatly. The multi-camera system described in this
paper is used for ground observation or visual measurement. The yaw angle of the cameras varies
greatly, while the pitch angle and roll angle of the cameras varies little. Therefore, due to the small
change of roll angle and pitch angle, the gimbal lock phenomenon will not occur in practical
applications.
To demonstrate the advantages of this method more clearly, a comparative analysis is conducted
with the existing method. These are typical calibration methods for cameras without common FOVs
in recent years. Since these papers are all a description of the accuracy of the translation, we use the
translation accuracy as a standard to compare. The results are shown in Table 7.

Table 7. Comparison with previous work.

Algorithm Accuracy RMS Error Operation Economy


3D reconstruction method [19] 0.10 mm No mentioned √ High cost
Reconfigurable target method [20] 0.04 mm 0.0394 Not easy to operate √
Laser projection method [25] 1.2 mm 1.14 Harm to human eyes √
Ours 0.08 mm 0.0286 √ √

It can be seen from Table 7, the accuracy of our method is close to the reconfigurable target
method [20], and better than the other two methods. The reconfigurable target method requires
artificial marks to calculate the pose relationship between sub-targets before each measurement, the
operation process is complicated. The 3D reconstruction method [19] needs large and high-precision
targets, which is difficult to create especially when the angle between the optical axis of the camera
is large. These three methods are only applicable to the laboratory environment. The method in this
paper does not require the aid of lasers or other special targets. It is easy to operate and has high
Remote Sens. 2018, 10, 1298 16 of 18

precision. Our method is not limited by the angle between the cameras’ optical axis. Therefore it is
suitable for on-site calibration in airborne remote sensing.

5. Conclusions
In this article, we propose a flexible multi-camera calibration method without common field of
view to be used in an airborne remote sensing field. The characteristics of this method include: (1)
the target is flexible and does not need a specific mechanical device. It is very suitable for rapid and
on-site calibration in airborne remote sensing. (2) A multi-camera system calibration and
optimization method is proposed. Experiments show that the optimization method based on the
projection error effectively improves the accuracy of the linear solution in binocular calibration. For
multi camera calibration, the global optimization method can effectively improve the effect of a single
calibration error or system error on the final calibration result. (3) The experimental results show that
when the baseline distance between the cameras is approximately 0.5 m, the calibration accuracy of
the rotation angle of the binocular system is within 0.002 rad, and the calibration accuracy of the
translation is within 0.15 mm. For the multi camera system, the accuracy of the three rotation angles
is within 0.001 rad, and the accuracy of the three translations is within 0.08 mm by using the global
optimization. Experiments show that the optimization algorithm proposed in this paper has better
robustness and higher accuracy, this method can guarantee the quality of the tasks such as multi-
camera image mosaic and terrain reconstruction in remote sensing field. The next work is conducted
to reduce the computational complexity of the projection error optimization. In this way, the
calculation speed of this method will be further accelerated.

Author Contributions: Conceptualization, L.Y. and X.W.; Data curation, L.Y. and J.Z.; Funding acquisition,
X.W.; Investigation, L.Y., Y.N., K.Z. and J.Z.; Methodology, L.Y., X.W., Y.N. and K.Z.; Software, L.Y.; Validation,
L.Y.; Writing—original draft, Lei Yin; Writing—review & editing, X.W.

Funding: This research was funded by National Natural Science Foundation (Grant No. 60872097 and No.
61179043), the National High Technology Research and Development Program 863(Grant No.2010AA122200).

Acknowledgments: We would like to thank the anonymous reviewers for their critical, helpful, and constructive
comments and suggestions.

Conflicts of Interest: The authors declare no conflict of interest.

References
1. Colomina, I.; Molina, P. Unmanned aerial systems for photogrammetry and remote sensing: A
review. ISPRS J. Photogramm. Remote Sens. 2014, 92, 79–97.
2. Everaerts, J. The use of unmanned aerial vehicles (UAVs) for remote sensing and mapping. Int.
Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2008, 37, 1187–1192.
3. Rupnik, E.; Nex, F.; Toschi, I.; Remondino, F. Aerial multi-camera systems: Accuracy and block
triangulation issues. ISPRS J. Photogramm. Remote Sens. 2015, 101, 233–246.
4. Turner, D.; Lucieer, A.; Watson, C. An automated technique for generating georectified mosaics
from ultra-high resolution unmanned aerial vehicle (UAV) imagery, based on structure from
motion (SfM) point clouds. Remote Sens. 2012, 4, 1392–1410.
5. Mancini, F.; Dubbini, M.; Gattelli, M.; Stecchi, F.; Fabbri, S.; Gabbianelli, G. Using unmanned
aerial vehicles (UAV) for high-resolution reconstruction of topography: The structure from
motion approach on coastal environments. RemoteSens. 2013, 5, 6880–6898.
6. Harwin, S.; Lucieer, A. Assessing the accuracy of georeferenced point clouds produced via
multi-view stereopsis from unmanned aerial vehicle (UAV) imagery. Remote Sens. 2012, 4, 1573–
1599.
7. d'Oleire-Oltmanns, S.; Marzolff, I.; Peter, K.D.; Ries, J.B. Unmanned aerial vehicle (UAV) for
monitoring soil erosion in Morocco. Remote Sens. 2012, 4, 3390–3416.
8. Dandois, J.P.; Ellis, E.C. Remote sensing of vegetation structure using computer vision. Remote
Sens. 2010, 2, 1157–1176.
Remote Sens. 2018, 10, 1298 17 of 18

9. Lee, J.J.; Shinozuka, M. A vision-based system for remote sensing of bridge displacement. NDT
E Int. 2006, 39, 425–431.
10. Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell.
2000, 22, 1330–1334.
11. Zhang, Z. Camera calibration with one-dimensional objects. IEEE Trans. Pattern Anal. Mach.
Intell. 2004, 26, 892–899.
12. Gurdjos, P.; Crouzil, A;Payrissat, R. Another way of looking at plane-based calibration: The
centre circle constraint. In Proceedings of the European Conference on Computer Vision,
Copenhagen,Denmark, 28–31 May 2002; Springer: Berlin/Heidelberg, Germany, 2002; pp.252–
266.
13. Gurdjos, P.; Sturm, P. Methods and geometry for plane-based self-calibration. In Proceedings of
the IEEE Computer Society Conference onComputer Vision and Pattern Recognition, Madison,
WI, USA, 18–20 June 2003; pp.491–496.
14. Tsai, R.Y. An efficient and accurate camera calibration technique for 3D machine vision. In
Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern
Recognition, Miami Beach, FL, USA, 22–26 June1986; pp. 364–374.
15. Tsai, R. A versatile camera calibration technique for high-accuracy 3D machine vision metrology
using off-the-shelf TV cameras and lenses. IEEE J. Robot. Autom. 1987, 3, 323–344.
16. Agrawal, M.; Davis, L. Complete camera calibration using spheres: A dual space approach. IEEE
Int. Conf. Comput. Vis. 2003, 206, 782–789.
17. Zhang, H.; Zhang, G.; Wong, K.Y. Camera calibration with spheres: Linear approaches. In
Proceedings of the IEEE International Conference on Image Processing 2005 (ICIP 2005).
Genova, Italy, 14 September 2005; IEEE: NY, USA; pp. II-1150.
18. Gong, Z.; Liu, Z.; Zhang, G. Flexible global calibration of multiple cameras with nonoverlapping
fields of view using circular targets. Appl. Opt. 2017, 56, 3122–3131.
19. Dong, S.; Shao, X.; Kang, X.; Yang, F.; He, X. Extrinsic calibration of a non-overlapping camera
network based on close-range photogrammetry. Appl. Opt. 2016, 55, 6363–6370.
20. Xia, R.; Hu, M.; Zhao, J.; Chen, S.; Chen, Y. Global calibration of multi-cameras with non-
overlapping fields of view based on photogrammetry and reconfigurable target. Meas. Sci.
Technol. 2018, 29, 065005, doi:10.1088/1361-6501/aab028.
21. Lu, R.S.; Li, Y.F. A global calibration method for large-scale multi-sensor visual measurement
systems. Sens. Actuators A Phys. 2004, 116, 384–393.
22. Lébraly, P.; Deymier, C.; Ait-Aider, O.; Royer, E.; Dhome, M. Flexible extrinsic calibration of
non-overlapping cameras using a planar mirror: Application to vision-based robotics. In
Proceedings of the IEEE/RSJ International Conference on IEEE Intelligent Robots and Systems
(IROS), Taipei, Taiwan, 18–22 October 2010, pp. 5640–5647.
23. Kumar, R.K.; Ilie, A.; Frahm, J.M.; Pollefeys, M. Simple calibration of non-overlapping cameras
with a mirror. In Proceedings of the CVPR 2008. IEEE Conference on Computer Vision and
Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–7.
24. Pagel, F. Extrinsic self-calibration of multiple cameras with non-overlapping views in vehicles.
In Proceedings of the International Society for Optics and Photonics, Video Surveillance and
Transportation Imaging Applications, San Francisco, CA, USA, 3–5 February 2014; Volume 9026,
pp. 902606.
25. Liu, Z.; Wei, X.; Zhang, G. External parameter calibration of widely distributed vision sensors
with non-overlapping fields of view. Opt. Lasers Eng. 2013, 51, 643–650.
26. Lepetit, V.; Moreno-Noguer, F.; Fua, P. Epnp: An accurate o (n) solution to the PNP problem.
Int. J. Comput. Vis. 2009, 81, 155, doi:10.1007/s11263-008-0152-6.
27. Madsen, K.; Nielsen, H.B.; Tingleff, O. Methods for Non-Linear Least Squares Problems; Informatics
and Mathematical Modelling, Technical University of Denmark, DTU: Lyngby, Denmark, 1999;
pp. 24–29.
Remote Sens. 2018, 10, 1298 18 of 18

28. Martinec, D.; Pajdla, T. Robust rotation and translation estimation in multiview reconstruction.
In Proceedings of the CVPR’07. IEEE Conference on Computer Vision and Pattern Recognition,
Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8.
29. Jiang, N.; Cui, Z.; Tan, P. A global linear method for camera pose registration. In Proceedings of
the 2013 IEEE International Conference on Computer Vision (ICCV), Sydney,Australia, 1–8
December 2013; pp. 481–488.
30. Camera Calibration Toolbox for Matlab. Available online:
http://robots.stanford.edu/cs223b04/JeanYvesCalib/index.html#links (accessed on 5 June 2018).
31. Horn, B.K. Closed-form solution of absolute orientation using unit quaternions. JOSA A 1987, 4,
629–642.

© 2018 by the authors. Submitted for possible open access publication under the terms and
conditions of the Creative Commons Attribution (CC BY) license
(http://creativecommons.org/licenses/by/4.0/).

You might also like