Extrinsic Parameters Calibration Method of Cameras PDF
Extrinsic Parameters Calibration Method of Cameras PDF
Abstract: Multi-camera systems are widely used in the fields of airborne remote sensing and
unmanned aerial vehicle imaging. The measurement precision of these systems depends on the
accuracy of the extrinsic parameters. Therefore, it is important to accurately calibrate the extrinsic
parameters between the onboard cameras. Unlike conventional multi-camera calibration methods
with a common field of view (FOV), multi-camera calibration without overlapping FOVs has certain
difficulties. In this paper, we propose a calibration method for a multi-camera system without
common FOVs, which is used on aero photogrammetry. First, the extrinsic parameters of any two
cameras in a multi-camera system is calibrated, and the extrinsic matrix is optimized by the re-
projection error. Then, the extrinsic parameters of each camera are unified to the system reference
coordinate system by using the global optimization method. A simulation experiment and a
physical verification experiment are designed for the theoretical arithmetic. The experimental
results show that this method is operable. The rotation error angle of the camera’s extrinsic
parameters is less than 0.001rad and the translation error is less than 0.08 mm.
Keywords: stereo vision; camera calibration; extrinsic parameter; non-overlapping field of view;
airborne remote sensing
1. Introduction
Photogrammetry is widely used in aerial remote sensing [1–3]. More and more airborne remote
sensing platforms are equipped with multi-camera systems to complete tasks such as image mosaic
[4], terrain reconstruction [5,6] and monitoring [7–9]. In some multi-camera remote sensing
photography systems, the angle between the optical axes of the cameras is large enough to obtain a
large field of view (FOV), resulting in the non-overlapping of the FOV, and increasing the calibration
difficulty of extrinsic parameters. However, the quality of terrain reconstruction and panoramic
mosaic tasks depend on the accuracy of the camera extrinsic parameters. To solve this problem, we
introduce a universal multi-camera extrinsic calibration method without common FOVs. The
commonly used method for calibrating the extrinsic parameters is based on a two-dimensional target
proposed by Zhang Z [10]. This method calculates the extrinsic matrix by placing the target in the
public FOV of the cameras. In addition, there are also some calibration methods based on other
targets such as: 1D target [11], plane-based targets [12,13], stereo targets [14,15], spheres target [16,17]
and so on. When calibrating multiple cameras, the above method requires that the different cameras
must have overlapping FOV, and the traditional calibration method is no longer applicable when the
cameras have no common FOV, because the same target cannot be observed by different cameras.
Remote Sens.2018, 10, 1298;doi:10.3390/rs10081298 www.mdpi.com/journal/remotesensing
Remote Sens. 2018, 10, 1298 2 of 18
To solve this problem, Gong et al. [18] and Dong et al. [19] used the three-dimensional
reconstruction method to calculate the extrinsic matrix of cameras without a common FOV. These
authors reconstruct the coding landmarks on the calibration plan by 3D scanner or visual method.
Cameras without a common field of view capture different section of the plane, the non-overlapping
FOVs are connected by the coding marks, and then the extrinsic parameter matrix can be solved. This
method requires reconstruction of the large calibration plane in advance. The reconstruction process
is complex and the calibration accuracy depends on the reconstruction accuracy. Moreover, it is
difficult to ensure the reconstruction accuracy for the large-sized target planes. Xia et al. [20] used
two fixed targets connected by metal rods to calibrate the non-overlapping-field camera. First, two
targets are captured by a large FOV camera, and the pose relationship between the two fixed targets
is computed by the pose recovery method. Then, the transformation matrix between targets is used
to solve the extrinsic parameters between the cameras. This method requires a special target and is
only suitable for binocular camera calibration. When the angle between the optical axes of the two
cameras is large, it is difficult to create the target. Lu et al. [21] used a theodolite pair called a
theodolite coordinate measurement system (TCMS) to calibrate the extrinsic parameters of the
cameras. A calibration block was placed in the camera’s FOV by using the theodolite to measure the
spatial coordinates of the marker points on the target. The transform matrix between the target
coordinate system and the global coordinate system can be obtained, and then the transform matrix
between the camera coordinate system and the global coordinate system can be obtained.
Lébraly et al. [22] introduced a calibration method by using plane mirrors. The key of this
method is to build a virtual camera through the reflection of the flat mirror and to use the virtual
camera to associate real cameras. There are some marks on the plane mirror, and the pose of the
virtual camera relative to the real camera can be calculated by these marks. However, when the
number of cameras is large, it is difficult to control the reflection scope of the flat mirror. Kumar et al.
[23] also calibrated multiple cameras using flat mirrors. The method removes the need for cameras
to observe the calibration object directly. The advantage of this method is that there is no need to
place markers on the mirror and through a novel set of constraints the mirror pose is not needed for
calibration.
Pagel et al. [24] introduced an extrinsic calibration method of multiple cameras with non-
overlapping views in vehicles. In this paper, the calibration problem is divided into the problems of
visual odometry and Hand-Eye Calibration. Visual odometry is calculated by the structure from the
motion method, which is based on the corresponding image features of each camera, and then the
visual odometry is optimized by sparse bundle adjustment (SBA). An Iterated Extended Kalman
Filter (IEKF) algorithm is used to calculate the uncertainties of the transformation, and merge the
redundant information to improve the calibration accuracy. By using the IEKF algorithm the extrinsic
parameters can be continuously estimated and the system is able to perform a self-calibration. Liu et
al. [25] proposed a calibration method for vision sensors with non-overlapping FOVs using a spot
laser. By placing the laser in suitable positions, the laser line passes through the FOVs of multiple
vision sensors. Utilizing the plane target to obtain the spatial coordinates of the intersection point of
the laser line and the target in each camera coordinate system, the equation is established by using
collinear conditions of the intersection points. By moving the laser line at least two positions,
sufficient constraint equations can be established to solve the extrinsic parameter matrix.
In this paper, we propose a calibration method of multi-cameras without a common field of view
based on multiple chessboard targets. The difference from this method and the traditional non-
common-field calibration method is that the target does not need to be rigidly connected through a
mechanical structure or with the aid of other auxiliary equipment, and the configuration is more
flexible. This method has high accuracy and is suitable for the on-site calibration of the remote sensing
imaging system. The number of chessboard targets is consistent with the number of cameras, so that
each camera corresponds to a target. We keep the target stationary during the calibration process,
and make the multi-camera system move multiple positions and shoot their corresponding targets.
Sufficient equations can be obtained to solve the extrinsic parameters by moving at least two
locations.
Remote Sens. 2018, 10, 1298 3 of 18
Rm,tm
Target1 Target2
Camera5 Camera3
Target5 Camera4
Target3
Target4
As shown in Figure 1, the FOVs of the five cameras do not overlap with each other, and the 5
cameras correspond to 5 checkerboard targets respectively. We first select any two cameras (such as
camera1 and camera2) to introduce the calibration method between two cameras in the system. In
the figure, Rc and tc are the rotation matrix and translation matrix between the two cameras, and
Rm and tm are the rotation and translation matrix between the targets. RTC1(1 i ) , tTC1(1 i ) , RCT 2(
2 T2
i ) , tC 2(i ) are
the rotation and translation matrix between the cameras and their corresponding target at the ith
position of the multi-camera system. PTT22 is the coordinates of the corner points of target 2 in the
target 2 coordinate system; PTT12 is the coordinates of the corner points of target 2 in the target 1
coordinate system. It can be seen from the figure that the relationship between PTT22 and PTT12 can be
transformed by two different paths.
Remote Sens. 2018, 10, 1298 4 of 18
R tm T 2
PTT12 = m PT 2
1
(1)
0
A set of equations for the unknown Rc , t c , Rm , tm can be obtained from Equations (1) and(2):
R m , tm P2
P1
Camera1 Camera2
Rc , tc
The coordinate of the jth corner point on target 1 is transformed to the camera 1 coordinate
system by path 2, and we can obtain the re-projected image coordinate Pcm1,1 j of the corner point from
Remote Sens. 2018, 10, 1298 5 of 18
the intrinsic parameter matrix of camera 1. The coordinate of the corner point on target 2 is
transformed to the camera 2 coordinate system by path 1, and we can obtain the re-projected image
coordinate Pcm2,2 j of the corner point from the intrinsic parameter matrix of camera 2. According to
the camera model, we have:
−1
R tc RCT 22 tCT 22 Rm tm
Pcm1 1 = A1 I 0 c
1 0 1
P1,j (4)
0 1 0
−1 −1
R tc RTC11 tTC11 Rm tm
Pcm2 2 = A2 I 0 c (5)
1 0 1
P2,j
0 1 0
(
f ( Rc,Tc ) = min( j =1 d Pcm1,1 j , Pˆcm1,1 j ) + ( )
2 2
d Pcm2,2 j , Pˆcm2,2 j )
n n
j =1
(6)
where A1 and A2 denote the intrinsic parameters matrix of the two cameras, P1,j and P2,j denote the
coordinates of the jth corner point on the two targets in the target coordinate system, I denotes a 3 ×
3 unit matrix. Pˆ m1 and Pˆ m2 denote the image coordinates, which remove the lens distortion. By
c1,j c 2, j
using Levenberg–Marquardt method [27] to solve Equation (6), the optimized extrinsic parameter
matrix can be obtained.
where r1i , r2i , r3i denote columns of Ri ,where Ri = r1i r2i r3i , Equation (8) can be arranged into
a form of linear equations:
r1i
i
r2 0
ri 31
The rotation matrices between any two cameras can be organized into the equation as shown in
Equation (9). Taking a 5-camera system as an example, C (5, 2) =10 equations as shown in Equation
(9) can be obtained. The 10 sets of equations can be combined and written into an over-determined
linear equation like Ax = 0 . Where A is a matrix of 30 × 15 dimensions that is made by stacking all
relative rotation matrices Rij and x is made by stacking all columns of camera extrinsic parameters Ri .
By solving the linear equation, the rotation matrix of each camera relative to the reference coordinate
system can be obtained. The orthonormality constraints are forced by projecting the approximate
rotation to the closest rotation in the Frobenius norm using SVD. In this case, the global rotation
parameters of each camera are obtained by 10 sets of relative rotation, which can effectively reduce
the influence of single measurement noise on global calibration accuracy because of the use of the
redundancy rotation matrix.
ci − c j = R j −1tij (11)
ci1
2
ci t
c3 1
where ci = ci1 ci2 ci3 , c j = c1j c 2j c 3j , R j −1tij = t1 t2 t3 . The translation vector between any two
T T T
cameras can be organized into the equation as shown in Equation (12). Taking a 5-camera system as
an example, C (5, 2) =10 equations as shown in Equation (12) can be obtained. The 10 sets of equations
can be combined and written into an over-determined linear equation like Ax = b . Where A is a
matrix of 30 × 15 dimensions that is made by stacking identity matrix, and x is made by stacking
Remote Sens. 2018, 10, 1298 7 of 18
all camera translation parameters ci . b is a vector of 30 × 1 dimensions that is made by the elements
in vector R j −1tij . By solving the linear equation, we can obtain the translation vectors ci , (i = 1, 2 5)
of each camera relative to the reference coordinate system. In this case the influence of single
measurement noise can effectively be reduced by using the redundant translation vector.
After obtaining the matrix Ri and vector ci through the above two subsections, we can obtain
Ri ci
the extrinsic matrix of each camera relative to the reference coordinate system
1
.
0
3. Results
(a) (b)
Remote Sens. 2018, 10, 1298 8 of 18
(c) (d)
Figure 3. The RMS error curve of the calibration result affected by noise. The graph (a) shows the
three rotation angle error curves obtained with the linear solution method with ten noise levels. The
graph (b) shows the translation error curve obtained with the linear solution method with ten noise
levels. The graph (c) is the rotation angle error curve after optimization of the re-projection error. The
graph (d) is the translation error curve after optimization of the re-projection error.
3.2. Simulation Experiment of Multi-Camera Extrinsic Parameters Calibration with the Global Optimization
Method
After calculating the extrinsic parameter matrix between any two cameras, the global
optimization method is used to calculate the extrinsic parameter matrices of all cameras relative to
the reference coordinate system. We analyzed the error variation of the extrinsic parameters before
and after global optimization.
The simulation experiment takes a five-camera system as an example. Using the method in
Section 2.1 to calculate the extrinsic matrix between any two cameras, and we can obtain 10 sets of
extrinsic parameters. Ten levels of noise are added to the 10 sets of extrinsic parameters. The noise
added to the rotation angles range from 0.001rad to 0.01rad, and the translation noise ranges from
0.05 mm to 0.5 mm. For each noise level, 100 tests are carried out. The extrinsic parameters of the five
cameras relative to the reference coordinate system are shown in Table 1. Without loss of generality,
we take camera 3 as an example, before and after optimization, the changes in the RMS errors of the
extrinsic parameters are shown in Figure 4.
Table1. The extrinsic of each camera relative to the reference coordinate system.
Extrinsic Parameters
Camera Roll Yew Pitching X-Axis Y-Axis Y-Axis
Number Angle Angle Angle Translation Translation Translation
(rad) (rad) (rad) (mm) (mm) (mm)
1 0 0 0 0 0 0
2 0.698 0.175 0.157 500 −100 10
3 1.484 0.262 0.367 600 300 −50
4 2.426 0.559 0.716 −10 600 20
5 −1.047 −0.070 0.401 −590 400 20
Remote Sens. 2018, 10, 1298 9 of 18
(a) (b)
(c) (d)
Figure 4. The extrinsic parameters’ RMS error curve of camera 3 before and after global optimization.
The graph (a) shows the three rotation angle error curves obtained before optimization with ten noise
levels. The graph (b) shows the translation error curve obtained before optimization with ten noise
levels. The graph (c) is the rotation angle error curve after optimization with ten noise levels. The
graph (d) is the translation error curve after optimization with ten noise levels.
The experimental results show that the global optimization method can effectively suppress the
random noise of single calibration. Because the noise of this simulation experiment is directly added
to three Euler angles and three translations of the extrinsic parameters, so the error has a linear
relationship with the noise level before optimization. It can be seen from the graph that the rotation
angle error is obviously suppressed after optimization. From the figure we see within a certain noise
level, the relationship between the rotation angle RMS error and noise level is linear. From Section
2.2.2 we can see that since the global optimization of the translation vector depends on the rotation
angle, the noise of the rotation angle has a great influence on the translation vector. The experimental
results show that when the noise level is higher and the rotation angle error increases, the error of
the translation vector increases more intense.
To illustrate the applicability of the algorithm to all cameras, we give the extrinsic parameters
errors of the remaining cameras relative to the reference coordinate system (camera coordinate
system of the camera 1) at different noise levels (level 2 and level 5 which are consistent with Figure
4). The errors before and after optimization are shown in Table 2.
From the Table 2, we can see that the optimization results of the remaining cameras are
consistent with the results of the camera 3 as described in Figure 4. The calibration accuracy of all the
cameras have improved after global optimization.
Remote Sens. 2018, 10, 1298 10 of 18
Table 2. The errors before and after optimization of the remaining cameras.
Camera 2 Camera 4 Camera 5
Noise Extrinsic
Before After Before After Before After
Level Parameters
Optimization Optimization Optimization Optimization Optimization Optimization
a (rad) 0.00194 0.00135 0.00208 0.00165 0.00183 0.00122
b (rad) 0.00192 0.00114 0.00210 0.00194 0.00174 0.00116
Level c (rad) 0.00203 0.00156 0.00199 0.00118 0.00207 0.00095
2 x (mm) 0.103 0.064 0.121 0.071 0.104 0.063
y (mm) 0.089 0.077 0.009 0.069 0.009 0.054
z (mm) 0.101 0.068 0.110 0.078 0.100 0.077
a (rad) 0.00506 0.00398 0.00400 0.00334 0.00512 0.00265
b (rad) 0.00513 0.00351 0.00523 0.00416 0.00501 0.00296
Level c (rad) 0.00498 0.00324 0.00491 0.00359 0.00523 0.00312
5 x (mm) 0.251 0.172 0.254 0.192 0.249 0.156
y (mm) 0.262 0.150 0.251 0.209 0.250 0.165
z (mm) 0.249 0.149 0.236 0.186 0.251 0.227
The experimental setup is shown in Figure 5, we keep the target position stationary and the
camera is moved 10 positions, and then the extrinsic matrix between the two cameras can be solved.
Remote Sens. 2018, 10, 1298 11 of 18
(a) (b)
Figure 6. The error curves of the calibration results for the two-camera experiment. (a) the three
rotation angle error curves; (b) the translation error curves.
The experimental results show that the error of the three rotation angles is within 0.002 rad,
where the RMS error of the roll angle is 0.0041, the RMS error of the yaw angle is 0.0039, and the RMS
error of the pitch angle is 0.0056. The difference between the mean value and true value of the roll
angle is 9.023 × 10−4, the difference between the mean value and the true value of the yaw angle is
9.417 × 10−4, and the difference between the mean value and the true value of the pitch angle is 0.0014.
The error of the three translational is within 0.08 mm, where the RMS error in the x-direction
displacement is 0.1816, the RMS error in the y-direction displacement is 0.1102, and the RMS error in
the z-direction is 0.2206. The difference between the mean value and the true value in the x direction
is 0.0424, the difference between the mean value and the true value in the y direction is 0.0073, and
the difference between the mean value and the true value in the z direction is 0.0531.
optimization. To verify the accuracy of the calibration algorithm under real conditions, cameras in
this experiment have no overlapping FOVs. Since the extrinsic parameters’ true value of cameras
without a common field of view cannot be directly calculated, we verify the accuracy of extrinsic
parameters by indirect methods with a coordinate measuring machine. The experimental device is
shown in Figure 7.
Camera 2
Target
Camera 4
Camera 1
Camera 3
Figure 7. The left figure shows the experimental setup, and the right figure shows the measurement
points marked by the red box on the target.
In this experiment, 4 checkerboard targets are used to calibrate 4 cameras. First, using the
method described in Section 2.1 to calculate the extrinsic parameter matrix between any two cameras,
a total of 6 combinations can be obtained. Then, using the global optimization method to calculate
the extrinsic parameters of each camera relative to the reference coordinate system through 6 sets of
extrinsic parameter data. Since the position of the camera relative to the corresponding checkerboard
can be calculated by EPNP algorithm [26], we can use the extrinsic parameters between the
checkerboards to obtain the extrinsic parameters between the corresponding cameras, and we used
it as ground truth.
Without loss of generality, we use the extrinsic parameters of camera 1 and camera 2 as an
example to verify the algorithm. The extrinsic parameters between the corresponding checkerboard
targets are obtained using a coordinate measuring machine. The measurement method of the
extrinsic parameters between the checkerboard targets is as follows:
(1) Select 8 corner points on target 1, and measure the coordinates of the 8 points using the
coordinate measuring machine. The position of the eight points on target 1 is shown in Figure 7.
(2) Measure the 8 points on target 2 at the same position as target 1.
(3) Establish constraint equations with Absolute Directional Algorithm [31], then we can obtain
the extrinsic parameters between the two targets using the coordinates of the 8 points. The equation
is shown in Equation (13):
n
min Rpi + t − qi , subject to RT R = I
2
R ,t
(13)
i =1
where pi denotes the coordinates of corner point on target 1 and qi denotes the coordinates of corner
point on target 2.
The spatial coordinates of the measured points on target 1 and target 2 in the coordinate system
of the three coordinates are shown in Figure 8 and the coordinate values of the 8 points are shown in
Table 4.
Remote Sens. 2018, 10, 1298 13 of 18
The coordinate measuring machine we used is the Romer model of the Hexagon company. The
measurement accuracy of the coordinate measuring machine is 5μm. The accuracy of the coordinate
measuring machine is much higher than the vision measurement system we used. Therefore, we can
use the coordinate measuring machine to evaluate the accuracy of the visual system in the
experiment. We have evaluated the measurement results of the coordinate measuring machine, and
calculated the residuals of Equation (13) using the measured data. Using the data in Table 4, the
residual of Equation (13) is 0.2365 mm (the sum of the residuals of the 8 corner points). The smaller
residual value indicates that the coordinate measuring machine has higher measurement accuracy
and can be used to verify the accuracy of the proposed algorithm.
Target2
Target 1
Figure 8. The spatial coordinates of the measured points on target 3 and target 4.The red circle
represents the corner point of the targets, and the blue crosses represent the measured points.
A total of 15 experiments were carried out, the experiment results are shown in Table 5, and the
errors between calculated results and the ground truth data measured using a coordinate measuring
machine are shown in Figure 9.
Remote Sens. 2018, 10, 1298 14 of 18
Table 5. The experiment results and the truth value measured by a coordinate measuring machine.
Rotate Translation
Index a (rad) b (rad) c (rad) x (mm) y (mm) z (mm)
1 0.412535 −0.073078 −0.084716 −165.3740 −497.73 −43.4709
2 0.413592 −0.072140 −0.084732 −165.3776 −497.759 −43.5122
3 0.414233 −0.072867 −0.084769 −165.4139 −497.785 −43.4713
4 0.413808 −0.072550 −0.085103 −165.4118 −497.771 −43.4689
5 0.412850 −0.072154 −0.084903 −165.3968 −497.78 −43.5357
6 0.413207 −0.072816 −0.085246 −165.3841 −497.77 −43.5158
7 0.413391 −0.072739 −0.085004 −165.3767 −497.757 −43.5191
8 0.414433 −0.072085 −0.085072 −165.3915 −497.75 −43.4718
9 0.412782 −0.072789 −0.085163 −165.3668 −497.756 −43.4529
10 0.4141816 −0.071671 −0.08515 −165.3728 −497.723 −43.5063
11 0.413759 −0.071354 −0.085006 −165.3578 −497.751 −43.4597
12 0.413222 −0.071859 −0.085203 −165.3841 −497.717 −43.471
13 0.412851 −0.072632 −0.084900 −165.3964 −497.742 −43.5423
14 0.413326 −0.072151 −0.084977 −165.4096 −497.716 −43.4788
15 0.413434 −0.073104 −0.084842 −165.3793 −497.773 −43.5017
truth values 0.41347 −0.07232 −0.08496 −165.391 −497.746 −43.524
std 0.000546 0.000527 0.000172 0.01679 0.02255 0.02860
(a) (b)
Figure 9. Errors between the calculated results and the ground truth data. (a) The figure on the left
shows three rotation angle errors; (b) the figure on the right shows three translation errors.
It can be seen from the above experimental results that the three rotation angle errors are within
0.001rad, the RMS error of the roll angle is 0.0021, the RMS error of the yaw angle is 0.0020, and the
RMS error of the pitch angle is 0.00065. The difference between the mean value and the true value is
−2.323 × 10−5, the difference between the mean value and the true value of the yaw angle is −7.779 ×
10−5, and the difference between the mean value and the true value of the pitch angle is −2.632 × 10−5.
The errors of the three translational values are within 0.08 mm, the RMS value in the x direction is
0.0655, the RMS value in the y direction is 0.0877, and the RMS value in the z direction is 0.1611. The
difference between the average value and the true value of the x-direction displacement is 4.734 ×
10−3. The difference between the mean value and the true value of the y-direction displacement is
−6.176 × 10−3, and the difference between average value and the true value the z-direction
displacement is 3.111 × 10−2.
To illustrate the applicability of the algorithm to all cameras, we present the extrinsic parameters
errors of the remaining cameras relative to the reference coordinate system (camera coordinate
system of the camera 1). The maximum values of the errors in 15 experiments are shown in Table 6.
Remote Sens. 2018, 10, 1298 15 of 18
From the Table 6, we can see that the optimization results of the remaining cameras are
consistent with the results of camera 2 as shown in Figure 9. The three rotation angle errors of these
cameras are within 0.001rad, and the three translational errors of these cameras are within 0.08 mm.
The accuracy of the extrinsic parameters is improved after optimization. Compared with the two-
camera experiment in Section 3.3, after global optimization, the calibration accuracy has been
obviously improved. The main performance is: the three angles of the extrinsic parameters and the
average of the three translations are all close to the true value, and the RMS error has also been
suppressed to a certain extent.
4. Discussion
In this paper, two steps of the optimization method are used to improve the calibration accuracy.
In the first step, the re-projection error optimization method is used to optimize the extrinsic
parameters of any two cameras in the system. In the second step, the global optimization method is
used to calculate the extrinsic parameters of each camera. The extrinsic parameters are relative to the
reference coordinate system and are optimized by redundant information. From the experimental
results in Section 3, it can be found that after the two optimization processes, the noise has been
effectively suppressed. It should be noted that when the multi-camera system is moved, if the camera
pose changes with a large range, the time of iterative calculation in the re-projection error
optimization process will be reduced. As the number of movements increase, the initial value of the
linear solution is closer to the true value. When the number of the movement is greater than 8, the
initial value is more stable. We use the Euler angle to calculate the extrinsic parameters, which will
result in gimbal lock when the angle changes greatly. The multi-camera system described in this
paper is used for ground observation or visual measurement. The yaw angle of the cameras varies
greatly, while the pitch angle and roll angle of the cameras varies little. Therefore, due to the small
change of roll angle and pitch angle, the gimbal lock phenomenon will not occur in practical
applications.
To demonstrate the advantages of this method more clearly, a comparative analysis is conducted
with the existing method. These are typical calibration methods for cameras without common FOVs
in recent years. Since these papers are all a description of the accuracy of the translation, we use the
translation accuracy as a standard to compare. The results are shown in Table 7.
It can be seen from Table 7, the accuracy of our method is close to the reconfigurable target
method [20], and better than the other two methods. The reconfigurable target method requires
artificial marks to calculate the pose relationship between sub-targets before each measurement, the
operation process is complicated. The 3D reconstruction method [19] needs large and high-precision
targets, which is difficult to create especially when the angle between the optical axis of the camera
is large. These three methods are only applicable to the laboratory environment. The method in this
paper does not require the aid of lasers or other special targets. It is easy to operate and has high
Remote Sens. 2018, 10, 1298 16 of 18
precision. Our method is not limited by the angle between the cameras’ optical axis. Therefore it is
suitable for on-site calibration in airborne remote sensing.
5. Conclusions
In this article, we propose a flexible multi-camera calibration method without common field of
view to be used in an airborne remote sensing field. The characteristics of this method include: (1)
the target is flexible and does not need a specific mechanical device. It is very suitable for rapid and
on-site calibration in airborne remote sensing. (2) A multi-camera system calibration and
optimization method is proposed. Experiments show that the optimization method based on the
projection error effectively improves the accuracy of the linear solution in binocular calibration. For
multi camera calibration, the global optimization method can effectively improve the effect of a single
calibration error or system error on the final calibration result. (3) The experimental results show that
when the baseline distance between the cameras is approximately 0.5 m, the calibration accuracy of
the rotation angle of the binocular system is within 0.002 rad, and the calibration accuracy of the
translation is within 0.15 mm. For the multi camera system, the accuracy of the three rotation angles
is within 0.001 rad, and the accuracy of the three translations is within 0.08 mm by using the global
optimization. Experiments show that the optimization algorithm proposed in this paper has better
robustness and higher accuracy, this method can guarantee the quality of the tasks such as multi-
camera image mosaic and terrain reconstruction in remote sensing field. The next work is conducted
to reduce the computational complexity of the projection error optimization. In this way, the
calculation speed of this method will be further accelerated.
Author Contributions: Conceptualization, L.Y. and X.W.; Data curation, L.Y. and J.Z.; Funding acquisition,
X.W.; Investigation, L.Y., Y.N., K.Z. and J.Z.; Methodology, L.Y., X.W., Y.N. and K.Z.; Software, L.Y.; Validation,
L.Y.; Writing—original draft, Lei Yin; Writing—review & editing, X.W.
Funding: This research was funded by National Natural Science Foundation (Grant No. 60872097 and No.
61179043), the National High Technology Research and Development Program 863(Grant No.2010AA122200).
Acknowledgments: We would like to thank the anonymous reviewers for their critical, helpful, and constructive
comments and suggestions.
References
1. Colomina, I.; Molina, P. Unmanned aerial systems for photogrammetry and remote sensing: A
review. ISPRS J. Photogramm. Remote Sens. 2014, 92, 79–97.
2. Everaerts, J. The use of unmanned aerial vehicles (UAVs) for remote sensing and mapping. Int.
Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2008, 37, 1187–1192.
3. Rupnik, E.; Nex, F.; Toschi, I.; Remondino, F. Aerial multi-camera systems: Accuracy and block
triangulation issues. ISPRS J. Photogramm. Remote Sens. 2015, 101, 233–246.
4. Turner, D.; Lucieer, A.; Watson, C. An automated technique for generating georectified mosaics
from ultra-high resolution unmanned aerial vehicle (UAV) imagery, based on structure from
motion (SfM) point clouds. Remote Sens. 2012, 4, 1392–1410.
5. Mancini, F.; Dubbini, M.; Gattelli, M.; Stecchi, F.; Fabbri, S.; Gabbianelli, G. Using unmanned
aerial vehicles (UAV) for high-resolution reconstruction of topography: The structure from
motion approach on coastal environments. RemoteSens. 2013, 5, 6880–6898.
6. Harwin, S.; Lucieer, A. Assessing the accuracy of georeferenced point clouds produced via
multi-view stereopsis from unmanned aerial vehicle (UAV) imagery. Remote Sens. 2012, 4, 1573–
1599.
7. d'Oleire-Oltmanns, S.; Marzolff, I.; Peter, K.D.; Ries, J.B. Unmanned aerial vehicle (UAV) for
monitoring soil erosion in Morocco. Remote Sens. 2012, 4, 3390–3416.
8. Dandois, J.P.; Ellis, E.C. Remote sensing of vegetation structure using computer vision. Remote
Sens. 2010, 2, 1157–1176.
Remote Sens. 2018, 10, 1298 17 of 18
9. Lee, J.J.; Shinozuka, M. A vision-based system for remote sensing of bridge displacement. NDT
E Int. 2006, 39, 425–431.
10. Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell.
2000, 22, 1330–1334.
11. Zhang, Z. Camera calibration with one-dimensional objects. IEEE Trans. Pattern Anal. Mach.
Intell. 2004, 26, 892–899.
12. Gurdjos, P.; Crouzil, A;Payrissat, R. Another way of looking at plane-based calibration: The
centre circle constraint. In Proceedings of the European Conference on Computer Vision,
Copenhagen,Denmark, 28–31 May 2002; Springer: Berlin/Heidelberg, Germany, 2002; pp.252–
266.
13. Gurdjos, P.; Sturm, P. Methods and geometry for plane-based self-calibration. In Proceedings of
the IEEE Computer Society Conference onComputer Vision and Pattern Recognition, Madison,
WI, USA, 18–20 June 2003; pp.491–496.
14. Tsai, R.Y. An efficient and accurate camera calibration technique for 3D machine vision. In
Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern
Recognition, Miami Beach, FL, USA, 22–26 June1986; pp. 364–374.
15. Tsai, R. A versatile camera calibration technique for high-accuracy 3D machine vision metrology
using off-the-shelf TV cameras and lenses. IEEE J. Robot. Autom. 1987, 3, 323–344.
16. Agrawal, M.; Davis, L. Complete camera calibration using spheres: A dual space approach. IEEE
Int. Conf. Comput. Vis. 2003, 206, 782–789.
17. Zhang, H.; Zhang, G.; Wong, K.Y. Camera calibration with spheres: Linear approaches. In
Proceedings of the IEEE International Conference on Image Processing 2005 (ICIP 2005).
Genova, Italy, 14 September 2005; IEEE: NY, USA; pp. II-1150.
18. Gong, Z.; Liu, Z.; Zhang, G. Flexible global calibration of multiple cameras with nonoverlapping
fields of view using circular targets. Appl. Opt. 2017, 56, 3122–3131.
19. Dong, S.; Shao, X.; Kang, X.; Yang, F.; He, X. Extrinsic calibration of a non-overlapping camera
network based on close-range photogrammetry. Appl. Opt. 2016, 55, 6363–6370.
20. Xia, R.; Hu, M.; Zhao, J.; Chen, S.; Chen, Y. Global calibration of multi-cameras with non-
overlapping fields of view based on photogrammetry and reconfigurable target. Meas. Sci.
Technol. 2018, 29, 065005, doi:10.1088/1361-6501/aab028.
21. Lu, R.S.; Li, Y.F. A global calibration method for large-scale multi-sensor visual measurement
systems. Sens. Actuators A Phys. 2004, 116, 384–393.
22. Lébraly, P.; Deymier, C.; Ait-Aider, O.; Royer, E.; Dhome, M. Flexible extrinsic calibration of
non-overlapping cameras using a planar mirror: Application to vision-based robotics. In
Proceedings of the IEEE/RSJ International Conference on IEEE Intelligent Robots and Systems
(IROS), Taipei, Taiwan, 18–22 October 2010, pp. 5640–5647.
23. Kumar, R.K.; Ilie, A.; Frahm, J.M.; Pollefeys, M. Simple calibration of non-overlapping cameras
with a mirror. In Proceedings of the CVPR 2008. IEEE Conference on Computer Vision and
Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–7.
24. Pagel, F. Extrinsic self-calibration of multiple cameras with non-overlapping views in vehicles.
In Proceedings of the International Society for Optics and Photonics, Video Surveillance and
Transportation Imaging Applications, San Francisco, CA, USA, 3–5 February 2014; Volume 9026,
pp. 902606.
25. Liu, Z.; Wei, X.; Zhang, G. External parameter calibration of widely distributed vision sensors
with non-overlapping fields of view. Opt. Lasers Eng. 2013, 51, 643–650.
26. Lepetit, V.; Moreno-Noguer, F.; Fua, P. Epnp: An accurate o (n) solution to the PNP problem.
Int. J. Comput. Vis. 2009, 81, 155, doi:10.1007/s11263-008-0152-6.
27. Madsen, K.; Nielsen, H.B.; Tingleff, O. Methods for Non-Linear Least Squares Problems; Informatics
and Mathematical Modelling, Technical University of Denmark, DTU: Lyngby, Denmark, 1999;
pp. 24–29.
Remote Sens. 2018, 10, 1298 18 of 18
28. Martinec, D.; Pajdla, T. Robust rotation and translation estimation in multiview reconstruction.
In Proceedings of the CVPR’07. IEEE Conference on Computer Vision and Pattern Recognition,
Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8.
29. Jiang, N.; Cui, Z.; Tan, P. A global linear method for camera pose registration. In Proceedings of
the 2013 IEEE International Conference on Computer Vision (ICCV), Sydney,Australia, 1–8
December 2013; pp. 481–488.
30. Camera Calibration Toolbox for Matlab. Available online:
http://robots.stanford.edu/cs223b04/JeanYvesCalib/index.html#links (accessed on 5 June 2018).
31. Horn, B.K. Closed-form solution of absolute orientation using unit quaternions. JOSA A 1987, 4,
629–642.
© 2018 by the authors. Submitted for possible open access publication under the terms and
conditions of the Creative Commons Attribution (CC BY) license
(http://creativecommons.org/licenses/by/4.0/).