In a multicamera measurement system, the determination of the external parameters is one of the vital tasks, referred to as the calibration of the system. In this paper, a new geometrical calibration method, which is based on the theory of the vanishing line, is proposed. Using a planar target with three equally spaced parallel lines, the normal vector of the target plane can be confirmed easily in every camera coordinate system of the measurement system. By moving the target into more than two different positions, the rotation matrix can be determined from related theory, i.e., the expression of the same vector in different coordinate systems. Moreover, the translation matrix can be derived from the known distance between the adjacent parallel lines. In this paper, the main factors effecting the calibration are analyzed. Simulations show that the proposed method achieves robustness and accuracy. Experimental results show that the calibration can reach 1.25 mm with the range about 0.5m. Furthermore, this calibration method also can be used for autocalibration of the multicamera mefasurement system as the feature of parallels exists widely.
I. INTRODUCTION
In computer vision, multicamera measurement systems have been widely utilized in various areas
[1
,
2]
, e.g., the 3D measurement, object reconstruction, object detecting and tracking, etc. Cameras in the multicamera measurement system are fixed; therefore, the relationship between each pair of cameras is constant. One simple but typical kind of multicamera measurement system is the stereo measurement system, which consists of two cameras with a common field of view. Based on the point pairs, the information about the point in the scene can be obtained from many limitations, such as epipolar constraint, etc. As a welldeveloped theory, the stereo measurement system has been widely utilized in many industrial areas. On the other hand, the cameras in the multicamera measurement system are not always with a common field of view according to different tasks. Therefore, when the cameras in the system have no common field of view, it is called a nonoverlapping multiple vision system. Although the nonoverlapping multiple vision system has no common field of view, the range of the measurement is larger. The most important task is the calibration of cameras in the multicamera measurement system, i.e. obtaining the relationship between each two cameras.
The approaches to obtaining the relationship between cameras, which is called the calibration of the external parameters of the system, are various. These approaches can be classified into two categories according to the target which is used: one uses the exact coordinates of the feature points while the other relies on the geometric or algebraic properties.
In
[3]
, M. Knorr presented an approach to calibrate the extrinsic parameters of a multicamera system using a ground plane, from which the homographies can be deduced. As a natural reference object, the ground plane can be estimated using a single iterated extended Kalman filters. Obviously, the relationship between each pair of cameras can be obtained. In
[4]
, J.P. Barreto et al. estimated the extrinsic parameters of the system by moving an
LED
in thousands of random positions in front of cameras, and then the correspondences between views can be obtained. The projection matrices and radial distortion parameters are simultaneously computed using a factorization approach. The approximate solution of a matrix is projected to the manifold of the parameter space to modify. In
[5]
, with the help of a mirror, a common calibration object can be seen directly through all the nonoverlapping view cameras. By formulating constraints between cameras and the mirrored calibration object, the extrinsic parameters can be deduced simply. For the purpose of overcoming the shortcomings of view angle, a calibration method using a nonplanar target is presented in
[6]
. As the projection of a sphere is not related to its orientation, a target consisting of several spheres is used to finish the task of calibration. As the coordinates of the spheres can be obtained under each coordinate system, the relation is easily confirmed.
In recent years, there are more and more calibration methods which are based on the geometric features, such as the feature of a straight line. In
[7]
, a spot laser projector was utilized to obtain two collinear feature points, which are located in different fields of view. The feature points are the cross points of the target plane and the light beam. More feature points can be obtained by moving the spot laser to different positions. Based on the constraint of the straight line, and the internal relationship between the image plane and the target plane, the calibration of the multicamera system can be carried out easily.
All the calibration methods described above can complete the task, but in the process, the image of the feature point is necessary, so the calibration result will always be affected by the accuracy of the feature point.
In this paper, a new calibration method based on the theory of vanishing point and vanishing line, which are important features in geometry, is proposed. At least three equally spaced parallel lines are required, whether from an artificial target or a natural scene, to obtain the vanishing line of the target. Then the corresponding normal vector can be deduced. When the normal vector in each camera coordinate system is obtained, i.e. the same vector in different coordinate systems, the rotation matrix between the two coordinate systems can be easily derived. As the distance between each pair of adjacent parallel lines is known exactly, the translation matrix can be obtained. In our calibration method, all the points in the view field of each camera can be utilized to calculate the vanishing line, so our proposed method achieves robustness. And also, the feature used in our calibration method is easily obtained, whether from the artificial target or the natural scene. On the other hand, the proposed calibration method can be used to selfcalibrate.
II. PRINCIPLE
As the stereo measurement system can be treated as a special one of the multicamera system, in our approach, we assume two cameras located in two different positions with nonoverlapping view field.
The structure of the measurement system is illustrated in
Fig. 1
.
O
_{l}

X
_{l}
Y
_{l}
Z
_{l}
is the coordinate system of the left camera while
O
_{r}

X
_{r}
Y
_{r}
Z
_{r}
is the right coordinate system of the right camera.
O
_{l}

X
_{l}
Y
_{l}
Z
_{l}
is also treated as the global camera coordinate system and
is the notation of the transformation matrix from
O
_{l}

X
_{l}
Y
_{l}
Z
_{l}
to
O
_{r}

X
_{r}
Y
_{r}
Z
_{r}
.
The structure of the calibration system.
The planar target used in our calibration method is required to be at least three parallel lines with equal space. The planar target should be located in the view field of each camera. When the target is formed into a rectangle, it is easy to satisfy this requirement. The images by each camera can be captured simultaneously.
The main steps of our calibration method are listed as follows:

1) The intrinsic parameters of each camera can be calculated by Zhang’s calibration method[8,9]. The relative position between each two camera is unchangeable while the cameras are fixed in the proper position due to the requirement of the measurement.

2) The planar target with three equally space parallel lines is used to finish the calibration. Put the target in the view field of the two cameras simultaneously (as illustrated inFig. 1).

3) Capture the image of the target from each camera separately, and then obtain the normal vector of the target plane in each camera coordinate system. So the rotation matrix of the two camera coordinate systems can be derived from the different expressions of the same vector, i.e. the normal vector of the target plane.

4) As the distance between each pair of adjacent parallel lines is known exactly, we can get some information about the translation matrix. When the target is moved into several different positions, enough consistency can be confirmed to calculate the translation matrix.
 2.1. Related Theories
A set of parallels in 3D space projects on the image plane of perspective geometry and the intersection is called the vanishing point. The vanishing point depends only on the direction of the line rather than on its position. The vanishing point can be obtained by intersecting the image plane with a ray parallel to the line and passing through the origin of the camera coordinate system. By definition and using the principle of perspective geometry, we have the following property when the direction vector of the line is defined as
where
K
is the intrinsic parameters matrix of the camera and
V
is the vanishing point of the line.
In
[10]
, a new approach of calculating the vanishing line from three coplanar equally spaced parallel lines is proposed. When the homogeneous coordinates of the three parallel lines on the image plane are defined as
l_{0}
,
l_{1}
,
l_{2}
, the vanishing line of the plane can be confirmed from the following equation:
Moreover, we can obtain the relationship between the vanishing line and the normal vector of its corresponding plane:
where
K
is the intrinsic parameters matrix of the camera, and
l
is the vanishing line while
is the normal vector of its corresponding plane.
 2.2. Obtaining the Rotation Matrix
From the knowledge above, the normal vector of the target in the coordinate system of each camera can be derived from Eq. (3). Now the rotation matrix of the two coordinate systems can be obtained from the corresponding normal vectors using the method mentioned in
[11
,
12]
.
 2.3. The Confirmation of the Translation Matrix
The normal vector of the target plane is known as
and the direction vector of the parallel lines is
, which is derived in Section 2.1. Furthermore, the perpendicular vector of the parallel lines on the target plane can be confirmed from the cross of the two vectors:
As illustrated in
Fig. 2
, three random points locating in two parallel lines are point
A
point
B
and point
C
, their corresponding image points are assumed as point
a
point
b
and point
c
.
The relationship between the points in the camera coordinate system.
So the relationship between the image coordinate system and the camera coordinate system is:
where
f
_{0}
is the effective focal length (
EFL
), and [
x
y
1] is the image point of the point [
X Y Z
] under the camera coordinate system.
Obviously, the direction vector from point
A
to point
B
is the direction vector of the parallel lines, i.e.
The projection of the vectors
and
onto the vector
, which is the vertical vector of the parallel lines, is the distance
d
between each pair of adjacent parallel lines. As the vector
is a unit vector, we have the relationship as below:
Combined Eq. (5) with Eq. (6) and Eq. (7), the coordinates of the points
A
,
B
and
C
can be derived. Moreover, the coordinates of the point
A
and the point
B
, which are two random points in the left camera coordinate system, and the coordinates of the point
A
’ and the point
B
’, which are two random points in the right camera coordinate system, can be obtained in the corresponding camera coordinate system.
Point
A
and point
B
can be transformed to the right camera coordinate system according to the constraint of the collineation. So the relation can be expressed as
When the target is moved into more than two different positions, more equations like Eq. (8) can be obtained, from which we can derive the translation matrix simply.
III. SIMULATIONS AND DISCUSSIONS
Simulations will be conducted to evaluate the impacts of some vital factors. The assumed multicamera measurement system consists of two identical cameras, and the intrinsic parameters of each camera, which are listed in
Table 1
, are also identical.
The intrinsic parameters of the camera
The intrinsic parameters of the camera
where
f_{0}
is the effective focal length,
f_{x}
is the scale factor in the
x
coordinate direction,
f_{y}
is the scale factor in the
y
coordinate direction, (
u_{0}
,
v_{0}
)T are the coordinates of the principal point. Now the simulations are detailed below.
 3.1. The Number of Target Positions
In the proposed method, one camera can capture one image of the target when the target is located in one position. And the normal vector of the target plane can be confirmed based on the property of the vanishing line. Theoretically, the rotation matrix can be determined by two vector pairs. But as the noise is inevitable, the normal will involve additive error. If more vector pairs can be used to confirm the rotation matrix, the additive error will be diminished. For the translation matrix, the number of target positions is the main factor, which can be obtained from the confirmation of the translation matrix. Therefore, the number of target positions is the only factor related to the translation matrix in this paper.
For the purpose of verifying the affection of the number of target positions, we conducted a simulation. In the simulation, the distance of each two adjacent lines is 20 mm, and the Gaussian noise with mean 0 and standard deviations of 0.5 pixels is added to the captured images. The number of target positions is varied from 3 to 10. In order to make the illustration clear, we transform the rotation matrix to the rotation vector and then calculate the angle between the ideal vector and the obtained one. The RootMeanSquare (
RMS
) error and the mean absolute error (
MAE
) as the function of the number of target positions are illustrated in
Fig. 3(a)
. For the translation matrix, which is defined as [
T
_{x}
T
_{y}
T
_{z}
]
^{T}
, the MAEs of the three elements are calculated, and the result is illustrated in
Fig. 3(b)
. Each point in
Fig. 3
represents the result averaged over 300 uniformly distributed rotations.
(a) The RMS error and the MAE of the rotation matrix as the function of the number of target positions; (b) The MAE of the translation matrix.
It is obvious that the
RMS
error and the
MAE
decline as the number of the target position increases, which means that the error of the rotation matrix is decreasing. The similar situation still can be found in the determination of the translation matrix. Therefore, if possible, the target should be located in different positions to make the calibration more accurate.
 3.2. The Distance between Two Adjacent Parallel Lines
As it has been known, the normal vector of the target plane is determined by the parallel lines on the target. Nevertheless, when the distance between two adjacent parallel lines is large, the representation of the feature of vanishing is notable. In this case, the effect of the additive noise will be decreased. A simulation will be conducted to verify the conclusion. Similarly, the Gaussian noise with mean 0 and standard deviation of 0.5 pixels is added to the idealized ones to generate the perturbed images. The number of target positions in the simulation is 5. The distance between adjacent parallel lines varies from 5 mm to 50 mm with the interval of 5 mm. The criterion of the simulation is the same as mentioned in Section 3.1, the angle between the idealized rotation vector and the obtained one is used to measure the error involved in the determination of the rotation matrix. The
RMS
error and the
MAE
of the angle as a function of the distance between the two adjacent parallel lines are illustrated in
Fig. 4
.
The RMS error and the MAE of the angle as the function of the distance between the two adjacent parallel lines.
As illustrated in
Fig. 4
, the angle decreases as the distance between the two adjacent parallel lines increases. So it is appropriate to increase the distance if it is possible. Meanwhile, the distortion of the lens cannot be neglected, if possible, the parallel lines should form an image in the center of the image plane, while the increasing of the space may make it more difficult.
 3.3. The Number of Points Used to Determine the Parallel Line
Obviously, the normal vector is one of the vital factors affecting the rotation matrix directly, while the normal vector is determined by the extraction of the parallel lines. In the proposed method, the extraction of the center of the line is based on C.Steger’s method mentioned in
[13]
, which has the capacity of reaching subpixel accuracy. Then the function of the line can be determined by the fitting of the separate points. So the number of image points is another important factor affecting the calibration result. For this purpose, the related simulation has been conducted to evaluate the situation. In this simulation, the number of points used to fit each parallel line varies from 100 to 1000 points with the interval of 100 points. The distance between each pair of adjacent parallel lines is 20 mm, while the number of target positions is 5. The criterion evaluating the calibration results is the same as the description mentioned in Section 3.1, and the angle between the ideal rotation vector and the obtained one is the target. The related result is illustrated in
Fig. 5
.
The RMS error and the MAE as the function of the number of points.
 3.4. The Total Simulation
For the purpose of evaluating the practicability of the proposed method, the simulation will be conducted. Gaussian noise with different standard deviations is added to the ideal image to generate a perturbed one. The distance between the two adjacent parallel lines is 20 mm, while the number of target positions is 5. The rotation matrix is still evaluated by the RMS error of the angle between the ideal rotation vector and the obtained one. The translation matrix is evaluated by the absolute bias of each element. The calibration results are plotted in
Fig. 6
.
The calibration result with different additive noise.
IV. THE REAL EXPERIMENT AND DISCUSSIONS
In the real experiment, the multicamera measurement system consists of two cameras, which are
Mikrotron MC4082’s
with a resolution of 2336×1728 pixels. The intrinsic parameters of each camera, listed in
Table 2
, are obtained by Zhang’s method. The notations
f_{x}
,
f_{y}
,
u_{0}
,
v_{0}
are defined as mentioned in
Table 1
, while
k
_{1}
,
k
_{2}
are the radial distortion coefficients of the lens.
The intrinsic parameters of the two cameras
The intrinsic parameters of the two cameras
The target used in our experiment is a planar target with a series of parallel lines on it (
Fig. 7
), but actually, only three equally space parallel lines are used to determine the vanishing line, and further, the normal vector of the target plane.
The construct of the real experiment.
Then the rotation matrix can be confirmed, while the translation matrix is determined by the steps described in section 2. According to the proposed method, the rotation matrix we obtained is
, while the translation matrix is
.
 4.1. The Experiment for Repeatability
In order to verify the repeatability of the proposed method, we calibrate the multicamera measurement system 8 times. In each time, the same experiment condition is requested. The standard deviation of the rotation vector and the translation vector are calculated, and the result is listed in
Table 3
.
Experimental result of the repeatability
Experimental result of the repeatability
 4.2. The Experiment for Determining the Accuracy
For the purpose of evaluating the accuracy of the proposed method, the method mentioned in Ref.
[7]
is used. Similarly, a 1D target with several feature points at each side is utilized (
Fig. 7
). The feature points on the left side are captured by the left camera while the feature points on the right side are captured by the right camera. As the constraint of the length, the coordinates of the feature points in each camera coordinate system can be determined. Then the feature points can be transformed to the right camera coordinate system according to the calibration result. The distance between the feature point captured by the left camera and the feature point captured by the right camera can be derived, while it is known exactly beforehand. The results are listed in
Table 4
.
The accuracy test for the calibration result
The accuracy test for the calibration result
[
X
_{r}
,
Y
_{r}
,
Z
_{r}
]
^{T}
in the table is the coordinate of the feature point in the right camera coordinate system while the [
X
_{l}
,
Y
_{l}
,
Z
_{l}
]
^{T}
is the coordinate of the feature point which is captured by the left camera but transformed to the right camera coordinate system. The notation
d
m is the measured distance while the notation
d
s is the sign of the standard length.
V. CONCLUSION
In this paper, a new calibration method for the external parameters of the multicamera measurement system is proposed. Using a target with three equally spaced parallel lines, the normal vector of the target plane can be confirmed. By moving the target into at least three different positions, the rotation matrix can be determined. As the distance between two adjacent parallel lines is known exactly, the translation matrix can also be obtained. Then all the external parameters are confirmed. The simulations and real experiments show that the proposed method is accurate and with good robustness. The calibration results can reach about 1.25 mm with the range of 0.5 m.
Moreover, as the feature of parallel lines exists widely, the proposed calibration method can also be used for autocalibration. For instant, the natural feature of parallel lines, such as the edges of the windows or the building, can be utilized to finish the calibration, especially when the target is not proper or for a measurement system with a wide field of view.
View Fulltext
Cho M.
,
Shin D.
2013
“Depth resolution analysis of axially distributed stereo camera systems under fixed constrained resources,”
J. Opt. Soc. Korea
17
500 
505
DOI : 10.3807/JOSK.2013.17.6.500
Hwang J.
,
Park S.
,
Park C.
,
Lee G.
,
Kim K.
2012
“A displaybased visual stimulator for psychophysical and electrophysiological color sensitivity measurements,”
J. Opt. Soc. Korea
16
145 
150
DOI : 10.3807/JOSK.2012.16.2.145
Moritz K.
,
Wolfgang N.
,
Christoph S.
2013
“Online extrinsic multicamera calibration using ground plane induced homographies,”
Proc. IEEE Intelligent Vehicles Symposium
Gold Coast, Australia
236 
241
Barreto J. P.
,
Daniilidis K.
2004
“Wide area multiple camera calibration and estimation of radial distortion,”
Proc. OMNIVIS 2004  Workshop on Omnidirectional Vision and Camera Networks
Prague, CZ
Kumar R. K.
,
Ilie A.
,
Frahm J.
,
Pollefeys M.
2008
“Simple calibration of nonoverlapping cameras with a mirror,”
Proc. Computer Vision and Pattern Recognition (CVPR)
Anchorage, AK
1 
7
Shen E.
,
Hornsey R.
2011
“Multicamera network calibration with a nonplanar target,”
IEEE Sensors Journal
11
2356 
2364
Liu Z.
,
Wei X. G.
,
Zhang G. J.
2013
“External parameter calibration of widely distributed vision sensors with nonoverlapping fields of view,”
Optics and Lasers in Engineering
51
643 
650
DOI : 10.1016/j.optlaseng.2012.11.009
Zhang Z.
2000
“A flexible new technique for camera calibration,”
IEEE Trans. Pattern Anal. Mach. Intell.
22
1330 
1334
DOI : 10.1109/34.888718
Bouguet J.
2013
“Camera calibration toolbox for Matlab,”
http://www.vision.caltech.edu/bouguetj/calib_doc/
Hartley R.
,
Zisserman A.
2003
Multiple View Geometry in Computervision
Cambridge University Press
Cambridge, UK
Chapter 28
Shuster M. D.
,
Oh S. D.
1981
“Threeaxis attitude determination from vector observations,”
J. Guidance and Control
4
70 
77
DOI : 10.2514/3.19717
Horn K. P.
1987
“Closedform solution of absolute orientation using unit quaternions,”
Optical Society of America A
4
629 
642
Steger C.
1998
“Unbiased extraction of curvilinear structures from 2D and 3D image,” Ph. D. Dissertation
Technische Universitaet Muenchen