Smart Camera Network Localization Using A 3D Target
Smart Camera Network Localization Using A 3D Target
Camera y (x,y)
designed and created a small target with 288 feature points
(u,v) v
Coordinate
Frame (CCF) x
set across 6 differently angled grids. We have also designed
OC
OI
z and implemented a simple and efficient detection algorithm.
f
y
World
Detection of our target (in a 640x480 image) begins by
stepping to find a green pixel. Then: find all contiguous
Image Coordinate
Coordinate Frame (WCF)
Frame (ICF)
x
green pixels on the row; from the line’s midpoint, find all
z contiguous green pixels on the column; consider this vertical
Figure 1. Projection of 3D points to 2D pixel points line to be the vertical diameter of the sphere atop the target;
Figure 2. Detecting the 3D localization target’s feature points
use the midpoint as a starting reference for finding all grid- Figure 6 shows the change in the estimate of the z-axis
side edges of the colored areas beside each grid. These orientation at each hop. This error is consistently higher than
edges define target-relative horizontal and vertical lines that error in other orientation estimates, and may be due to the
bound a grid and define scan line length and orientation for fact that in our setup the z-coordinate value of target
finding edge fits to all sides of squares in the grid. positions was much greater than the other two.
Intersecting edge fits gives corners of squares, which are the
feature points of the target, shown in Figure 2. Our results 4.3 Message Passing
are generated with the target upright for ground truth Message passing in our solution consists only of determining
measurement purposes, but the detection algorithm does not simultaneous target detections between pairs of cameras, and
require it to be so. We have also verified detection functions
well under various lighting conditions. Figures 3 and 4: Accuracy of single camera localizations
4. EVALUATION
4.1 Single Camera Localization
Our algorithm’s accuracy is dependent upon the accuracy of
the projection matrix estimation by individual cameras upon
observing the target. This is because each camera computes
and passes to at least one neighbor a transformation between
two target coordinate frames it has localized to. Any error in
these localizations is propagated through the network via the
passed transformations.
Figures 3 and 4 demonstrate the accuracy of single camera
localizations using our indoor localization target. Figure 3
shows the position error of a single camera’s localization as
the target is placed successively further from the camera.
Position error is defined as the percentage of the Euclidean
error of the camera position estimation to the camera-to-
target distance. Percentage of frame area occupied by point
matches means the percentage of the area, in square pixels,
of the bounding box around all target feature points to the
total image area. The graph shows that the localization error
when the target is close to the camera is the same at
subsequent target positions–thus the decreasing percentage.
This suggests that the error is likely from an inaccuracy in
manual measurement, because it is consistent. Figure 4
shows that estimated orientation angles fluctuate less than
0.3 of a degree over the same configuration.
4.2 Network Localization
Figure 5 shows the position error of cameras realigned to the
network’s global coordinate frame. Due to the propagation
of single camera errors in transformations passed to
neighbors, the error increases at each hop away from the
camera chosen as origin of the global coordinate frame.
Figures 5 and 6: Accuracy of network localization takes the next step of a full-featured 3D target that not only
resolves scale, but also reduces both message passing and
computation.
7. ACKNOWLEDGEMENTS
This material is based upon work supported by the National
Science Foundation under Grant No. CNS-0722063. Any
opinions, findings, and conclusions or recommendations
expressed in this material are those of the authors and do not
necessarily reflect the views of the National Science
Foundation.
8. REFERENCES
[1] Hartley, R. and Zisserman, A. Multiple View Geometry
in Computer Vision. Cambridge University Press, 2000.
[2] Faugeras, O. The Geometry of Multiple Images. MIT
Press, 2004.
[3] Zhang, Z. A Flexible New Technique for Camera
Calibration. IEEE Transactions on Pattern Analysis
and Machine Intelligence, 22, 11, (Nov 2000),
1330-1334.
[4] Funiak, S., Guestrin, C., Paskin, M., and Sukthankar, R.
Distributed localization of networked cameras.
Information Processing In Sensor Networks, (Apr
2006), 34-42.
[5] Devarajan, D., Radke, R., and Chung, H. Distributed
Metric Calibration of Ad-Hoc Camera Networks. ACM
Transactions on Sensor Networks, 2, 3 (Aug. 2006),
380-403.
[6] Lymberopoulos, D., Barton-Sweeny, A., and Savvides,
A. Sensor localization and camera calibration using low
power cameras. ENALAB Technical Report, 090105,
September 2005.
[7] Mantzel, W., Choi, H., Baraniuk, R.G. Distributed
camera network localization. 38th Asilomar Conference
on Signals, Systems and Computers, (Nov 2004)
[8] Kurillo, G., Li, Z., Bajcsy, R. Wide-area external multi-
the passing of realignment transformations. Because camera calibration using vision graphs and virtual
projection matrix estimation occurs between the target and calibration target. Second ACM/IEEE International
Conference on Distributed Smart Cameras. (Sep 2008)
one camera, there is no need to pass or correlate detected
[9] Medeiros, H., Iwaki, H., Park, J. Online distributed
feature point sets between pairs of cameras. calibration of a large network of wireless cameras using
5. FUTURE WORK dynamic clustering. Second ACM/IEEE International
Conference on Distributed Smart Cameras. (Sep 2008)
Due to the uncertainty manual measurement errors cast over [10] Taylor, C., Rahimi, A., Bachrach, J., Shrobe, H., and
our real deployment results, we are implementing a Grue. A. Simultaneous localization, calibration, and
simulator. We will also implement a centralized bundle tracking in an ad hoc sensor network. Information
adjustment for comparison purposes, as well as the use of Processing In Sensor Networks, (2006), 27-33.
local pairwise bundle adjustments, although both would [11] Rahimi, A., Dunagan, B., Darrell, T. Simultaneous
increase message passing if adopted into the solution. calibration and tracking with a network of non-
overlapping sensors. Computer Vision and Pattern
6. CONCLUSION Recognition, 1, (Jul 2004), 187-194.
We have presented a new solution for smart camera network [12] Lowe, D. Distinctive Image Features from Scale-
localization in 3D that addresses both the point Invariant Keypoints. International Journal of Computer
correspondence problem and high amount of processing Vision, 60, 2 (2004), 91-110.
required in epipolar geometry-based computer-vision [13] Bindel, D., Demmel, J., and Kahan, W. On computing
givens rotations reliably and efficiently. ACM
localization algorithms. Our solution also addresses the Transactions on Mathematical Software, 28, 2, (Jun
unknown scale issue inherent in using epipolar geometry to 2002), 206-238.
determine relative pose between cameras. Recent epipolar [14] Feng, W., Code, B., Kaiser, E., Shea, M., Feng, W.,
geometry-based solutions [8,9] propose the use of a simple Panoptes: scalable low-power video sensor networking
2D calibration target to resolve the scale issue. Our solution technologies. ACM Multimedai (2003), 90-91.