View PDF - CiteSeerX

True Single View Point Cone Mirror Omni-Directional Catadioptric System1
Shih-Schön Lin, Ruzena Bajcsy
GRASP Laboratory, Computer and Information Science Department
University of Pennsylvania,
[email protected] , [email protected] (also, [email protected])
Abstract
Pinhole camera model is a simplified subset of geometric
optics. In special cases like the image formation of the
cone (a degenerate conic section) mirror in an omnidirectional view catadioptric system, there are more
complex optical phenomena involved that the simple
pinhole model can not explain. We show that using the
full geometric optics model a true single viewpoint cone
mirror omni-directional system can be built. We show
how such system is built first, and then show in detail
how each optical phenomenon works together to make
the system true single viewpoint. The new system
requires only simple off-the-shelf components and still
outperforms other single viewpoint omni-systems for
many applications.
1. Introduction
With the ideal pinhole camera model in mind, it has
been shown that conic section convex mirrors can be
used to create single viewpoint catadioptric omnidirectional view systems that can be unwarped
perspectively without systematic distortions [1]. It has
also been shown that cone mirror is a degenerate shape of
the conic section family that still possesses a single
viewpoint [1]. The problem was that, with a true pinhole
camera the single viewpoint of a cone (located at its tip)
seems not useable because all the visible rays seem to be
blocked by the cone itself [1], while if the virtual pinhole
was placed at some distance away from the tip of the
cone, the whole system no longer possesses a single
viewpoint [10;1;8].
With this problem in mind, cone mirrors have not
been used to generate precisely unwarped perspective
images. However, cone mirrors have been used to aid
navigation, collision avoidance, and pipe inspections
without single viewpoint [9;10;2;6;8] In those
1
applications, cone mirror images are used as is, and no
attempt was made to perform a perspectively correct
image unwarping. A cone mirror is also used to construct
mosaics [5], but that requires a special line scan camera,
time consuming scanning, and mechanical moving parts.
In real world applications we use finite aperture lens
systems to simulate the projective properties of a
theoretical pinhole camera. The real cameras have more
useful properties that a pure pinhole lacks. We show in
this work that under the geometric optics image
formation model, the single viewpoint of the cone mirror
is actually usable, i.e. we can physically place the
effective pinhole position at the single viewpoint and still
get the exact true single viewpoint image we need.
The geometric optics image formation model is more
powerful than pinhole camera model, i.e. it is a more
accurate description of what really happens in the
physical world. For example, when an object is placed
inside the front focal point of a real convex lens, it forms
an enlarged virtual image. This is completely
unexplainable if you treat the lens system as a perspective
pinhole. A perspective pinhole camera always gets real
images regardless of the object distance and image plane
distance, and it can never be out of focus. Baker and
Nayar [1] used a simple optical model to analyze some
omni-mirrors but they did not use any optical modeling
on cone mirrors. This turns out to be the key deficiency
in modeling the cone mirror/perspective camera
combination as we will elaborate shortly.
The cone system proposed here uses only off-the-shelf
ordinary lens and CCD camera. The cone is among the
simplest shape to produce, and it has much higher
resolution for scenes around the horizon [8]. It adds the
least optical distortion to the resulting image because it is
the only omni-view mirror with a longitudinally flat
surface. All other omni-mirror types are curved in both
longitudinal and lateral directions. The ability to pair
with readily available perspective camera instead of
expensive and complex orthographic cameras is also a
very important advantage over existing omni-view
This work has been supported in part by NSF-IIS-0083209, ARO/MURI-DAAH04-96-1-0007, NSF-CDS-97-03220, DARPA-ITO-DABT63-99-1-0017,
and Advanced Network and Services. Special thanks to members of MOOSE Project: V. J. Kumar, Kostas Daniilidis, Elli Angelopoulou, Oleg Naroditsky,
and Geoffrey Egnal for their invaluable comments and support. Also grateful thanks to members of GRASP Lab for their wonderful feedback and help.
systems. For example, when you want to build an omnisystem outside the visible range, it is often the case that
only perspective cameras are available. That is an added
reason why we go through all the trouble to prove and
build a cone mirror omni-directional camera system.
2. True single viewpoint cone mirror omniview system
As proved by Baker and Nayar [1], when the
viewpoint of a perspective camera coincides with the tip
of the cone, you have a single viewpoint omni-view
system. The only question then is whether or not you can
see an image in such configuration. The COPIS by Yagi
[10] did not put the lens camera viewpoint at the tip of
the cone because of this concern and thus did not form
single viewpoint omni-system. It turns out that we can
indeed see an image in this single viewpoint
configuration. We put the viewpoint of our lens camera
right at the tip of the cone and have a working true single
viewpoint cone mirror omni-view system. See Figure
1(geometry), Figure 2, Figure 3(full models) and Figure
8(real experimental setup).
image plane
effective viewpoint
of lens camera
p = (0,0)
effective viewpoint
v = (0,0)
A
a
B
M
mirror
N
b
Figure 1 True Single Viewpoint Cone Mirror
Geometry
In geometric optics image formation model [3], plane
mirror is an optical system by itself. The cross section of
a cone mirror through its axis of rotation consists of two
plane mirrors joined at the tip of the cone. The image
formation property of a plane mirror is that it always
forms virtual images behind its mirror surface. Notice
that the right side mirror does not actually extends to
point N, but that does not prevent the virtual image b
from forming. Within geometric optics, the position of
the virtual image point b does not change, either. As long
as there is some portion of the plane (ideally infinitely
large) is occupied with piece(s) of real mirror, the virtual
images corresponding to any world point on the front of
the mirror plane will be formed on the back of the mirror
plane. The positions of the virtual images are not effected
by the size of the actual mirror, but its “visibility” will
depend on the size and position of the actual mirror.
Since the virtual images are formed by the real plane
mirrors, the second converging optical system must ‘see’
the virtual image point through the real plane mirror.
One can regard real plane mirror as a “window” to look
at the “virtual world” on the other side of the mirror. The
surface of the cone mirror, after it accomplishes the task
of forming a virtual world behind it, acts as a window
opening for the 2nd converging optical system, i.e. the
perspective lens camera, to look into the virtual world. So
the cone is not blocking our view, it IS the view itself.
Whether the lens camera has its effective pinhole at the
cone tip or not, the lens camera “sees” the virtual image
points just like any other real world point. These virtual
points get imaged perspectively the same way as if there
is a real point there. In Figure 1, the virtual images a and
b formed by the catoptric system are imaged by the
dioptric system to form real images α and β respectively.
As proved by Baker and Nayar [1], if we can put the
effective pinhole of the lens camera at the tip of the cone
and still see image perspectively we have a single
viewpoint omni-view catadioptric system. We have just
shown that we can do this. We put the effective pinhole
of the lens camera right at tip of the cone and construct a
true single viewpoint cone mirror omni-directional view
catadioptric system. See Figure 8.
Full geometric optic ray tracing for the cone
combined with Gaussian optics ray tracing for the lens
camera will yield exactly the same results we just
showed. Notice here in geometric optics the term “ray
tracing” has very different meaning than the same term
used in computer graphics literatures. Computer graphics
“ray tracing” traces only one ray for one world point
while geometric optics “ray tracing” traces no less than 2
rays (ideally infinitely many) for one world point. The
computer graphic “ray tracing” is a simplified way of
finding image position valid only for focused real images.
Analyses based on tracing single rays alone can not
model how the single viewpoint of cone mirror omniview system works.
Since we need to trace many rays for each world
point, we trace only point B in Figure 2 and Figure 3 to
avoid confusion (too many rays in one small graph if we
trace both points). Interested reader can trace point A and
find it imaged exactly at the image position α.
In Figure 2 the tip of the cone is cut out a little so the
cone tip is not physically present. This does not alter the
single viewpoint geometry. Only the brightness of the
single viewpoint image is reduced a little. The two
configurations (Figure 2 and Figure 3) yield exactly the
same image for scenes points that are in focus vertically
(geometry-wise, the physical focus settings are different).
We shall explain details in the following section.
3. System Characteristics
3.1 Perfect vertical imaging
Curved surfaces, do not form exact image points for
all scene points. For example, parabolic mirror forms
perfect image only for points at infinity, while hyperbolic
mirror forms perfect virtual image only for one point, its
outside focal point. Baker and Nayar [1] has good
analysis of vertical defocus blur caused by vertically
curved surfaces. Because cone is the only vertically flat
omni-view mirror, the cone mirror based omni-view
device is the only device free of vertical defocus blur.
back focal point
O is projected to I via ray 2. In image space the two
image positions are always the same because the lengths
CH and VI are always the same. The lengths are the
same because under Gaussian optics conditions ray 1
goes parallel to VW after refracted by the lens and will
hit point I when the object is in-focus. Here the
conceptual camera at F does not produce any physical
image but the real image seen by camera at C will be
exactly the same since they are 1:1 orthographic
projection of each other.
Gauss developed the theory for paraxial rays only and
founded Gaussian optics [3]. Within Gaussian optics all
the rays that come from the same object point in the
world and enter the focused optical system converge at
the same point. This is an approximation but in practice
is very close to perfect in many situations. Given that,
trace any two such rays from a world point and we can
find its image point.
effective viewpoint
of lens camera #2
front focal point
p = (0,0)
real image plane
(CCD position)
lens
Virtual image plane
effective viewpoint
v = (0,0)
1.000
f
image plane
effective viewpoint
of lens camera #1
p = (0,0 )
effective viewpoint
v = (0,0 )
B
lens iris
B
4.000
So
mirror
N
mirror
b
N
b
Figure 3 Optical ray tracing using front focal point
of the lens as its virtual pin-hole
O
Object
Figure 2 Optical ray tracing using lens center as
virtual pin-hole
3.2 Dual effective pinholes for lens camera
This applies to most ordinary perspective lens
cameras that are in focus. In Figure 4, with a conceptual
perspective camera with viewpoint at F and image plane
at C(perpendicular to the axis WV), an arbitrary object
point O is projected to H via ray 1. While with a second
conceptual perspective camera with viewpoint at C and
image plane at V(perpendicular to WV), the same object
Lens(Aperture)
Back Focus Screen
(CCD Chip)
W
V
1
2
3
H
F
Front Focus C
(Pin Hole#2) Lens Center
(Pin Hole #1
I
Image of Object
Figure 4 Thin lens geometry implies dual effective
viewpoints
For convenience, we usually pick the two most easily
traced rays.
1. The ray that passes through the front focal point
will be refracted by the lens and then advances parallel to
the optical axis until it hits the image plane.
2. The ray that passes through the center of the lens
will pass straight through unaltered.
The Gaussian Thin Lens formula:[3] can be derived
directly from the same similar triangles we just used to
derive the dual effective pinholes.
BB
effective viewpoint
of lens camera #1
p = (0,0)
back focal point
image plane
effective viewpoint
v = (0,0)
lens iris
L
B
R
mirror
Equation 1 Gaussian Thin Lens Formula
1
so
+
1
si
=
N
1
f
where so object distance, si is image distance, and f is
focal length. The term “focal length” may cause some
confusion as many computer vision literature uses “focal
length” to mean the distance from the viewpoint to the
image plane. The term “focal length” in Gaussian optics
is a lens parameter that remains fixed while we change
the position of image plane. The two “conceptual
cameras” has different “object distance”/”image plane
distance” parameters but their resulting images are the
same.
This is why we can use the setup of either Figure 2 or
Figure 3 and get the same results. The choice of effective
pinhole effects only the value of so, si , and f, but not the
pixel position of any image point (compare Figure 2,
Figure 3, and Figure 4). Krishna and Ahuja [4] also use
the front focal point as the viewpoint for their panoramic
imaging device.
b
Figure 5 World points on the opposite side form
extremely weak images which are not visible in most
practical situations
Cone Omnisystem Vertical FOV
Real Perspective Camera
33.4°
FOV with cone mirror
cone tip
16./2
7°
Mirror Image of Half of the Camera
cone edge
53.3°
/2
106.6°
Cone Mirror
/2
53.3°
3.3 Image brightness and visibility
Since the profile of a cone consists of two plane
mirrors, each point inside the cone should be the virtual
image point of two world points simultaneously, one on
each side, see Figure 5.
In practice, we only “see” images on the same side of
the mirror. Figure 5 shows clearly that the point b is also
the position of the virtual image of point BB on the other
side. But in the same figure we can also see that very few
rays from point BB can actually reach the image plane
because of occlusion. Also, due to the symmetry of the
geometry, for a point b that is to the right of the central
axis, there is always more area of the mirror on the right
side (from tip of the cone p to point R) than the area of
mirror on the opposite side (from p to L) that is useful in
terms of imaging. So we always see B instead of BB in
practice because the image intensity of BB is far weaker
than that of B and thus the image of B always dominates
in practice.
/2
16.7°
FOV without cone mirror
Figure 6 Cone Shape and Vertical FOV
4. Unwarping algorithm
Unlike most existing omni-view system, the vertical
field of view for cone omni-system is not continuous
across the zenith. As shown in Figure 6, if the tip of the
cone subtends an angle α, and the normal angular
FOV(Field Of View) of the perspective camera is φ, then
the upper limit of viewable elevation is α-π/2. The
elevation here is defined to be 0 at level, 90 degree
straight up and -90 degree straight down. The same as
the system used for a gun turret. The lower bound of
viewable elevation is α-(π/2)-(φ/2).
When “unwarping”, assume the camera is perfectly
positioned and lined up, we establish a 2D image polar
coordinate system. For a given azimuth and elevation
angle in the “unwarped view”, the azimuth match the
polar angle directly. The polar radius variable r is related
to the elevation as
Equation 2
1.0044 pix^2 in the horizontal direction and 0.4425
pix^2 in the vertical direction. The square root of mean
squared distance differences is 1.2029 pix. This is
equivalent to 0.5 degree in view angle.
r = f p × tan(α − π2 − θ )
where fp is the effective focal length(image plane to
viewpoint distance) of real camera in real camera
pixels(i.e. the pixel unit in the original omni-view image,
not the pixels in the unwarped image) and θ is the
elevation angle of a point in the ‘unwarped’ view we
want to create.
Figure 10 Cone and Parabolic mirror placed side by
side; Right: Test pattern
5. Experiments
Figure 11 Omni-views, Left: Cone/perspective, Right:
parabolic/orthographic
Figure 7 Cone mirror prototype. Right: Overview of the
prototype setup.
Figure 12 Unwarped images. Left: Cone, more stripes
Figure 8 The tip of the cone actually coincides with the
effective pinhole. Right: cone omni-view seen by the
camera. The left side is a test pattern with dots 4” apart.
Figure 9 Actual perspective image taken by real
perspective camera (left) The unwarped perspective
image from the above looking at the same test pattern is
shown at the right.
To demonstrate the ability to correctly unwarp image
into perspective view, a test pattern of rectangular grids
is imaged in a omni-view picture. The right picture in
Figure 8 shows the omni-view containing the grids.
Figure 9 shows the unwarped result compared to an
actual picture taken by a normal perspective camera.
There is no visible radial distortion in Figure 9, which
means the radial distortion within our FOV is very small.
Tsai’s algorithm is suitable for compensating small
radial lens distortions. [7]
For the pattern of 45 points grids, the best match
between the two pictures has mean squared error of
preserved. Right: Parabolic, stripes tend to merge and not
separable due to lower vertical resolution.
To demonstrate the high vertical angular resolution
we put a custom made integrated parabolic/orthographic
system along side the cone prototype and let them look at
the same color foam pattern placed about 1m away.
In traditional non-omni cameras, there is a preferred
direction one can move the camera closer to the scene to
improve resolution, but in our omni camera case there is
no special direction to “move closer”. If you move the
camera “closer” toward one direction, the scene in the
opposite direction becomes “farther” at the same time. In
other words, cone mirror gets high resolution in all
direction easily in normal environment while other omni
cam system gets high resolution only in very confined
space where everything is close by.
The lens used is COSMICAR 6 mm 1:1.2 TV lens
stopped down to f/11 or f/16. A SONY XC-77 camera is
connected through a video box model PS-12SU made by
CHORI AMERICA. If your lens cannot get proper focus
within normal adjustment range, using extender rings
and/or use smaller aperture settings. Small aperture
setting in a dark room may need to digitally enhance the
brightness of the image. Some lens set has effective
viewpoints surrounded by light shield which blocks most
FOV. Avoid using such lens or remove the light shield
before use.
6. Discussions
The cone mirror is the only system that can get
perfect image vertically when you use a perfect lens
system. All other longitudinally curved mirrors introduce
defocus blurs before the light enter any lens system [1].
So even with a perfect lens system you can not get a
perfect image in those mirrors. In the rotational
symmetric direction, all omni-mirrors are curved with
the same horizontal cross-section shape: circles.
As a degenerate case of conic section, the cone mirror
does not have a full hemispherical view. It sees all 360
degrees in the horizontal direction, but the vertical field
of view (FOV) is not continuous over the zenith, see
Figure 6. Other omni-mirror system may in theory have
complete hemispherical view but in practice the lens
camera or secondary mirrors are always blocking the
view around the zenith, so in practice no similar system
can see the whole hemispherical view anyway. Smaller
vertical angular FOV is actually good for many
applications because we get higher angular resolution
with the same number of pixels concentrating on the
most interesting scene. The region not visible is occupied
by the lens camera, the sky or the floor, all of which are
of secondary interests in many applications like driving a
ground vehicle.
7. Conclusion
The Cone mirror has been identified as having a
single viewpoint before but that single viewpoint has
been regarded as useless because previous analysis was
done with the idealized pinhole camera geometric model.
This work proves that with more accurate modeling of
the optical properties of real cameras used in the real
world, the single viewpoint inherent in the cone mirror
geometry is usable. Prototype system based on our theory
is constructed and experiments confirm our theoretical
predictions. The total field of view is smaller but more
flexible. The new system is simple to build and maintain,
easier to analyze, costs a lot less, and yet yields much
higher performance in the peripheral region where most
interesting scenes reside.
References
[1] Baker, S. and Nayar, S. K., "A Theory of Single-Viewpoint
Catadioptric Image Formation," International Journal of
Computer Vision, vol. 35, no. 2, pp. 175-196, Nov.1999.
[2] Bogner, S. Introduction to panoramic imaging. 3100-3106.
1995. Proceedings of the IEEE SMC Conference.
[3] Hecht, E., Optics, 3 ed. Reading, MA, USA: Addison Wesley
Longman, Inc., 1998, pp. 1-694.
[4]
Krishna, A. and Ahuja, N. Panoramic image acquisition. 379384. 6-18-1996. San Francisco, CA. International Conference
on Computer Vision and Pattern Recognition.
[5] Nayar, S. K. and Karmarkar, A. 360 x 360 Mosaics. 2000.
Hilton Head Island, South Carolina. Proceedings of IEEE
Conference on Computer Vision and Pattern Recognition.
[6] Southwell, D., Vandegriend, B., and Basu, A. A Conical Mirror
Pipeline Inspection System. 4, 3253-3258. 4-22-1996.
Minneapolis, Minnesota, USA. Proceedings of the 1996 IEEE
International Conference on Robotics and Automation.
[7] Tsai, R. Y., "A Versatile Camera Calibration Technique for
High-Accuracy 3D Machine Vision Metrology Using Off-theShelf TV Cameras and Lenses," IEEE Journal of Robotics and
Automation, vol. RA-3, no. 4, pp. 323-344, Aug.1987.
[8] Yagi, Y., "Omnidirectional sensing and its applications," IEICE
TRANSACTIONS ON INFORMATION AND SYSTEMS, vol.
E82D, no. 3, pp. 568-579, Mar.1999.
[9] Yagi, Y. and Kawato, S. Panoramic scene analysis with conic
projection. 1990. Proceedings of the International Conference on
Robots and Systems.
[10] Yagi, Y., Kawato, S., and Tsuji, S., "Real-Time
Omnidirectional Image Sensor (COPIS) for Vision-Guided
Navigation," IEEE TRANSACTIONS ON ROBOTICS AND
AUTOMATION, vol. 10, no. 1, pp. 11-22, Feb.1994.