Underwater Optical Mapping Rafael Garcia and Nuno Gracias Computer Vision and Robotics Group University of Girona, 17071 Girona, Spain Tel. +34 972 419 812 [email protected], [email protected] Abstract ─ This work presents ongoing work at the University of Girona towards development and application of vision-based seafloor survey methodologies, including large area 2D mosaicing (>1sqkm), monocular-based 3D mosaicing, and stereo seafloor modeling. The developed tools set a first step towards using Autonomous Underwater Vehicles for detecting and documenting the temporal variations associated with the active processes operating at these sites. Keywords ─ mosaicing, mapping, underwater optical imaging, 3D reconstruction. 1. INTRODUCTION Robot mapping has greatly advanced in the last few years as a tool for environmental monitoring and seafloor characterization. Seafloor imagery is routinely acquired during near-bottom mapping surveys conducted with both remotely operated vehicles (ROVs) and autonomous underwater vehicles (AUVs). Deep-sea hydrothermal fields or shallow-water coral reef communities are, for instance, two scenarios that have long been the target of such studies. Imagery is useful to characterize the nature and distribution of geological features and biological communities, extract ecological indicators, and to provide a permanent visual record of the seafloor condition. However, imaging studies often yield large numbers of images (several tens of thousands, especially in deep-sea cruises) that are frequently underutilized largely because of the difficulties inherent in processing and visualizing large data sets [1]. Moreover, light suffers from a rapid and nonlinear attenuation underwater that affects the acquired images. In this paper we describe ongoing work at the University of Girona towards development and application of vision-based seafloor survey methodologies, including large area 2D mosaicing (>1sqkm), monocular-based 3D mosaicing, and stereo seafloor modeling. The developed tools set a first step towards detecting and documenting the temporal variations associated with the active processes operating at these sites. 2. OPTICAL MAPPING A large number of underwater studies rely on visual inspection, frequently using optical maps when the area to be explored is larger than what can be covered by a single image. Optical mapping can be divided into two large groups: (a) twodimensional mosaicing and (b) three-dimensional mapping. Two-dimensional image mosaicing deals with the process of combining the information from multiple images of the same area, to create a single representation with extended field of view [2,3] In order to construct ocean floor photomosaics, the individual images forming the mosaic are usually obtained by setting a camera on an underwater vehicle [4]. The camera looks down and the acquired images cover a small area of the ocean floor. The automatic detection of a number of points in one image, and their correspondences in another, allows the estimation of the motion between the images [5]. Using the motion estimations, the images are aligned and merged, forming composite view of the surveyed area. This approach needs to face additional challenges such as non-uniform illumination, light scattering and moving objects. All these effects cause strong visual artifacts when several images are stitched together to form a photo-mosaic. For this reason, once we have a globally aligned mosaic, image blending techniques are required to obtain a seamless mosaic. Conventional blending techniques used in terrestrial imagery are not always adequate in the underwater context due to light attenuation, suspended particles (producing light scattering), strong parallax and frequent moving elements, which are typical in underwater imagery. Fig. 1 illustrates the result of blending a composite mosaic into a seamless high-resolution image to provide a meaningful representation of the seafloor. Although very valuable, there are cases in which the information provided by 2D mosaics is not sufficient for comprehensive understanding of the region of interest. Studies on the evolution of coral reefs in biology and ecology, hydrothermal vents 100m 1m 3m Fig. 1. Sample 2D photo-mosaic of the MOMAR08 cruise in the mid-Atlantic Ridge. The mosaic represents 3 full days of survey with a Remotely Operated Vehicle and covers an area of nearly one square kilometer. After global alignment of all the images, the mosaic has been blended to obtain a high-resolution seamless picture of the seafloor. and lava formations in geology or underwater archeological sites, among others, heavily benefit from three-dimensional (3D) maps. Side-scanning sonars and acoustic cameras offer a solution to 3D seafloor imaging. Although the precision of sonars has improved significantly [6], sonars do not provide photometric information required to obtain textured 3D models. The combined use of sonars and cameras [7,8], to provide optical texture has so far been very limited, mainly due to the differences between the effective ranges of the two sensors. Side-scanning sonars are employed at altitudes exceeding 10 m, which generally surpasses the range of cameras due to light scattering and absorption. A more flexible approach is to use optical data to obtain both photometric and geometric information of the seafloor. This can be achieved by estimating the scene geometry using disparity information either between multiple calibrated cameras, known as stereovision [9], or by a single moving camera [10]. When a single camera is used, the 3D reconstruction problem becomes more complex as it requires determining the 3D camera motion for each frame [11]. This problem, referred to as structure from motion (SFM), has largely been studied by the computer vision community [12]. Oriented toward underwater imaging, [13] proposed a SLAM approach using navigation priors. This approach targets applications that require maintaining the vehicle position and velocity and their associated uncertainty. More oriented toward 3D mapping, [14] proposed a SFM framework that deals with large sequences by independently processing local submaps and registering them using global alignment techniques. While accurate, this approach has somewhat limited applications as it uses navigation priors for submap generation. More recently, our group [15] proposed a SFM-based large scale 3D mapping approach that increases the flexibility of the data acquisition process and the mapping accuracy using camera pose registration techniques. Figure 2 illustrates the result of our approach to create accurate three-dimensional textured models of the seafloor using monocular video sequences. The method takes into account the geometry of the scene through a 3D vertex selection mechanism which results in a reduction in the complexity of the final 3D model, with minimal loss of precision [15]. The technique does not require any navigation priors, and data can be acquired by submersible robots or divers using either video or still cameras, under various lighting conditions. Moreover, as the estimates of the camera position do not depend directly on the previous estimations, it is more robust to misregistration problems and scene occlusions than most state-of-the-art approaches. Although impressive technical progress is being achieved in underwater mapping technology, it has so far mainly concentrated in the use of a few sensor modalities (for example optical cameras and position data) or in increasing the autonomy of autonomous vehicles. There is a clear comparative lack of effort on the joint processing and integration of subsea data, focused on end-user scientist needs. This project is directed towards this goal. Despite of the extensive work in 2D and 3D optical mapping of the ocean floor, we still face a series of open issues for the efficient deployment of underwater imaging systems, mainly due to the intrinsic challenges of the medium [16]: light attenuation, backward and forward scattering and need for artificial illumination. These effects degrade the quality of the acquired images, resulting in image blur, very low contrast, reduced visibility range, and loss of color content. Thus, underwater optical imaging remains a challenging problem. 3. CONCLUSIONS We have presented a powerful set of processing tools to obtain a highly detailed maps of underwater interest areas, surveyed using either a stereo system or a monocular camera as the main sensor. The proposed methodologies do not assume any kind of a priori geometry on the reconstructed structure, so they are able to accurately reconstruct complex shapes over relatively large areas. Within this context, our approach is well-suited for conducting temporal studies through repeated surveys of the same area, enabling the detection of changes in these environments. The developments proposed in this paper will ultimately provide techniques and tools to a broad scientific community (biologists, ecologists, geologists and geophysicists), and that will allow advancements that are possible with existing data and processing tools. Furthermore, these tools will open the door to numerous non-scientific applications ranging from industry (offshore operations, Figure 2. Estimation of the 3D relief of the seafloor using a single calibrated camera. The model has approximately 240,000 vertices and covers an area of 12 × 19 m. drilling, mining) to the public domain (environmental impact and remediation, infrastructure construction, public works, risk assessment and education.), and will close the gap that exists in our knowledge of the seafloor with respect to the subaerial Earth’s surface. REFERENCES [1] [2] [3] [4] [5] [6] [7] J. Escartin, R. Garcia, O. Delaunoy, N. Gracias, A. Elibol, X. Cufí, L. Neumann, D.J.Fornari, S.E. Humphris, J. Renard, 2008, Globally aligned photomosaic of the Lucky Strike hydrothermal vent field (Mid-Atlantic Ridge, 37º18.5'N): Release of georeferenced data, mosaic construction, and viewing software, Geochemistry, Geophysics and Geosystems, vol. 9, no. 12, pp. 1-17 Irani, M., P. Anandan, S. Hsu, 1995. Mosaic based representations of video sequences and their applications. In Proc. of the 5th IEEE International Conference on Computer Vision, pp. 605-611, Cambridge, Massachusetts. Garcia, R., J. Batlle, X. Cufí, J. Amat, 2001. Positioning an Underwater Vehicle through Image Mosaicking. IEEE Int. Conf. on Robotics and Automation, vol. 3, pp. 2779– 2784, Seoul, Rep. of Korea. Eustice, R., H. Singh, J. Leonard, M. Walter, R. Ballard, 2005. Visually Navigating the RMS Titanic with SLAM Information Filters, In Proceedings of Robotics: Science and Systems (RSS) pp. 57 – 64, Cambridge, MA, USA. Negahdaripour, S., C. Barufaldi, A. Khamene, 2006. Integrated system for robust 6 DOF positioning utilizing new closed-form visual motion estimation methods over planar terrains. IEEE J. Oceanic Engineering Ruiz, I. T., de Raucourt, S., Petillot, Y., 2004. Concurrent Mapping and Localization Using Sidescan Sonar, IEEE J. of Oceanic Engineering, vol. 29, no. 2, pp. 442-456. I. Mahon and S. Williams, 2004. "Slam using natural features in an underwater environment,” in IEEE Control, Automation, Robotics and Vision Conference, vol. 3, pp 2076–2081. [8] [9] [10] [11] [12] [13] [14] [15] [16] S. Negahdaripour, H. Sekkati, H. Pirsiavash, 2009. “Optiacoustic stereo imaging: On system calibration and 3-D target reconstruction," IEEE Trans. Image Processing, IEEE Trans. Image Processing, Vol 18(6) pp. 1203-1214. Negahdaripour, S., H. Madjidi, 2003. Stereovision imaging on submersible platforms for 3-d mapping of benthic habitats and sea floor structures. IEEE J. Ocean. Eng, 28, pp. 625650. T. Nicosevici and R. Garcia, 2008. "On-line robust 3D Mapping using structure from motion cues," MTS/IEEE Techno-Ocean Conference (Oceans'08) Kobe. Y. Ma, S. Soatto, J. Kosecka, S. Shankar Sastry, 2003. An Invitation to 3-D Vision: From Images to Geometric Models. SpringerVerlag, vol. 26. A. W. Fitzgibbon, A. Zisserman, 1998. Automatic Camera Recovery for Closed or Open Image Sequences. European Conference on Computer Vision, pp. 311-326. R. M. Eustice, H. Singh, J. J. Leonard, 2006. Exactly Sparse Delayed-State Filters for View-Based SLAM, IEEE Transactions on Robotics, vol. 22, no. 6, pp. 11001114. O. Pizarro and R. Eustice and H. Singh, 2004. Large area 3D reconstructions from underwater surveys. MTS/IEEE OCEANS, vol. 2, pp. 678-687. T. Nicosevici, N. Gracias, S. Negahdaripour, R. Garcia, 2009. "Efficient three-dimensional Scene Modeling and Mosaicing," Journal of Field Robotics, vol. 26, no. 10, pp. 759-788. W. S. Pegau, D. Gray, J. R. V. Zaneveld, 1997. Absorption and attenuation of visible and near-infrared light in water: dependence on temperature and salinity. no.36, pp.60356046. Acknowledgement: This work has been funded in part by MICINN under grant CTM2010-15216/MAR, and in part by the EU under grants FP7-ICT-2009-248497 and FP7- IEF-2009-253322. N. Gracias has been funded by the Ramon y Cajal programme.
© Copyright 2026 Paperzz