Abstract
Purpose of Review
This paper addresses the benefits and challenges of mixed reality (MR) for the exploration of deep-sea environments with remotely operated vehicles. The approach is twofold: virtual reality (VR) let the scientist explore the environment via a visual 3D model, overcoming limitations of local perception. Augmented reality (AR) concepts are designed in order to improve environment perception and interaction.
Recent Findings
The key to such concepts is the implementation of 3D visual geo-referenced terrain models from the imaging feedback gathered by the vehicle exploring its unknown surroundings. Image processing, underwater vehicle navigation, and user-friendly displays for robotic intervention are addressed in an integrated concept. A broad development programme carried out at the French Institute for Ocean Science, IFREMER, is described and illustrates technical topics and use cases.
Summary
3D perception derived from camera vision is shown to enable AR concepts that will significantly improve remote exploration and intervention in unknown natural environments. Cumulative geo-referenced 3D model building is in the process of being taken to reliable functioning in real-world underwater applications, accomplishing a milestone change in the capacity to view and understand the obscure and inaccessible deep-sea world.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
Introduction
The scientific exploration of deep-sea environments represents continuously renewing challenges for underwater technology. Investigation and research are related to major societal questions such as biodiversity, global change, living resources, mineral, or fossil reservoirs and in questions related to the impact of human activity on our planet. Relying primarily on the use of remotely operated deep-sea vehicles (ROV), the achievement of underwater research missions depends on the technical capacity to precisely navigate, to provide reliable visual and spatial information, to carry out precise measurements, to collect representative samples of various nature (sediment, mineral, fauna, water, etc.), and to deploy stationary equipment on the seabed. ROVs used by IFREMER for scientific research are shown in Fig. 1.
Highly relevant ecosystems and habitats are generally located in rugged terrain with strong slopes, such as canyons, cliffs, and deposits of mineral resources originated from seismic or volcanic activity, for example, around highly vertical hydrothermal vents. As a result, the scientific investigator and the operator of deep-sea vehicles must obtain a broad and accurate understanding of the local topography and develop situation awareness regarding the local working environment, in order to accomplish efficient intervention tasks. Live video images from several cameras located on the front end of the vehicle represent today the main source of operator feedback. Real-time observation allows the operator and scientific end users alike to interpret the work environment, to identify objects of interest, and to perform tasks and manoeuvres. Direct observation over 2D images represents however a limiting factor to performance and efficiency of the intervention tasks: the human operator must supervise the task execution from several two-dimensional viewpoints in order to deduce a three-dimensional (3D) approach of the scene. This approximated 3D interpretation of the scene is necessary even when performing simple tasks such as placing or handling objects (tools, instruments, samples, etc.) and estimating the position and shape of target samples, which is often made before collection.
The challenge is thus to move towards a concept of augmented perception that implements the idea of transporting humans into deep-waters—this transport will ultimately take place by virtual means. Recent virtual or augmented reality techniques often involve linking known structural models to sensory data. In the present field considering investigation of the natural environment, 3D visual models are in general not available before the dive. In order to make use of emerging techniques related to augmented reality, a model of the environment needs to be reconstructed on-the-fly, and the robotic tasks must be set in perspective with the continuously upgraded model.
This paper is organised as follows: Section 2 presents the recent advances in underwater augmented reality (UWAR), Section 3 develops the concept of virtual transportation into the deep-sea, Section 4 presents the current developments at IFREMER towards intuitive and reliable UWAR applications, and Section 5 draws conclusion and perspectives.
Mixed Reality for Underwater Situational Awareness
Divers, ROV pilots, and oceanographic researchers may benefit from advances in mixed reality (MR) technologies [2] by using it as an enriched 3D perception tool to increase situational awareness. The prime benefit being a better understanding of the terrain topology and features, MR allows to avoid collision with the sea floor, to improve navigation security and intervention efficiency.
Virtual reality (VR) has been used offline to virtually visit underwater sites [3] and to train operators by simulating control interfaces to a simulated environment [4], for near real-time visualisation assistance to track objects in underwater construction and maintenance interventions [5] and for trajectory tracking [6]. Augmented reality [7] was used in the context of underwater cultural heritage [8, 9]. Waterproof tablets have been designed where guidance notes, navigation, and a virtual reconstruction of the archaeological site are displayed [9, 10]. At the current state, applications rely on a prior environment model as is available, e.g. in industrial construction sites. This is a key difference opposing most of such applications to oceanographic exploration.
In order to blend virtual information into the current user view, it is needed to register the virtual world with real world. Thus, the 3D environment must either be previously known or, at least locally, reconstructed. The 3D terrain model is built by processing and integrating information from several sensors, such as monocular [11] and stereoscopic cameras [12••] [9], sonars, structured light devices, laser profilers, and LiDAR systems [13].
If the site can be previously equipped with landmarks, AR markers are widely used as anchors to superimpose 3D virtual objects or information in user view [8, 14, 15]. A combination of visual markers and real-time 3D reconstruction can be used in order to allow onshore operation of underwater robots through haptic and visual interfaces with a virtual world [16].
In the case of exploration of unknown site, 3D sensors such as stereo rig [9] or multibeam echo sounders give local 3D information through point clouds and depth maps. This representation can be used to estimate the pose of a manipulator w.r.t. the environment, to display the relative distance between the end effector and the environment [12••], or to project arm contours in the user video feedback [9].
As soon as several viewpoints are available, structure from motion (SfM) techniques [16, 17••] or simultaneous localisation and mapping (SLAM) [18] have proven their maturity in 3D environment modelling. They can use either monocular [19, 20] or stereo cameras [12••, 21•, 22,23,24,25]. Moreover, stereo VSLAM tends to achieve more accurate results for ego-motion estimation [26]. This will then allow to support the navigation process in cluttered environments and help and improve the consistence of large DTM.
The Perception Leap: Travelling Virtually into the Deep Sea
The ongoing technological development effort in the field of remote perception is specifically targeting to overcome the limitations of exploring the underwater environment through traditional ROV’s cameras. Novel techniques are being introduced based on augmented reality and 3D mapping that will allow the scientific and operational end users to benefit from three-dimensional real-time representation of the underwater environment.
ROV pilots ensure the key interface role between the scientists’ objectives, the intervention tool, and the environment. Following the tasks defined by the scientists, the pilot analyses the feasibility of the requested operations and estimates the best positioning for the ROV, in order to successfully carry out the manipulations. Under many respects, the quality of the scientific end result depends on the successful interaction among these key players. Developing tools to help improve and optimise this interaction is a constantly sought for objective. Consequently, we way witness the progressive evolution of roles where the scientist interacts in a more direct manner with the modelled environment in an augmented reality scenario while the ROV pilot assures a backseat technical supervision of the task. The technical machine will become transparent, giving the scientist a sensory, perceptive, and gestural whole relative to an extended scene. The operator will play a role of supervisor who guides the operation and monitors safety, providing advice on practical and operational aspects.
New remote operations techniques will enable virtual movements, observations, and actions that interface with the marine environment, part of an approach progressively imitating that of humans in terrestrial environments. Changes in piloting modes and the performance of scientific operations will be based on several key functionalities associated with enhanced reality:
-
a)
3D visual model of the underwater environment
Depth measurements from acoustic sounders and optical tools (photogrammetry, laser profiling, and LIDAR) will be integrated in a high-resolution 3D model that will integrate environmental data for the entire dive and for a group of dives. The 3D model is fitted with the visual texture provided by photo-video optical imagery. These techniques are now available offline. Real-time, entirely automated availability will be possible in the short to medium term. Viewing these models is possible either on standard or 3D screens or in a virtual reality approach through VR goggles or HoloLens. Given that the 3D model is geo-referenced, dimensional analysis is directly achievable by graphic operations.
-
b)
Integration of multimodal data
Digital terrain model (DTM) will be enriched with dedicated layers representing data from specific sensors, either from online or past acquisitions. Currently used in 2D or 3D GIS tools offline, the discipline layers will provide precious assistance for exploring sites. Differential analysis between earlier and new data represents a particularly rich perspective (characterisation of changes in the environment over time).
-
c)
On demand virtual viewpoint changing
Scientists will be able to navigate in a virtual environment, without being limited by the actual field of view of the imaging sensor (see Fig. 2). Visualisation of the working spaces that can be reached by equipment (arms, probes, samplers, etc.) in the modelled environment will also make it possible to plan and guide intervention. It also provides the basis for new graphical ergonomics that moves controls into the environment model.
-
d)
Towards the performance of entirely automated manipulation tasks
Scientists will plan intervention by using virtual tools in the DTM. All operations will be facilitated through fully automated functions such as virtual capture of manipulation targets (corals, rocks, etc.) and automatic management of standard tools (corers, suction devices, probes, etc.).
-
e)
Multiple use model and shared exploration
Progressive construction of the model will be adapted to the distribution by satellite link to terrestrial laboratories. Remote presence equipment will rely on immersive techniques in multidimensional models of the environment. Without being present at the place of operation, teams can participate in remote work (complementary skills, etc.) in real or differed time. Annotations and other points of interest will enable organising collective work through a shared interface.
Underwater Augmented Reality Developments at IFREMER
Several key functional aids for the vehicle operator and scientific end user are currently under development at IFREMER. Until now, video feedback from one or several cameras looking to the intervention site is the main source of information the operators use to control the manipulators and the vehicle. Through augmented reality, the operator can use an enriched representation of the environment with data overlay from several sources and processing algorithms without losing focus on the operation. On the top of that, 3D maps can be analysed collaboratively by scientists.
The acquisition and processing of the optical data will feed two distinct uses. The first one is the in-dive real-time 3D processing which provide a better understanding of the environment and assistance tools. The second use case occurs after the dive with data post-processing for increased accuracy on large-scale environment modelling for scientific use.
In-dive processing blends virtual data in the local field of view as the vehicle explores the scene, and it permits to acquire data to shape a wider environment model progressively until the work area has been covered sufficiently. As a major advantage, AR allows to use local 3D data in order to improve interaction between the vehicle manipulators and the scene. Techniques focus on merging video and analytic information and display ergonomics which come in as a key factor for operator acceptance (see Fig. 3).
Post-processing uses data sets over a defined time-frame–computation-intensive solutions can be employed in order to cover an extensive area of interest. Resulting 3D models, which have a high resolution and accuracy, will represent a standard for post-expedition scientific investigation and are used in virtual reality concepts. The technical solutions for 3D model construction as well as scientific analysis of said models must handle very large data sets englobing tens of thousands of high-resolution images (see Fig. 2).
The aim of the remote perception and intervention processes will be to merge the two above approaches in a single enhanced reality concept.
Underwater Optical Data Processing
On state-of-the-art scientific ROVs, the use of high-quality video or still images from monocular cameras is widespread, as optical images are an essential source of information for the understanding and analysis of the deep underwater environment.
However, in an underwater context, light propagation properties and disturbances in the sea water such as backscattering, diffusion, and colour and light attenuation degrade the quality of the image formation. Moreover, lighting is provided by artificial sources mounted on the vehicle, and, as a result, shadows are cast on the scene change as the robot moves. Colour attenuation leads features to appear differently when observed at varying distances. To cope with these issues, an image correction procedure is a pre-requisite to any image-based reconstruction. Our computation process includes a pre-processing stage to correct images for nonuniform illumination and colorimetric degradation [27].
Imaging surveys are accomplished to cover a work area with enough overlap in the successive image frames, allowing to reconstruct the scene into a single large and geo-referenced model of the seafloor. This is the purpose of 3D optical mapping.
Gathered and corrected images can be used to solve the structure from motion problem [17••] in order to obtain a sparse 3D reconstruction. It is progressively densified by spreading knowledge from the already triangulated points [28]. Finally, from the 3D point cloud, a mesh is constructed [29], on which texture from the initial image sequence is blended [30] at native resolution. Sensor fusion with the vehicle navigation data ensures geo-referencing the three-dimensional model.
In a complementary approach, the use of stereo sensors allows to immediately obtain three-dimensional information in the camera’s field of view. Stereovision photogrammetry provides a depth field directly associated with the image frame. The use of stereo rigs is hence envisaged in the aim of providing on-the-fly quantitative local information that aids piloting and intervention tasks.
Augmented Reality Tools
Persistent model perception and stereovision field-of-view depth computation are two alternative ways to produce 3D information of the scene. Augmented reality concepts can be applied to persistent digital terrain model as well as on the go 3D reconstruction.
Augmented reality tools can be implemented and deployed immediately in the local field of view (see Fig. 3). Scene maps may be augmented with information and interactive measurement tools that will assist pilots and scientists in the comprehension of the site topography as well as in the execution of tasks. A similar approach can be exploited in post-processing, this time benefitting from a posteriori integration of local point clouds into an extended environment model representation.
-
a)
Topological functionalities
Depth maps are processed to compute online topological measurements such as bathymetric level curves, linear or curvilinear distance, areas, and slopes. Three-dimensional distance w.r.t. the environment is displayed in gradient colour, yielding to a primary perception of the scene 3D structure.
Level curves such as isobaths can help to better understand the slope of the terrain and its orientation w.r.t. the robot as well as the size of objects. Level curves perpendicular to the camera optical axis are useful to understand the diameter and the longitudinal profile of confined spaces such as underwater caves.
Slopes can be calculated from the 3D information as the local average inclination of segmented areas of interest in the images. This is helpful during the exploration of very rugged terrains such as hydrothermal vents, canyons, and cliffs.
The computation of areas, which are defined in the image as 2D non-convex polygons, is based on a local meshing of the terrain inside the polygon [31, 32].
-
b)
Overlay of scientific data information
Physicochemical measurements collected during the dive using the scientific payload can be geo-referenced and displayed in real time, allowing a deeper understanding of the underlying phenomena that characterise the ecosystem’s dynamics. Finally, annotations, comments, and short analysis made by scientists can be geo-referenced and attached to the terrain model. They can thus be shared with colleagues and be used in further offline studies.
-
c)
Assistance for ROV manoeuvre
The processed 3D model of the environment allows computing distances between selected points in the image and specific reference frames of the robot, such as camera frame and robot arm frame. The manipulator arm’s reach on the environment can be calculated from the intersection of the manipulator envelop diagram and the local 3D model of the terrain. The projection of the end effector axis and the modelled environment can also be graphically visualised, to help the operator decide the suitable trajectory to control arm movement (in Cartesian control frame for instance) towards a predefined goal.
Before positioning the vehicle on the seabed, augmented reality display helps the pilots analyse the nature of the terrain. The same tools allow them to visualise the footprint of the vehicle on the seabed and the reachability diagram of the manipulator arm. These graphical representations allow to better identify the suitable landing configuration with regard to the planned intervention task. This in turn minimises the potential impact on the environment (sediment up stirring, contact with structures, etc.) through unnecessary repeated landing retries.
Post-processing Tools
-
a)
3D reconstruction for simplified non-expert use: Matisse
The request for 3D reconstruction tools of the underwater seabed by non-expert users such as oceanographers and biologists has led to the design of a fully automated, graphic user-friendly software called Matisse [33]. This software is based on several open source libraries such as open MVG [34] or mvs-texturing [30] that have emerged to solve the large-scale offline 3D reconstruction problem with good accuracy. An example of 3D geo-referenced textured model produced by Matisse is shown in Fig. 4. This kind of model generates a global representation of the site and provides accurate and measurable information of the environment in a way that goes far beyond the perception the user has during the actual dive.
-
b)
Handling of large 3D models for geo-spatial analysis: 3D Metrics
Multiple 3D processing and software visualisation tools (CloudCompare, Meshlab, AutoDesk suite) exist, some of which offering advanced analysis algorithms—yet none of them were fully adapted to scientific geo-referenced data exploitation. In order to handle geo-referenced 3D models, textured meshes, and point clouds, we developed a new software application called 3D Metrics. Simultaneous visualisation of multiple datasets such as large-scale acoustic models and optical mapping is a basic feature. Annotation functions, as well geometric operators such as distances, surface, and slope estimation (see Fig. 4), provide the environment for scientific analysis. Models and measurement layers can also be ortho-projected and exported to be used in classical 2D GIS software.
-
c)
Virtual immersion in digital terrain model
3D models can be used by end users to explore the 3D scene with a perception that is not limited by camera field of view or by the range of lighting. Virtual reality tools such as motion-sensing stereo headsets are used for collaborative virtual exploration as well as for demonstrations to the general public. The quantitative exploitation of 3D geo-referenced models through computer applications or VR immersive techniques has proven a strong potential for the remote work with the robot vehicle on the sea floor.
Conclusion and Outlook
This paper provides an oversight on current developments and applications on Underwater Augmented Reality (UWAR) at IFREMER. In recent years, techniques to build textured 3D maps of the seabed from video or photo sequences were successfully implemented. Three-dimensional models are computed by means of structure-from-motion (SfM) techniques from image sequences issued by high-end optical imaging devices. Geo-referencing is obtained by data fusion algorithms using the vehicle navigation data. Resulting high-resolution 3D maps are used by scientists for immersive revisiting of the work sites and for quantitative geometric analysis of the site topography. The offline process is evolving towards an on-dive 3D perception of the environment, thanks to fast 3D point cloud computation with stereo cameras and real-time visual simultaneous location and mapping (visual SLAM) algorithms. Fast, online processing allowed for augmented reality image overlays on operator camera views, greatly enhancing the ability to perform remotely operated tasks on complex environments.
The applications presented in the article include examples of UWAR functions designed to improve intervention task efficiency and reduce potential environment disturbance. Scientists and pilots will be able to discuss and exchange more precisely, since reliable information will be provided in real time aiding the decision-making process.
A more accurate 3D perception of the environment will lead to greater automation of robotic tasks, as the pilot will play a role of supervisor, guiding and monitoring the safety of the operation. Emerging techniques in recent years, such as artificial intelligence (AI), should also be used for shared virtual exploration of the environment model, combined with collaborative functions such as manual or automatic feature tagging for characterisation and classification of seabed sediments and specimen, which allow to further enhance the qualitative perception of the work site [35, 36].
Perspectives for future work concern semi-autonomous manoeuvring for positioning and grasping tasks [37]. The scientific user might then mimic the intervention in the virtual environment, the ROV being able to carry out the task in the real scene.
References
Papers of particular interest, published recently, have been highlighted as: • Of importance •• Of major importance
E. Raugel, J. Opderbecke, M. Fabri, L. Brignone and V. Rigaud, “Operational and scientific capabilities of Ariane, Ifremer’s hybrid ROV,” in Oceans, Marseille, 2019.
Freiherr von Lucas U, John Q, Kaklis P, Tim D. Underwater mixed environments. In: Virtual Realities; 2015.
Bruno F, Lagudi A, Barbieri L, Cozza M, Cozza A, Peluso R, et al. Virtual tour In the Sunken "Villa con ingresso a protiro", within the underwater archaeological park of Baiae. ISPRS - international archives of the photogrammetry, remote sensing and spatial information sciences. 2019:45–51.
Qingping L, Chengi K. On applying virtual reality to underwater robot tele-operation and pilot training. International Journal of Virtual Reality. 2001;5(1):71–91.
Santamaria JC, Opdenbosh A. Monitoring underwater operations with virtual environments. In: Offshore Technology Conference; 2002.
Davis B, Patron P, Arredondo M, Lane D. Augmented reality and data fusion techniques for enhanced situational awareness of the underwater domain. In: OCEANS; 2007.
Azuma R, Baillot Y, Behringer R, Feiner S, Julier S, Macintre B. Recent advances in augmented reality. IEEE Computer Graphics and Applications. 2001:34–47.
Cejka J, Zsiros A, Liarokapis F. A hybrid augmented reality guide for underwater cultural heritage sites. In: Personal and Ubiquitous Computing; 2020.
A. Casals, J. Fernández and J. Amat, “Augmented reality to assist teleoperation working with reduced visual conditions,” in IEEE International Conference on Robotics and Automation, Washington, DC, USA, USA, 2002.
R. Morales, P. Keitler, P. Maier and G. Klinker, "An underwater augmented reality system for commercial diving operations, in OCEANS 2009, MTS/IEEE - Marine Technology for Our Future: Global and Local Challenges, Biloxi, 2009.
M. Chouiten, C. Domingues, J.-Y. Didier, S. Otmane and M. Mallem, “Distributed mixed reality for diving and underwater tasks using remotely operated vehicles,” International Journal on Computational Sciences & Applications, AIRCC, vol. 5, no. 4, p. (elec. Proc.), 2014.
•• Bruno F, Lagudi A, Barbieri L, Rizzo D, Muzzupappa M, De Napoli L. Augmented reality visualization of scene depth for aiding ROV pilots in underwater manipulation. Ocean Engineering. 2018;168:140–54. In this paper, experimental results of underwater augmented reality are provided in the context of underwater cultural heritage, proving the feasibility of using real-time AR function in underwater missions.
Massot-Campos M, Oliver-Codina G. Optical sensors and methods for underwater 3D reconstruction. Sensors. 2015;12:31525–31,557.
B. Abdelkader, D. Christophe, O. Samir, B. Samir and D. Alain, “Augmented reality for underwater activities with the use of the DOLPHYN.,” in 10th IEEE International Conference on Networking, Sensing and Control (ICNSC 2013), Evry2013
Oppermann L, Blum L, Lee J-Y, Seo J-H. AREEF multi-player underwater augmented reality experience. In: IEEE International Games Innovation Conference (IGIC): Vancouver; 2013.
Bryson M, Johnson-Roberson M, Friedman A, Pizarro O, Troni G, Ozog P, et al. High-resolution underwater robotic vision-based mapping and three-dimensional reconstruction for archaeology. Journal of Field Robotics. 2017;34:625–43.
•• Istenič K, Gracias N, Arnaubec A, Escartín J, Garcia R. Automatic scale estimation of structure from motion based 3D models using laser scalers in underwater scenarios. ISPRS Journal of Photogrammetry and Remote Sensing. 159:13–25. This paper provides a comprehensive explanation on the structure from motion problem pipeline. Novel approaches are proposed to solve the scaling problem and experimental results are evaluated.
F. Hidalgo and T. Bräunl, “Review of underwater SLAM techniques,” in International Conference on Automation, Robotics and Applications, 2015.
V. Creuze, “Odometry for underwater vehicles with online estimation of the scale factor,” in IFAC 2017World Congress, Toulouse, France, 2017.
Ferrera M, Moras J, Trouvé-Peloux P, Creuze V. Real-time monocular visual odometry for turbid and dynamic underwater environments: Sensors; 2019.
• S. Rahman, A. Q. Li and I. Rekleitis, “Sonar visual inertial SLAM of underwater structures,” in IEEE International Conference on Robotics and Automation, Brisbane, QLD, 2018. This paper provides recent results on real-time mapping of underwater environments through the fusion of visual, acoustic and inertial sensors. Such techniques are useful for in-dive augmented reality applications
A. Palomer, P. Ridao and D. Riba, “Multibeam 3D underwater SLAM with probabilistic registration sensors,” Sensors, vol. 16, no. 4, p. (elec. proc), 2016
Lagudi A, Bianco G, Muzzupappa M, Bruno F. An alignment method for the integration of underwater 3D data captured by a stereovision system and an acoustic camera. Sensors. 2016;16:536.
Bianco G, Gallo A, Bruno F, Muzzupappa M. A comparative analysis between active and passive techniques for underwater 3D reconstruction of close-range objects. Sensors. 2013;13:11007–11,031.
Letessier T, Juhel J-B, Vigliola L, Meeuwig J. Low-cost small action cameras in stereo generates accurate underwater measurements of fish. Journal of Experimental Marine Biology and Ecology. 2015;466:120–6.
R. Giubilato, M. Pertile and S. Debei, “A comparison of monocular and stereo visual FastSLAM implementations,” in IEEE Metrology for Aerospace (MetroAeroSpace), Florence, Italy, 2016.
Bryson M, Johnson-Roberson M, Pizarro O, Williams SB. True color correction of autonomous underwater vehicle imagery. Journal of Field Robotics. 2016;6:33.
Shen S. Accurate multiple view 3D reconstruction using patch-based. IEEE Transactions on Image Processing. 2013;22:1901–14.
Jancosek M, Pajdla T. Exploiting visibility information in surface reconstruction to preserve weakly supported surfaces: Hindawi Publishing Corporation; 2014.
M. Waechter, M. N and M. Goesele “Let there be color! Large-scale texturing of 3D reconstructions.,” Cham, 2014.
Shewchuk J. Triangle: engineering a 2D quality mesh generator and Delaunay triangulator. Applied Computational Geometry: Towards Geometric Engineering. 1996:203–22.
Centin M, Pezzotti N, Signoroni A. Poisson-driven seamless completion of triangular meshes. Computer Aided Geometric Design. 2015;35–36:42–55.
Arnaubec A, Opderbecke J, Allais A-G, Brignone L. Optical mapping with the ARIANE HROV at Ifremer: The MATISSE processing tool. In: Oceans; 2015.
Moulon P, Monasse P, Perrot R, Marlet R. “Openmvg,” in International Workshop on Reproducible Research in Pattern Recognition, Open multiple view geometry. International Workshop on Reproducible. .
Diesing M, Stephens D. A multi-model ensemble approach to seabed mapping. Journal of Sea Research. 2015;100:62–9.
Salman A, Jalal A, Shafait F, Mian A, Shortis M, Seager J, et al. Fish species classification in unconstrained underwater environments based on deep learning. Limnology and Oceanography: Methods. 2016;14:570–85.
S. Sivčev, M. Rossi, J. Coleman, G. Dooly, E. Omerdić and D. Toal, “Fully automatic visual servoing control for work-class marine intervention ROVs,” 2018.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of Interest
The authors declare that they have no conflict of interest.
Human and Animal Rights and Informed Consent
This article does not contain any studies with human or animal subjects performed by any of the authors.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
This article belongs to the Topical Collection on Underwater Robotics
Rights and permissions
About this article
Cite this article
Laranjeira, M., Arnaubec, A., Brignone, L. et al. 3D Perception and Augmented Reality Developments in Underwater Robotics for Ocean Sciences. Curr Robot Rep 1, 123–130 (2020). https://doi.org/10.1007/s43154-020-00014-5
Published:
Issue Date:
DOI: https://doi.org/10.1007/s43154-020-00014-5