Ítem


Online View Planning for Inspecting Unexplored Underwater Structures

In this letter, we propose a method to automate the exploration of unknown underwater structures for autonomous underwater vehicles (AUVs). The proposed algorithm iteratively incorporates exteroceptive sensor data and replans the next-best-view in order to fully map an underwater structure. This approach does not require prior environment information. However, a safe exploration depth and the exploration area (defined by a bounding box, parameterized by its size, location, and resolution) must be provided by the user. The algorithm operates online by iteratively conducting the following three tasks: (1) Profiling sonar data are first incorporated into a 2-D grid map, where voxels are labeled according to their state (a voxel can be labeled as empty, unseen, occluded, occplane, occupied, or viewed). (2) Useful viewpoints to continue exploration are generated according to the map. (3) A safe path is generated to guide the robot toward the next viewpoint location. Two sensors are used in this approach: a scanning profiling sonar, which is used to build an occupancy map of the surroundings, and an optical camera, which acquires optical data of the scene. Finally, in order to demonstrate the feasibility of our approach, we provide real-world results using the Sparus II AUV

This work was supported by the EXCELLABUST and ARCHROV Projects under Grants H2020-TWINN-2015, CSA, ID: 691980, and DPI2014-57746-C3-3-R. The work of E. Vidal was supported by the Spanish Government through Ph.D. grant FPU14/05493

Institute of Electrical and Electronics Engineers (IEEE)

Director: Ministerio de Economía y Competitividad (Espanya)
Autor: Vidal Garcia, Eduard
Hernández Vega, Juan David
Istenič, Klemen
Carreras Pérez, Marc
Data: 17 febrer 2017
Resum: In this letter, we propose a method to automate the exploration of unknown underwater structures for autonomous underwater vehicles (AUVs). The proposed algorithm iteratively incorporates exteroceptive sensor data and replans the next-best-view in order to fully map an underwater structure. This approach does not require prior environment information. However, a safe exploration depth and the exploration area (defined by a bounding box, parameterized by its size, location, and resolution) must be provided by the user. The algorithm operates online by iteratively conducting the following three tasks: (1) Profiling sonar data are first incorporated into a 2-D grid map, where voxels are labeled according to their state (a voxel can be labeled as empty, unseen, occluded, occplane, occupied, or viewed). (2) Useful viewpoints to continue exploration are generated according to the map. (3) A safe path is generated to guide the robot toward the next viewpoint location. Two sensors are used in this approach: a scanning profiling sonar, which is used to build an occupancy map of the surroundings, and an optical camera, which acquires optical data of the scene. Finally, in order to demonstrate the feasibility of our approach, we provide real-world results using the Sparus II AUV
This work was supported by the EXCELLABUST and ARCHROV Projects under Grants H2020-TWINN-2015, CSA, ID: 691980, and DPI2014-57746-C3-3-R. The work of E. Vidal was supported by the Spanish Government through Ph.D. grant FPU14/05493
Format: application/pdf
Accés al document: http://hdl.handle.net/10256/14452
Llenguatge: eng
Editor: Institute of Electrical and Electronics Engineers (IEEE)
Col·lecció: info:eu-repo/semantics/altIdentifier/doi/10.1109/LRA.2017.2671415
info:eu-repo/semantics/altIdentifier/issn/2377-3766
info:eu-repo/grantAgreement/MINECO//DPI2014-57746-C3-3-R/ES/ARQUEOLOGIA MARINA MEDIANTE LA COOPERACION HROV%2FAUV/
info:eu-repo/grantAgreement/EC/H2020/691980/EU/Excelling LABUST in marine robotics/EXCELLABUST
Drets: Tots els drets reservats
Matèria: Vehicles submergibles
Submersibles
Robots autònoms
Autonomous robots
Algorismes computacionals
Computer algorithms
Títol: Online View Planning for Inspecting Unexplored Underwater Structures
Tipus: info:eu-repo/semantics/article
Repositori: DUGiDocs

Matèries

Autors