FN Archimer Export Format PT C TI Combined use of a frame and a linear pushbroom camera for deep-sea 3D hyperspectral mapping BT AF Kumar, Peeyush Gracias, Nuno Istenic, Klemen Garcia, Rafael Arnaubec, Aurelien Ferrera, Maxime Bajjouk, Touria AS 1:1;2:2;3:2;4:2;5:3;6:3;7:3; FF 1:;2:;3:;4:;5:PDG-DFO-SM-PRAO;6:PDG-DFO-SM-PRAO;7:PDG-ODE-DYNECO-LEBCO; C1 Institute of Geodesy and Geoinformation, University of Bonn, Suisse Research Centre in Underwater Robotics (CIRS), University of Girona, Spain French Research Institute for Exploitation of the Sea (IFREMER), France C2 UNIV BONN, SWITZERLAND UNIV GIRONA, SPAIN IFREMER, FRANCE SI TOULON BREST SE PDG-DFO-SM-PRAO PDG-ODE-DYNECO-LEBCO IN WOS Ifremer UPR copubli-europe copubli-int-hors-europe UR https://archimer.ifremer.fr/doc/00759/87116/92616.pdf LA English DT Proceedings paper DE ;Geometry;Sea surface;Three-dimensional displays;Navigation;Nonlinear distortion;Cameras;Sensors AB Hyperspectral (HS) imaging produces an image of an object across a large range of the visible spectrum, and not just the primary colors (R, G, B) of conventional cameras. It can provide valuable information for object detection, analysis of materials and processes in environmental science in the deep-sea, especially for the study of benthic environments and pollution monitoring. In this paper, we address the problem of camera calibration towards 3D hyperspectral mapping where GPS is not available, and the platform navigational sensors are not accurate enough to allow direct georeferencing of linear sensors, as is the case with traditional aerial platform methods. Our approach presents a preliminary method for 3D hyperspectral mapping that uses only image processing techniques to reduce reliance on GPS or navigation sensors. The method is based on the use of standard RGB camera coupled with the hyperspectral pushbroom camera. The main contribution is the implementation and preliminary testing of a method to relate the two cameras using image information alone. The experiments presented in this paper analyze the estimation of relative orientation and time synchronization parameters for both cameras through experiments based on epipolar geometry and Monte-Carlo simulation. All methods are designed to work with real world data. PY 2021 PD SEP CT OCEANS 2021: San Diego – Porto, 2021, Electronic ISBN:978-0-692-93559-0, ISSN: 0197-7385. pp. 1-9 DI 10.23919/OCEANS44145.2021.9706053 ID 87116 ER EF