Deep-Sea Fauna Segmentation: A Comparative Analysis of Convolutional and Vision Transformer Architectures at Lucky Strike Vent Field

Due to recent technological developments, the acquisition and availability of deep-sea imagery has increased exponentially in the last years, leading to an increasing backlog in image annotation and processing, attributable to limited specialized human resources. In this work, we investigate the performance of well-established convolutional neural networks and Vision Transformer (ViT) based architectures, namely, DeepLabv3+ and UNETR, for the segmentation of fauna in deep-sea images. The dataset consists of images captured at the Lucky Strike Vent field, located on the mid-Atlantic ridge, of three edifices named Montsegur, White Castle, and Eiffel Tower. Our experimental investigation reveals that the Vision Transformer consistently outperforms the fully convolutional deep learning architecture, by approximately 14% in terms of F1-Score, demonstrating the effectiveness of ViTs in capturing intricate patterns and long-range dependencies present in deep-sea imagery. Our findings highlight the potential of ViTs as a promising approach for accurate semantic segmentation in challenging environmental contexts, paving the way for improved understanding and analysis of deep-sea ecosystems.

Keyword(s)

Fauna Segmentation, Deep Learning, Hydrothermal Vents, Vision Transformers, Convolutional Networks

Full Text

FilePagesSizeAccess
Publisher's official version
91 Mo
How to cite
Soto Vega Pedro J., Andrade-Miranda Gustavo X., Da Costa Gilson A. O. P., Papadakis Panagiotis, Matabos Marjolaine, Napoleon Thibault, Karine Ayoub, Fagundes Gasparoto Henrique (2024). Deep-Sea Fauna Segmentation: A Comparative Analysis of Convolutional and Vision Transformer Architectures at Lucky Strike Vent Field. ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences. X-3-2024. 387-395. https://doi.org/10.5194/isprs-annals-X-3-2024-387-2024, https://archimer.ifremer.fr/doc/00918/103033/

Copy this text