Exploring Machine Learning to Correct Satellite-Derived Sea Surface Temperatures

Type Article
Date 2018-02
Language English
Author(s) Picart Stephane Saux1, Tandeo Pierre2, Autret EmmanuelleORCID3, Gausset Blandine1
Affiliation(s) 1 : Meteo France Ctr Meteorol Spatiale, Ave Lorraine,BP 50747, F-22307 Lannion, France.
2 : IMT Atlantique, Lab STICC, UBL, F-29238 Brest, France.
3 : IFREMER, Lab Oceanog Phys & Spatiale, ZI Pointe Diable CS 10070, F-29280 Plouzane, France.
Source Remote Sensing (2072-4292) (Mdpi), 2018-02 , Vol. 10 , N. 2 , P. 224 (11p.)
DOI 10.3390/rs10020224
WOS© Times Cited 8
Note This article belongs to the Collection Sea Surface Temperature Retrievals from Remote Sensing
Keyword(s) machine learning, systematic error, sea surface temperature, random forest
Abstract

Machine learning techniques are attractive tools to establish statistical models with a high degree of non linearity. They require a large amount of data to be trained and are therefore particularly suited to analysing remote sensing data. This work is an attempt at using advanced statistical methods of machine learning to predict the bias between Sea Surface Temperature (SST) derived from infrared remote sensing and ground “truth” from drifting buoy measurements. A large dataset of collocation between satellite SST and in situ SST is explored. Four regression models are used: Simple multi-linear regression, Least Square Shrinkage and Selection Operator (LASSO), Generalised Additive Model (GAM) and random forest. In the case of geostationary satellites for which a large number of collocations is available, results show that the random forest model is the best model to predict the systematic errors and it is computationally fast, making it a good candidate for operational processing. It is able to explain nearly 31% of the total variance of the bias (in comparison to about 24% for the multi-linear regression model).

Full Text
File Pages Size Access
Publisher's official version 11 2 MB Open access
Top of the page