JpGU-AGU Joint Meeting 2020

Presentation information

[E] Oral

A (Atmospheric and Hydrospheric Sciences ) » A-CG Complex & General

[A-CG50] Earth & Environmental Sciences and Artificial Intelligence

convener:Tomohiko Tomita(Faculty of Advanced Science and Technology, Kumamoto University), Shigeki Hosoda(Japan Marine-Earth Science and Technology), Ken-ichi Fukui(Osaka University), Satoshi Ono(Kagoshima University)

[ACG50-04] Robustness of optical remote sensing of cloud by deep convolutional neural network using multi-channel imaging spectroradiometer

*Hironobu Iwabuchi1, Hana Kato1, Aki Masuda1 (1.Graduate School of Science, Tohoku University)

Keywords:remote sensing, cloud, deep neural network

Deep neural networks (DNNs) have recently been used in many problems of remote sensing. We have applied the DNNs to optical remote sensing of cloud from satellite- and ground-based imagers, for which the DNNs have been trained by a radiative transfer model (physics model). In optical remote sensing, three-dimensional (3D) radiative transfer effects significantly influence the cloud property retrieval from high-resolution optical instruments, operating across spatial elements of the cloudy atmosphere at a wide range of spatial scales. Because multispectral images contain spectral and spatial features, convolutional neural networks (CNNs) naturally trace the 3D radiative effects that appear across image pixels, which traditional single-pixel approaches cannot capture. Training data of synthetic radiances are made from simulations using a Monte Carlo 3D radiative transfer model for cloud fields simulated by a large eddy simulation model. The DNN can represent the cloud structure and 3D radiative effects that depend on many factors including sun-cloud-view geometry, cloud structure, aerosols, and surface reflection. Because satellite observation usually provide useful metadata information such as solar and satellite geometrical parameters, aerosol state, and surface reflectance, we have recently developed a new DNN architecture that is suitable for multimodal fusion of image and metadata. Furthermore, we found a way to properly treat measurement noise and bias in the DNN. If a DNN is trained on datasets of multispectral images with synthetic noise and bias of appropriate magnitude, cloud retrieval by the DNN is robust, not strongly depending on the absolute calibration and noise of instrument. If measured radiances are more uncertain, the DNN becomes to more strongly rely spatial features to estimate the cloud properties. We will show the robustness and sensitivity to noise and bias.