A Unified Multimodal Deep Learning Framework For Remote Sensing Imagery Classification
2021
In this paper, we present a unified deep learning framework for multimodal remote sensing image classification, U-MDL for short. U-MDL attempts to develop a general network architecture that consists of two subnetworks for feature extraction and feature fusion, respectively, with a focus on "which", "when", and "how" to fuse. For this purpose, we detail several common but effective fusion modules in the networks, e.g., early fusion, middle fusion, late fusion, and encoder-decoder fusion. These modules can be generalized well into our U-MDL framework. More significantly, we also emphasize to investigate a special case of multi-modality learning (MML), that is, cross-modality learning (CML) which widely exists in real applications. Moreover, extensive experiments are conducted to demonstrate the superiority and effectiveness of the proposed U-MDL framework in the remote sensing image classification task. The codes and datasets are available at: https://github.com/danfenghong/IEEE_TGRS_MDL-RS for the sake of reproducibility.
Keywords:
- Correction
- Source
- Cite
- Save
- Machine Reading By IdeaReader
20
References
0
Citations
NaN
KQI