Adapting Cross-Sensor High-Resolution Remote Sensing Imagery for Land Use Classification

High-resolution visible remote sensing imagery, as a fundamental contributor to Earth observation, has found extensive application in land use classification. However, the heterogeneous array of optical sensors, distinguished by their unique design architectures, exhibit disparate spectral responses...

Full description

Saved in:
Bibliographic Details
Main Authors: Wangbin Li, Kaimin Sun, Jinjiang Wei
Format: Article
Language:English
Published: MDPI AG 2025-03-01
Series:Remote Sensing
Subjects:
Online Access:https://www.mdpi.com/2072-4292/17/5/927
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:High-resolution visible remote sensing imagery, as a fundamental contributor to Earth observation, has found extensive application in land use classification. However, the heterogeneous array of optical sensors, distinguished by their unique design architectures, exhibit disparate spectral responses and spatial distributions when observing ground objects. These discrepancies between multi-sensor data present a significant obstacle to the widespread application of intelligent methods. In this paper, we propose a method tailored to accommodate these disparities, with the aim of achieving a smooth transfer for the model across diverse sets of images captured by different sensors. Specifically, to address the discrepancies in spatial resolution, a novel positional encoding has been incorporated to capture the correlation between the spatial resolution details and the characteristics of ground objects. To tackle spectral disparities, random amplitude mixup augmentation is introduced to mitigate the impact of feature anisotropy resulting from discrepancies in low-level features between multi-sensor images. Additionally, we integrate convolutional neural networks and Transformers to enhance the model’s feature extraction capabilities, and employ a fine-tuning strategy with dynamic pseudo-labels to reduce the reliance on annotated data from the target domain. In the experimental section, the Gaofen-2 images (4 m) and the Sentinel-2 images (10 m) were selected as training and test datasets to simulate cross-sensor model transfer scenarios. Also, Google Earth images of Suzhou City, Jiangsu Province, were utilized for further validation. The results indicate that our approach effectively mitigates the degradation in model performance attributed to image source inconsistencies.
ISSN:2072-4292