LandNet: Combine CNN and Transformer to Learn Absolute Camera Pose for the Fixed-Wing Aircraft Approach and Landing

Camera localization approaches often degrade in challenging environments characterized by illumination variations and significant viewpoint changes, presenting critical limitations for fixed-wing aircraft landing applications. To address these challenges, we propose LandNet—a novel absolute camera p...

Full description

Saved in:
Bibliographic Details
Main Authors: Siyuan Shen, Guanfeng Yu, Lei Zhang, Youyu Yan, Zhengjun Zhai
Format: Article
Language:English
Published: MDPI AG 2025-02-01
Series:Remote Sensing
Subjects:
Online Access:https://www.mdpi.com/2072-4292/17/4/653
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Camera localization approaches often degrade in challenging environments characterized by illumination variations and significant viewpoint changes, presenting critical limitations for fixed-wing aircraft landing applications. To address these challenges, we propose LandNet—a novel absolute camera pose estimation network specifically designed for airborne scenarios. Our framework processes images from forward-looking aircraft cameras to directly predict 6-DoF camera poses, subsequently enabling aircraft pose determination through rigid transformation. As a first step, we design two encoders from Transformer and CNNs to capture complementary spatial–temporal features. Furthermore, a novel <b>Feature Interactive Block (FIB)</b> is employed to fully utilize spatial clues from the CNN encoder and temporal clues from the Transformer encoder. We also introduce a novel Attentional Convtrans Fusion Block <b>(ACFB)</b> to fuse the feature maps from encoder and transformer encoder, which can enhance the image representations to promote the accuracy of the camera pose. Finally, two <b>Multi-Layer Perceptron (MLP)</b> heads are applied to estimate 6-DOF of camera position and orientation, respectively. Thus the estimated position and orientation of our LandNet can be further used to acquire the pose and orientation of the aircraft through the rigid connection between the airborne camera and the aircraft. The experimental results from simulation and real flight data demonstrate the effectiveness of our proposed method.
ISSN:2072-4292