PLL-VO: An Efficient and Robust Visual Odometry Integrating Point-Line Features and Neural Networks

Visual odometry is crucial for the navigation and planning of autonomous robots, but low-light conditions, dramatic lighting changes, and low-texture scenes pose significant challenges to odometry estimation. This paper proposes PLL-VO, which integrates point-line features and deep learning. To over...

Full description

Saved in:
Bibliographic Details
Main Authors: L. Zhao, Y. Yang, D. Ma, X. Lin, W. Wang
Format: Article
Language:English
Published: Copernicus Publications 2025-07-01
Series:ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences
Online Access:https://isprs-annals.copernicus.org/articles/X-G-2025/1045/2025/isprs-annals-X-G-2025-1045-2025.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1850112485135220736
author L. Zhao
Y. Yang
D. Ma
X. Lin
W. Wang
author_facet L. Zhao
Y. Yang
D. Ma
X. Lin
W. Wang
author_sort L. Zhao
collection DOAJ
description Visual odometry is crucial for the navigation and planning of autonomous robots, but low-light conditions, dramatic lighting changes, and low-texture scenes pose significant challenges to odometry estimation. This paper proposes PLL-VO, which integrates point-line features and deep learning. To overcome the impact of complex lighting conditions, a self-supervised learning method for interest point detection and a line detection algorithm that combines line optical flow tracking with cross-constraints is presented. After selecting keyframes based on point feature counts and line feature overlap angles, we integrate convolutional neural networks (CNNs) and graph neural networks (GNNs) to enhance sparse matching, thereby improving both accuracy and computational efficiency. PLL-VO system are evaluated in multiple datasets under various lighting conditions, demonstrating a 6.3% reduction in absolute trajectory error for pose estimation compared to state-of-the-art (SOTA) algorithms, the average computation time for visual odometry (43 ms) shows a 29.74% decrease compared to the state-of-the-art (SOTA) algorithms.
format Article
id doaj-art-16fdf389c1f64f728a58ee3ad1c72bb4
institution OA Journals
issn 2194-9042
2194-9050
language English
publishDate 2025-07-01
publisher Copernicus Publications
record_format Article
series ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences
spelling doaj-art-16fdf389c1f64f728a58ee3ad1c72bb42025-08-20T02:37:21ZengCopernicus PublicationsISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences2194-90422194-90502025-07-01X-G-20251045105210.5194/isprs-annals-X-G-2025-1045-2025PLL-VO: An Efficient and Robust Visual Odometry Integrating Point-Line Features and Neural NetworksL. Zhao0Y. Yang1D. Ma2X. Lin3W. Wang4School of Architecture and Urban Planning, Shenzhen University, Shenzhen 518060, ChinaSchool of Geomatics and Geographical Sciences, Liaoning Technical University, Fuxin 123000, ChinaSchool of Architecture and Urban Planning, Shenzhen University, Shenzhen 518060, ChinaSchool of Architecture and Urban Planning, Shenzhen University, Shenzhen 518060, ChinaSchool of Architecture and Urban Planning, Shenzhen University, Shenzhen 518060, ChinaVisual odometry is crucial for the navigation and planning of autonomous robots, but low-light conditions, dramatic lighting changes, and low-texture scenes pose significant challenges to odometry estimation. This paper proposes PLL-VO, which integrates point-line features and deep learning. To overcome the impact of complex lighting conditions, a self-supervised learning method for interest point detection and a line detection algorithm that combines line optical flow tracking with cross-constraints is presented. After selecting keyframes based on point feature counts and line feature overlap angles, we integrate convolutional neural networks (CNNs) and graph neural networks (GNNs) to enhance sparse matching, thereby improving both accuracy and computational efficiency. PLL-VO system are evaluated in multiple datasets under various lighting conditions, demonstrating a 6.3% reduction in absolute trajectory error for pose estimation compared to state-of-the-art (SOTA) algorithms, the average computation time for visual odometry (43 ms) shows a 29.74% decrease compared to the state-of-the-art (SOTA) algorithms.https://isprs-annals.copernicus.org/articles/X-G-2025/1045/2025/isprs-annals-X-G-2025-1045-2025.pdf
spellingShingle L. Zhao
Y. Yang
D. Ma
X. Lin
W. Wang
PLL-VO: An Efficient and Robust Visual Odometry Integrating Point-Line Features and Neural Networks
ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences
title PLL-VO: An Efficient and Robust Visual Odometry Integrating Point-Line Features and Neural Networks
title_full PLL-VO: An Efficient and Robust Visual Odometry Integrating Point-Line Features and Neural Networks
title_fullStr PLL-VO: An Efficient and Robust Visual Odometry Integrating Point-Line Features and Neural Networks
title_full_unstemmed PLL-VO: An Efficient and Robust Visual Odometry Integrating Point-Line Features and Neural Networks
title_short PLL-VO: An Efficient and Robust Visual Odometry Integrating Point-Line Features and Neural Networks
title_sort pll vo an efficient and robust visual odometry integrating point line features and neural networks
url https://isprs-annals.copernicus.org/articles/X-G-2025/1045/2025/isprs-annals-X-G-2025-1045-2025.pdf
work_keys_str_mv AT lzhao pllvoanefficientandrobustvisualodometryintegratingpointlinefeaturesandneuralnetworks
AT yyang pllvoanefficientandrobustvisualodometryintegratingpointlinefeaturesandneuralnetworks
AT dma pllvoanefficientandrobustvisualodometryintegratingpointlinefeaturesandneuralnetworks
AT xlin pllvoanefficientandrobustvisualodometryintegratingpointlinefeaturesandneuralnetworks
AT wwang pllvoanefficientandrobustvisualodometryintegratingpointlinefeaturesandneuralnetworks