Multi-scale sparse convolution and point convolution adaptive fusion point cloud semantic segmentation method

Abstract Semantic segmentation of LIDAR point clouds is essential for autonomous driving. However, current methods often suffer from low segmentation accuracy and feature redundancy. To address these issues, this paper proposes a novel approach based on adaptive fusion of multi-scale sparse convolut...

Full description

Saved in:
Bibliographic Details
Main Authors: Yuxuan Bi, Peng Liu, Tianyi Zhang, Jialin Shi, Caixia Wang
Format: Article
Language:English
Published: Nature Portfolio 2025-02-01
Series:Scientific Reports
Online Access:https://doi.org/10.1038/s41598-025-88905-5
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1823862480160423936
author Yuxuan Bi
Peng Liu
Tianyi Zhang
Jialin Shi
Caixia Wang
author_facet Yuxuan Bi
Peng Liu
Tianyi Zhang
Jialin Shi
Caixia Wang
author_sort Yuxuan Bi
collection DOAJ
description Abstract Semantic segmentation of LIDAR point clouds is essential for autonomous driving. However, current methods often suffer from low segmentation accuracy and feature redundancy. To address these issues, this paper proposes a novel approach based on adaptive fusion of multi-scale sparse convolution and point convolution. First, addressing the drawbacks of redundant feature extraction with existing sparse 3D convolutions, we introduce an asymmetric importance of space locations (IoSL) sparse 3D convolution module. By prioritizing the importance of input feature positions, this module enhances the sparse learning performance of critical feature information. Additionally, it strengthens the extraction capability of intrinsic feature information in both vertical and horizontal directions. Second, to mitigate significant differences between single-type and single-scale features, we propose a multi-scale feature fusion cross-gating module. This module employs gating mechanisms to improve fusion accuracy between different scale receptive fields. It utilizes a cross self-attention mechanism to adapt to the unique propagation features of point features and voxels, enhancing feature fusion performance. Experimental comparisons and ablation studies conducted on the SemanticKITTI and nuScenes datasets validate the generality and effectiveness of the proposed approach. Compared with state-of-the-art methods, our approach significantly improves accuracy and robustness.
format Article
id doaj-art-592b6f67b78846ac9b00063d2a86a726
institution Kabale University
issn 2045-2322
language English
publishDate 2025-02-01
publisher Nature Portfolio
record_format Article
series Scientific Reports
spelling doaj-art-592b6f67b78846ac9b00063d2a86a7262025-02-09T12:29:45ZengNature PortfolioScientific Reports2045-23222025-02-0115111710.1038/s41598-025-88905-5Multi-scale sparse convolution and point convolution adaptive fusion point cloud semantic segmentation methodYuxuan Bi0Peng Liu1Tianyi Zhang2Jialin Shi3Caixia Wang4School of Electronic Information Engineering, Changchun University of Science and TechnologySchool of Electronic Information Engineering, Changchun University of Science and TechnologySchool of Electronic Information Engineering, Changchun University of Science and TechnologySchool of Electronic Information Engineering, Changchun University of Science and TechnologySchool of Electronic Information Engineering, Changchun University of Science and TechnologyAbstract Semantic segmentation of LIDAR point clouds is essential for autonomous driving. However, current methods often suffer from low segmentation accuracy and feature redundancy. To address these issues, this paper proposes a novel approach based on adaptive fusion of multi-scale sparse convolution and point convolution. First, addressing the drawbacks of redundant feature extraction with existing sparse 3D convolutions, we introduce an asymmetric importance of space locations (IoSL) sparse 3D convolution module. By prioritizing the importance of input feature positions, this module enhances the sparse learning performance of critical feature information. Additionally, it strengthens the extraction capability of intrinsic feature information in both vertical and horizontal directions. Second, to mitigate significant differences between single-type and single-scale features, we propose a multi-scale feature fusion cross-gating module. This module employs gating mechanisms to improve fusion accuracy between different scale receptive fields. It utilizes a cross self-attention mechanism to adapt to the unique propagation features of point features and voxels, enhancing feature fusion performance. Experimental comparisons and ablation studies conducted on the SemanticKITTI and nuScenes datasets validate the generality and effectiveness of the proposed approach. Compared with state-of-the-art methods, our approach significantly improves accuracy and robustness.https://doi.org/10.1038/s41598-025-88905-5
spellingShingle Yuxuan Bi
Peng Liu
Tianyi Zhang
Jialin Shi
Caixia Wang
Multi-scale sparse convolution and point convolution adaptive fusion point cloud semantic segmentation method
Scientific Reports
title Multi-scale sparse convolution and point convolution adaptive fusion point cloud semantic segmentation method
title_full Multi-scale sparse convolution and point convolution adaptive fusion point cloud semantic segmentation method
title_fullStr Multi-scale sparse convolution and point convolution adaptive fusion point cloud semantic segmentation method
title_full_unstemmed Multi-scale sparse convolution and point convolution adaptive fusion point cloud semantic segmentation method
title_short Multi-scale sparse convolution and point convolution adaptive fusion point cloud semantic segmentation method
title_sort multi scale sparse convolution and point convolution adaptive fusion point cloud semantic segmentation method
url https://doi.org/10.1038/s41598-025-88905-5
work_keys_str_mv AT yuxuanbi multiscalesparseconvolutionandpointconvolutionadaptivefusionpointcloudsemanticsegmentationmethod
AT pengliu multiscalesparseconvolutionandpointconvolutionadaptivefusionpointcloudsemanticsegmentationmethod
AT tianyizhang multiscalesparseconvolutionandpointconvolutionadaptivefusionpointcloudsemanticsegmentationmethod
AT jialinshi multiscalesparseconvolutionandpointconvolutionadaptivefusionpointcloudsemanticsegmentationmethod
AT caixiawang multiscalesparseconvolutionandpointconvolutionadaptivefusionpointcloudsemanticsegmentationmethod