Multi-scale sparse convolution and point convolution adaptive fusion point cloud semantic segmentation method
Abstract Semantic segmentation of LIDAR point clouds is essential for autonomous driving. However, current methods often suffer from low segmentation accuracy and feature redundancy. To address these issues, this paper proposes a novel approach based on adaptive fusion of multi-scale sparse convolut...
Saved in:
Main Authors: | , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Nature Portfolio
2025-02-01
|
Series: | Scientific Reports |
Online Access: | https://doi.org/10.1038/s41598-025-88905-5 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1823862480160423936 |
---|---|
author | Yuxuan Bi Peng Liu Tianyi Zhang Jialin Shi Caixia Wang |
author_facet | Yuxuan Bi Peng Liu Tianyi Zhang Jialin Shi Caixia Wang |
author_sort | Yuxuan Bi |
collection | DOAJ |
description | Abstract Semantic segmentation of LIDAR point clouds is essential for autonomous driving. However, current methods often suffer from low segmentation accuracy and feature redundancy. To address these issues, this paper proposes a novel approach based on adaptive fusion of multi-scale sparse convolution and point convolution. First, addressing the drawbacks of redundant feature extraction with existing sparse 3D convolutions, we introduce an asymmetric importance of space locations (IoSL) sparse 3D convolution module. By prioritizing the importance of input feature positions, this module enhances the sparse learning performance of critical feature information. Additionally, it strengthens the extraction capability of intrinsic feature information in both vertical and horizontal directions. Second, to mitigate significant differences between single-type and single-scale features, we propose a multi-scale feature fusion cross-gating module. This module employs gating mechanisms to improve fusion accuracy between different scale receptive fields. It utilizes a cross self-attention mechanism to adapt to the unique propagation features of point features and voxels, enhancing feature fusion performance. Experimental comparisons and ablation studies conducted on the SemanticKITTI and nuScenes datasets validate the generality and effectiveness of the proposed approach. Compared with state-of-the-art methods, our approach significantly improves accuracy and robustness. |
format | Article |
id | doaj-art-592b6f67b78846ac9b00063d2a86a726 |
institution | Kabale University |
issn | 2045-2322 |
language | English |
publishDate | 2025-02-01 |
publisher | Nature Portfolio |
record_format | Article |
series | Scientific Reports |
spelling | doaj-art-592b6f67b78846ac9b00063d2a86a7262025-02-09T12:29:45ZengNature PortfolioScientific Reports2045-23222025-02-0115111710.1038/s41598-025-88905-5Multi-scale sparse convolution and point convolution adaptive fusion point cloud semantic segmentation methodYuxuan Bi0Peng Liu1Tianyi Zhang2Jialin Shi3Caixia Wang4School of Electronic Information Engineering, Changchun University of Science and TechnologySchool of Electronic Information Engineering, Changchun University of Science and TechnologySchool of Electronic Information Engineering, Changchun University of Science and TechnologySchool of Electronic Information Engineering, Changchun University of Science and TechnologySchool of Electronic Information Engineering, Changchun University of Science and TechnologyAbstract Semantic segmentation of LIDAR point clouds is essential for autonomous driving. However, current methods often suffer from low segmentation accuracy and feature redundancy. To address these issues, this paper proposes a novel approach based on adaptive fusion of multi-scale sparse convolution and point convolution. First, addressing the drawbacks of redundant feature extraction with existing sparse 3D convolutions, we introduce an asymmetric importance of space locations (IoSL) sparse 3D convolution module. By prioritizing the importance of input feature positions, this module enhances the sparse learning performance of critical feature information. Additionally, it strengthens the extraction capability of intrinsic feature information in both vertical and horizontal directions. Second, to mitigate significant differences between single-type and single-scale features, we propose a multi-scale feature fusion cross-gating module. This module employs gating mechanisms to improve fusion accuracy between different scale receptive fields. It utilizes a cross self-attention mechanism to adapt to the unique propagation features of point features and voxels, enhancing feature fusion performance. Experimental comparisons and ablation studies conducted on the SemanticKITTI and nuScenes datasets validate the generality and effectiveness of the proposed approach. Compared with state-of-the-art methods, our approach significantly improves accuracy and robustness.https://doi.org/10.1038/s41598-025-88905-5 |
spellingShingle | Yuxuan Bi Peng Liu Tianyi Zhang Jialin Shi Caixia Wang Multi-scale sparse convolution and point convolution adaptive fusion point cloud semantic segmentation method Scientific Reports |
title | Multi-scale sparse convolution and point convolution adaptive fusion point cloud semantic segmentation method |
title_full | Multi-scale sparse convolution and point convolution adaptive fusion point cloud semantic segmentation method |
title_fullStr | Multi-scale sparse convolution and point convolution adaptive fusion point cloud semantic segmentation method |
title_full_unstemmed | Multi-scale sparse convolution and point convolution adaptive fusion point cloud semantic segmentation method |
title_short | Multi-scale sparse convolution and point convolution adaptive fusion point cloud semantic segmentation method |
title_sort | multi scale sparse convolution and point convolution adaptive fusion point cloud semantic segmentation method |
url | https://doi.org/10.1038/s41598-025-88905-5 |
work_keys_str_mv | AT yuxuanbi multiscalesparseconvolutionandpointconvolutionadaptivefusionpointcloudsemanticsegmentationmethod AT pengliu multiscalesparseconvolutionandpointconvolutionadaptivefusionpointcloudsemanticsegmentationmethod AT tianyizhang multiscalesparseconvolutionandpointconvolutionadaptivefusionpointcloudsemanticsegmentationmethod AT jialinshi multiscalesparseconvolutionandpointconvolutionadaptivefusionpointcloudsemanticsegmentationmethod AT caixiawang multiscalesparseconvolutionandpointconvolutionadaptivefusionpointcloudsemanticsegmentationmethod |