RGB-D salient object detection based on BC2 FNet network

In the face of complex scene images, the introduction of depth information can greatly improve the performance of salient object detection. However, up-sampling and down-sampling operations in neural networks maybe blur the boundaries of objects in the saliency map, thereby reducing the performance...

Full description

Saved in:
Bibliographic Details
Main Authors: WANG Feng, CHENG Yongmei
Format: Article
Language:zho
Published: EDP Sciences 2024-12-01
Series:Xibei Gongye Daxue Xuebao
Subjects:
Online Access:https://www.jnwpu.org/articles/jnwpu/full_html/2024/06/jnwpu2024426p1135/jnwpu2024426p1135.html
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1825206546455330816
author WANG Feng
CHENG Yongmei
author_facet WANG Feng
CHENG Yongmei
author_sort WANG Feng
collection DOAJ
description In the face of complex scene images, the introduction of depth information can greatly improve the performance of salient object detection. However, up-sampling and down-sampling operations in neural networks maybe blur the boundaries of objects in the saliency map, thereby reducing the performance of salient object detection. Aiming at this problem, a boundary-driven cross-modal and cross-layer fusion network (BC2FNet) for RGB-D salient object detection is proposed in this paper, which preserves the boundary of the object by adding the guidance of boundary information to the cross-modal and cross-layer fusion, respectively. Firstly, a boundary generation module is designed to extract two kinds of boundary information from low-level features of RGB and depth modalities, respectively. Secondly, a boundary-driven feature selection module is designed, which is dedicated to simultaneously focusing on important feature information and preserving boundary details in the process of RGB and depth modality fusion. Finally, a boundary-driven cross-layer fusion module is proposed which simultaneously adds two kinds of boundary information in the process of up-sampling fusion on adjacent layers. By embedding this module into the top-down information fusion flow, the predicted saliency map can contain accurate objects and sharp boundaries. Simulation results on five standard RGB-D data sets show that the proposed model can achieve better performance.
format Article
id doaj-art-1d4a45d0f5f34abca12dc0331b52edf8
institution Kabale University
issn 1000-2758
2609-7125
language zho
publishDate 2024-12-01
publisher EDP Sciences
record_format Article
series Xibei Gongye Daxue Xuebao
spelling doaj-art-1d4a45d0f5f34abca12dc0331b52edf82025-02-07T08:23:13ZzhoEDP SciencesXibei Gongye Daxue Xuebao1000-27582609-71252024-12-014261135114310.1051/jnwpu/20244261135jnwpu2024426p1135RGB-D salient object detection based on BC2 FNet networkWANG Feng0CHENG Yongmei1School of Physics and Electrical Engineering, Weinan Normal UniversitySchool of Automation, Northwestern Polytechnical UniversityIn the face of complex scene images, the introduction of depth information can greatly improve the performance of salient object detection. However, up-sampling and down-sampling operations in neural networks maybe blur the boundaries of objects in the saliency map, thereby reducing the performance of salient object detection. Aiming at this problem, a boundary-driven cross-modal and cross-layer fusion network (BC2FNet) for RGB-D salient object detection is proposed in this paper, which preserves the boundary of the object by adding the guidance of boundary information to the cross-modal and cross-layer fusion, respectively. Firstly, a boundary generation module is designed to extract two kinds of boundary information from low-level features of RGB and depth modalities, respectively. Secondly, a boundary-driven feature selection module is designed, which is dedicated to simultaneously focusing on important feature information and preserving boundary details in the process of RGB and depth modality fusion. Finally, a boundary-driven cross-layer fusion module is proposed which simultaneously adds two kinds of boundary information in the process of up-sampling fusion on adjacent layers. By embedding this module into the top-down information fusion flow, the predicted saliency map can contain accurate objects and sharp boundaries. Simulation results on five standard RGB-D data sets show that the proposed model can achieve better performance.https://www.jnwpu.org/articles/jnwpu/full_html/2024/06/jnwpu2024426p1135/jnwpu2024426p1135.htmlsalient object detectionboundary-drivencross-modal fusioncross-layer fusion
spellingShingle WANG Feng
CHENG Yongmei
RGB-D salient object detection based on BC2 FNet network
Xibei Gongye Daxue Xuebao
salient object detection
boundary-driven
cross-modal fusion
cross-layer fusion
title RGB-D salient object detection based on BC2 FNet network
title_full RGB-D salient object detection based on BC2 FNet network
title_fullStr RGB-D salient object detection based on BC2 FNet network
title_full_unstemmed RGB-D salient object detection based on BC2 FNet network
title_short RGB-D salient object detection based on BC2 FNet network
title_sort rgb d salient object detection based on bc2 fnet network
topic salient object detection
boundary-driven
cross-modal fusion
cross-layer fusion
url https://www.jnwpu.org/articles/jnwpu/full_html/2024/06/jnwpu2024426p1135/jnwpu2024426p1135.html
work_keys_str_mv AT wangfeng rgbdsalientobjectdetectionbasedonbc2fnetnetwork
AT chengyongmei rgbdsalientobjectdetectionbasedonbc2fnetnetwork