Novel Fusion Technique for High-Performance Automated Crop Edge Detection in Smart Agriculture

Optimising vegetable production systems is crucial for maintaining and enhancing agricultural productivity, particularly for crops like lettuce. Separating the crop from the background poses a significant challenge when using automated tools. To address this, a novel technique has been developed to...

Full description

Saved in:
Bibliographic Details
Main Authors: F. Martinez, James B. Romaine, P. Johnson, A. Cardona Ruiz, Pablo Millan Gata
Format: Article
Language:English
Published: IEEE 2025-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10858138/
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1825207052102795264
author F. Martinez
James B. Romaine
P. Johnson
A. Cardona Ruiz
Pablo Millan Gata
author_facet F. Martinez
James B. Romaine
P. Johnson
A. Cardona Ruiz
Pablo Millan Gata
author_sort F. Martinez
collection DOAJ
description Optimising vegetable production systems is crucial for maintaining and enhancing agricultural productivity, particularly for crops like lettuce. Separating the crop from the background poses a significant challenge when using automated tools. To address this, a novel technique has been developed to automatically detect the vegetative area of lettuces, optimising time and eliminating subjectivity during crop inspections. The proposed deep learning model integrates the YOLOv10 object detector, the K-means classifier, and a segmentation method known as superpixel. This combination enables lettuce area identification using bounding box labels instead of contour labels during training, improving efficiency compared to other methods like YOLOv8 and Detectron2. Additionally, the combination of the YKMS method with YOLOv8 (YKMSV8) is evaluated, where YKMS serves as a label assistant. These methods are also used as benchmarks to compare the proposed approach. For the training of each methods, a custom database has been created using a low-cost, low-power custom IoT node deployed on a real farm to provide the most accurate data. Throughout the comparison, a custom metric is used to evaluate performance both in training and inference, balancing computational cost and area error, making it applicable in agriculture. Performance metric is associated with computational cost factor and accuracy factor whose value are respectively 65% and 35%, ensuring applicability for autonomous agricultural devices. Computational cost is prioritised to maintain battery life during extended campaigns. The results of the custom metric during inference indicated that the YKMSV8 method achieved the highest performance, followed by Detectron2, YOLOv8, and, lastly, YKMS. Regarding area error, YOLOv8 exhibited the lowest mean error, followed by Detectron2, while YKMSV8 and YKMS produced similar values. In terms of inference time, YKMSV8 was the most computationally efficient, followed by YOLOv8, YKMS, and, finally, Detectron2.
format Article
id doaj-art-caa16f200c0248c4ad153c1d8b7d351d
institution Kabale University
issn 2169-3536
language English
publishDate 2025-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj-art-caa16f200c0248c4ad153c1d8b7d351d2025-02-07T00:01:06ZengIEEEIEEE Access2169-35362025-01-0113224292244510.1109/ACCESS.2025.353670110858138Novel Fusion Technique for High-Performance Automated Crop Edge Detection in Smart AgricultureF. Martinez0https://orcid.org/0000-0002-5732-6154James B. Romaine1https://orcid.org/0000-0002-2117-5833P. Johnson2https://orcid.org/0000-0003-2379-9700A. Cardona Ruiz3https://orcid.org/0009-0004-4387-1209Pablo Millan Gata4https://orcid.org/0000-0002-6129-0094Departamento de Ingeniería, Universidad Loyola Andalucía, Seville, SpainDepartamento de Ingeniería, Universidad Loyola Andalucía, Seville, SpainSchool of Engineering, Liverpool John Moores University, Liverpool, U.K.Departamento de Ingeniería, Universidad Loyola Andalucía, Seville, SpainDepartamento de Ingeniería, Universidad Loyola Andalucía, Seville, SpainOptimising vegetable production systems is crucial for maintaining and enhancing agricultural productivity, particularly for crops like lettuce. Separating the crop from the background poses a significant challenge when using automated tools. To address this, a novel technique has been developed to automatically detect the vegetative area of lettuces, optimising time and eliminating subjectivity during crop inspections. The proposed deep learning model integrates the YOLOv10 object detector, the K-means classifier, and a segmentation method known as superpixel. This combination enables lettuce area identification using bounding box labels instead of contour labels during training, improving efficiency compared to other methods like YOLOv8 and Detectron2. Additionally, the combination of the YKMS method with YOLOv8 (YKMSV8) is evaluated, where YKMS serves as a label assistant. These methods are also used as benchmarks to compare the proposed approach. For the training of each methods, a custom database has been created using a low-cost, low-power custom IoT node deployed on a real farm to provide the most accurate data. Throughout the comparison, a custom metric is used to evaluate performance both in training and inference, balancing computational cost and area error, making it applicable in agriculture. Performance metric is associated with computational cost factor and accuracy factor whose value are respectively 65% and 35%, ensuring applicability for autonomous agricultural devices. Computational cost is prioritised to maintain battery life during extended campaigns. The results of the custom metric during inference indicated that the YKMSV8 method achieved the highest performance, followed by Detectron2, YOLOv8, and, lastly, YKMS. Regarding area error, YOLOv8 exhibited the lowest mean error, followed by Detectron2, while YKMSV8 and YKMS produced similar values. In terms of inference time, YKMSV8 was the most computationally efficient, followed by YOLOv8, YKMS, and, finally, Detectron2.https://ieeexplore.ieee.org/document/10858138/Computer visionobject detectionYOLOv8 segmentationYOLOv10superpixelK-means
spellingShingle F. Martinez
James B. Romaine
P. Johnson
A. Cardona Ruiz
Pablo Millan Gata
Novel Fusion Technique for High-Performance Automated Crop Edge Detection in Smart Agriculture
IEEE Access
Computer vision
object detection
YOLOv8 segmentation
YOLOv10
superpixel
K-means
title Novel Fusion Technique for High-Performance Automated Crop Edge Detection in Smart Agriculture
title_full Novel Fusion Technique for High-Performance Automated Crop Edge Detection in Smart Agriculture
title_fullStr Novel Fusion Technique for High-Performance Automated Crop Edge Detection in Smart Agriculture
title_full_unstemmed Novel Fusion Technique for High-Performance Automated Crop Edge Detection in Smart Agriculture
title_short Novel Fusion Technique for High-Performance Automated Crop Edge Detection in Smart Agriculture
title_sort novel fusion technique for high performance automated crop edge detection in smart agriculture
topic Computer vision
object detection
YOLOv8 segmentation
YOLOv10
superpixel
K-means
url https://ieeexplore.ieee.org/document/10858138/
work_keys_str_mv AT fmartinez novelfusiontechniqueforhighperformanceautomatedcropedgedetectioninsmartagriculture
AT jamesbromaine novelfusiontechniqueforhighperformanceautomatedcropedgedetectioninsmartagriculture
AT pjohnson novelfusiontechniqueforhighperformanceautomatedcropedgedetectioninsmartagriculture
AT acardonaruiz novelfusiontechniqueforhighperformanceautomatedcropedgedetectioninsmartagriculture
AT pablomillangata novelfusiontechniqueforhighperformanceautomatedcropedgedetectioninsmartagriculture