An autonomous navigation method for field phenotyping robot based on ground-air collaboration

High-throughput phenotyping collection technology is important in affecting the efficiency of crop breeding. This study introduces a novel autonomous navigation method for phenotyping robots that leverages ground-air collaboration to meet the demands of unmanned crop phenotypic data collection. The...

Full description

Saved in:
Bibliographic Details
Main Authors: Zikang Zhang, Zhengda Li, Meng Yang, Jiale Cui, Yang Shao, Youchun Ding, Wanneng Yang, Wen Qiao, Peng Song
Format: Article
Language:English
Published: KeAi Communications Co., Ltd. 2025-12-01
Series:Artificial Intelligence in Agriculture
Subjects:
Online Access:http://www.sciencedirect.com/science/article/pii/S2589721725000601
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1850235368438235136
author Zikang Zhang
Zhengda Li
Meng Yang
Jiale Cui
Yang Shao
Youchun Ding
Wanneng Yang
Wen Qiao
Peng Song
author_facet Zikang Zhang
Zhengda Li
Meng Yang
Jiale Cui
Yang Shao
Youchun Ding
Wanneng Yang
Wen Qiao
Peng Song
author_sort Zikang Zhang
collection DOAJ
description High-throughput phenotyping collection technology is important in affecting the efficiency of crop breeding. This study introduces a novel autonomous navigation method for phenotyping robots that leverages ground-air collaboration to meet the demands of unmanned crop phenotypic data collection. The proposed method employs a UAV equipped with a Real-Time Kinematic (RTK) module for the construction of high-precision Field maps. It utilizes SegFormor-B0 semantic segmentation models to detect crop rows, and extracts key coordinate points of these rows, and generates navigation paths for the phenotyping robots by mapping these points to actual geographic coordinates. Furthermore, an adaptive controller based on the Pure Pursuit algorithm is proposed, which dynamically adjusts the steering angle of the phenotyping robot in real-time, according to the distance (d), angular deviation (α) and the lateral deviation (ey) between the robot's current position and its target position. This enables the robot to accurately trace paths in field environments. The results demonstrate that the mean absolute error (MAE) of the proposed method in extracting the centerline of potted plants area's rows is 2.83 cm, and the cropland's rows is 4.51 cm. The majority of global path tracking errors stay within 2 cm. In the potted plants area, 99.1 % of errors lie within this range, with a mean absolute error of 0.62 cm and a maximum error of 2.59 cm. In the cropland, 72.4 % of errors remain within this range, with a mean absolute error of 1.51 cm and a maximum error of 4.22 cm. Compared with traditional GNSS-based navigation methods and single vision methods, this method shows significant advantages in adapting to the dynamic growth of crops and complex field environments, which not only ensures that the phenotyping robot accurately travels along the crop rows during field operations to avoid damage to the crops, but also provides an efficient and accurate means of data acquisition for crop phenotyping.
format Article
id doaj-art-c8335e384abf4e99b0963cc34a447782
institution OA Journals
issn 2589-7217
language English
publishDate 2025-12-01
publisher KeAi Communications Co., Ltd.
record_format Article
series Artificial Intelligence in Agriculture
spelling doaj-art-c8335e384abf4e99b0963cc34a4477822025-08-20T02:02:17ZengKeAi Communications Co., Ltd.Artificial Intelligence in Agriculture2589-72172025-12-0115461062110.1016/j.aiia.2025.05.005An autonomous navigation method for field phenotyping robot based on ground-air collaborationZikang Zhang0Zhengda Li1Meng Yang2Jiale Cui3Yang Shao4Youchun Ding5Wanneng Yang6Wen Qiao7Peng Song8National Key Laboratory of Crop Genetic Improvement, National Center of Plant Gene Research, Huazhong Agricultural University, Wuhan 430070, PR ChinaNational Key Laboratory of Crop Genetic Improvement, National Center of Plant Gene Research, Huazhong Agricultural University, Wuhan 430070, PR ChinaNational Key Laboratory of Crop Genetic Improvement, National Center of Plant Gene Research, Huazhong Agricultural University, Wuhan 430070, PR ChinaNational Key Laboratory of Crop Genetic Improvement, National Center of Plant Gene Research, Huazhong Agricultural University, Wuhan 430070, PR ChinaNational Key Laboratory of Crop Genetic Improvement, National Center of Plant Gene Research, Huazhong Agricultural University, Wuhan 430070, PR ChinaCollege of Engineering, Huazhong Agricultural University, Wuhan 430070, PR ChinaNational Key Laboratory of Crop Genetic Improvement, National Center of Plant Gene Research, Huazhong Agricultural University, Wuhan 430070, PR ChinaWuhan X-Agriculture Intelligent Technology Co., Ltd, Wuhan 430070, PR ChinaNational Key Laboratory of Crop Genetic Improvement, National Center of Plant Gene Research, Huazhong Agricultural University, Wuhan 430070, PR China; Corresponding author.High-throughput phenotyping collection technology is important in affecting the efficiency of crop breeding. This study introduces a novel autonomous navigation method for phenotyping robots that leverages ground-air collaboration to meet the demands of unmanned crop phenotypic data collection. The proposed method employs a UAV equipped with a Real-Time Kinematic (RTK) module for the construction of high-precision Field maps. It utilizes SegFormor-B0 semantic segmentation models to detect crop rows, and extracts key coordinate points of these rows, and generates navigation paths for the phenotyping robots by mapping these points to actual geographic coordinates. Furthermore, an adaptive controller based on the Pure Pursuit algorithm is proposed, which dynamically adjusts the steering angle of the phenotyping robot in real-time, according to the distance (d), angular deviation (α) and the lateral deviation (ey) between the robot's current position and its target position. This enables the robot to accurately trace paths in field environments. The results demonstrate that the mean absolute error (MAE) of the proposed method in extracting the centerline of potted plants area's rows is 2.83 cm, and the cropland's rows is 4.51 cm. The majority of global path tracking errors stay within 2 cm. In the potted plants area, 99.1 % of errors lie within this range, with a mean absolute error of 0.62 cm and a maximum error of 2.59 cm. In the cropland, 72.4 % of errors remain within this range, with a mean absolute error of 1.51 cm and a maximum error of 4.22 cm. Compared with traditional GNSS-based navigation methods and single vision methods, this method shows significant advantages in adapting to the dynamic growth of crops and complex field environments, which not only ensures that the phenotyping robot accurately travels along the crop rows during field operations to avoid damage to the crops, but also provides an efficient and accurate means of data acquisition for crop phenotyping.http://www.sciencedirect.com/science/article/pii/S2589721725000601Phenotyping robotGround-air collaborationSemantic segmentationPath planningAutonomous navigation
spellingShingle Zikang Zhang
Zhengda Li
Meng Yang
Jiale Cui
Yang Shao
Youchun Ding
Wanneng Yang
Wen Qiao
Peng Song
An autonomous navigation method for field phenotyping robot based on ground-air collaboration
Artificial Intelligence in Agriculture
Phenotyping robot
Ground-air collaboration
Semantic segmentation
Path planning
Autonomous navigation
title An autonomous navigation method for field phenotyping robot based on ground-air collaboration
title_full An autonomous navigation method for field phenotyping robot based on ground-air collaboration
title_fullStr An autonomous navigation method for field phenotyping robot based on ground-air collaboration
title_full_unstemmed An autonomous navigation method for field phenotyping robot based on ground-air collaboration
title_short An autonomous navigation method for field phenotyping robot based on ground-air collaboration
title_sort autonomous navigation method for field phenotyping robot based on ground air collaboration
topic Phenotyping robot
Ground-air collaboration
Semantic segmentation
Path planning
Autonomous navigation
url http://www.sciencedirect.com/science/article/pii/S2589721725000601
work_keys_str_mv AT zikangzhang anautonomousnavigationmethodforfieldphenotypingrobotbasedongroundaircollaboration
AT zhengdali anautonomousnavigationmethodforfieldphenotypingrobotbasedongroundaircollaboration
AT mengyang anautonomousnavigationmethodforfieldphenotypingrobotbasedongroundaircollaboration
AT jialecui anautonomousnavigationmethodforfieldphenotypingrobotbasedongroundaircollaboration
AT yangshao anautonomousnavigationmethodforfieldphenotypingrobotbasedongroundaircollaboration
AT youchunding anautonomousnavigationmethodforfieldphenotypingrobotbasedongroundaircollaboration
AT wannengyang anautonomousnavigationmethodforfieldphenotypingrobotbasedongroundaircollaboration
AT wenqiao anautonomousnavigationmethodforfieldphenotypingrobotbasedongroundaircollaboration
AT pengsong anautonomousnavigationmethodforfieldphenotypingrobotbasedongroundaircollaboration
AT zikangzhang autonomousnavigationmethodforfieldphenotypingrobotbasedongroundaircollaboration
AT zhengdali autonomousnavigationmethodforfieldphenotypingrobotbasedongroundaircollaboration
AT mengyang autonomousnavigationmethodforfieldphenotypingrobotbasedongroundaircollaboration
AT jialecui autonomousnavigationmethodforfieldphenotypingrobotbasedongroundaircollaboration
AT yangshao autonomousnavigationmethodforfieldphenotypingrobotbasedongroundaircollaboration
AT youchunding autonomousnavigationmethodforfieldphenotypingrobotbasedongroundaircollaboration
AT wannengyang autonomousnavigationmethodforfieldphenotypingrobotbasedongroundaircollaboration
AT wenqiao autonomousnavigationmethodforfieldphenotypingrobotbasedongroundaircollaboration
AT pengsong autonomousnavigationmethodforfieldphenotypingrobotbasedongroundaircollaboration