Advances in Object Detection and Localization Techniques for Fruit Harvesting Robots
Due to the short time, high labor intensity and high workload of fruit and vegetable harvesting, robotic harvesting instead of manual operations is the future. The accuracy of object detection and location is directly related to the picking efficiency, quality and speed of fruit-harvesting robots. B...
Saved in:
Main Authors: | , , , , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2025-01-01
|
Series: | Agronomy |
Subjects: | |
Online Access: | https://www.mdpi.com/2073-4395/15/1/145 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | Due to the short time, high labor intensity and high workload of fruit and vegetable harvesting, robotic harvesting instead of manual operations is the future. The accuracy of object detection and location is directly related to the picking efficiency, quality and speed of fruit-harvesting robots. Because of its low recognition accuracy, slow recognition speed and poor localization accuracy, the traditional algorithm cannot meet the requirements of automatic-harvesting robots. The increasingly evolving and powerful deep learning technology can effectively solve the above problems and has been widely used in the last few years. This work systematically summarizes and analyzes about 120 related literatures on the object detection and three-dimensional positioning algorithms of harvesting robots over the last 10 years, and reviews several significant methods. The difficulties and challenges faced by current fruit detection and localization algorithms are proposed from the aspects of the lack of large-scale high-quality datasets, the high complexity of the agricultural environment, etc. In response to the above challenges, corresponding solutions and future development trends are constructively proposed. Future research and technological development should first solve these current challenges using weakly supervised learning, efficient and lightweight model construction, multisensor fusion and so on. |
---|---|
ISSN: | 2073-4395 |