Loading…

Accurate detection and depth estimation of table grapes and peduncles for robot harvesting, combining monocular depth estimation and CNN methods

Precision agriculture is a growing field in the agricultural industry and it holds great potential in fruit and vegetable harvesting. In this work, we present a robust accurate method for the detection and localization of the peduncle of table grapes, with direct implementation in automatic grape ha...

Full description

Saved in:
Bibliographic Details
Published in:Computers and electronics in agriculture 2023-12, Vol.215, p.108362, Article 108362
Main Authors: Coll-Ribes, Gabriel, Torres-Rodríguez, Iván J., Grau, Antoni, Guerra, Edmundo, Sanfeliu, Alberto
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c352t-d4dfa27746ff066bf84c4930b5be350dc2591d47ad73df7a58208b1008e57b433
cites cdi_FETCH-LOGICAL-c352t-d4dfa27746ff066bf84c4930b5be350dc2591d47ad73df7a58208b1008e57b433
container_end_page
container_issue
container_start_page 108362
container_title Computers and electronics in agriculture
container_volume 215
creator Coll-Ribes, Gabriel
Torres-Rodríguez, Iván J.
Grau, Antoni
Guerra, Edmundo
Sanfeliu, Alberto
description Precision agriculture is a growing field in the agricultural industry and it holds great potential in fruit and vegetable harvesting. In this work, we present a robust accurate method for the detection and localization of the peduncle of table grapes, with direct implementation in automatic grape harvesting with robots. The bunch and peduncle detection methods presented in this work rely on a combination of instance segmentation and monocular depth estimation using Convolutional Neural Networks (CNN). Regarding depth estimation, we propose a combination of different depth techniques that allow precise localization of the peduncle using traditional stereo cameras, even with the particular complexity of grape peduncles. The methods proposed in this work have been tested on the WGISD (Embrapa Wine Grape Instance Segmentation) dataset, improving the results of state-of-the-art techniques. Furthermore, within the context of the EU project CANOPIES, the methods have also been tested on a dataset of 1,326 RGB-D images of table grapes, recorded at the Corsira Agricultural Cooperative Society (Aprilia, Italy), using a Realsense D435i camera located at the arm of a CANOPIES two-manipulator robot developed in the project. The detection results on the WGISD dataset show that the use of RGB-D information (mAP=0.949) leads to superior performance compared to the use of RGB data alone (mAP=0.891). This trend is also evident in the CANOPIES Grape Bunch and Peduncle dataset, where the mAP for RGB-D images (mAP=0.767) outperforms that of RGB data (mAP=0.725). Regarding depth estimation, our method achieves a mean squared error of 2.66 cm within a distance of 1 m in the CANOPIES dataset. •New bunch and peduncle detection method based on a combination of instance segmentation and monocular depth estimation.•New methods to estimate the peduncle depth.•Combination of depth estimation methods and direct stereo camera measurement.•Improvement of bunch and peduncle detection with respect to the state-of-art.•Tests in WGISD dataset and a new dataset created within the EU CANOPIES project.•Accurate detection of bunches and peduncles in real life experiments.
doi_str_mv 10.1016/j.compag.2023.108362
format article
fullrecord <record><control><sourceid>elsevier_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1016_j_compag_2023_108362</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0168169923007500</els_id><sourcerecordid>S0168169923007500</sourcerecordid><originalsourceid>FETCH-LOGICAL-c352t-d4dfa27746ff066bf84c4930b5be350dc2591d47ad73df7a58208b1008e57b433</originalsourceid><addsrcrecordid>eNp9kM1KAzEUhYMoWKtv4CIP4NT8zEwyG6EU_6DUja5DftspM5MhSQu-hY9sxroTXN17LvccDh8AtxgtMML1_X6hfT_K7YIgQvOJ05qcgRnmjBQMI3YOZvmNF7humktwFeMeZd1wNgNfS60PQSYLjU1Wp9YPUA4mqzHtoI2p7eXP0TuYpOos3AY52vjzNFpzGHSXlfMBBq98gjsZjpNt2N7B3Eq1Q15h7wevD50Mf4OnoNVmA3ubdt7Ea3DhZBftze-cg4-nx_fVS7F-e35dLdeFphVJhSmNk4SxsnYO1bVyvNRlQ5GqlKUVMppUDTYlk4ZR45isOEFcYYS4rZgqKZ2D8pSrg48xWCfGkCuFT4GRmKiKvThRFRNVcaKabQ8nm83djq0NIurWDtqaNmR8wvj2_4BvVVOFkg</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Accurate detection and depth estimation of table grapes and peduncles for robot harvesting, combining monocular depth estimation and CNN methods</title><source>ScienceDirect Freedom Collection 2022-2024</source><creator>Coll-Ribes, Gabriel ; Torres-Rodríguez, Iván J. ; Grau, Antoni ; Guerra, Edmundo ; Sanfeliu, Alberto</creator><creatorcontrib>Coll-Ribes, Gabriel ; Torres-Rodríguez, Iván J. ; Grau, Antoni ; Guerra, Edmundo ; Sanfeliu, Alberto</creatorcontrib><description>Precision agriculture is a growing field in the agricultural industry and it holds great potential in fruit and vegetable harvesting. In this work, we present a robust accurate method for the detection and localization of the peduncle of table grapes, with direct implementation in automatic grape harvesting with robots. The bunch and peduncle detection methods presented in this work rely on a combination of instance segmentation and monocular depth estimation using Convolutional Neural Networks (CNN). Regarding depth estimation, we propose a combination of different depth techniques that allow precise localization of the peduncle using traditional stereo cameras, even with the particular complexity of grape peduncles. The methods proposed in this work have been tested on the WGISD (Embrapa Wine Grape Instance Segmentation) dataset, improving the results of state-of-the-art techniques. Furthermore, within the context of the EU project CANOPIES, the methods have also been tested on a dataset of 1,326 RGB-D images of table grapes, recorded at the Corsira Agricultural Cooperative Society (Aprilia, Italy), using a Realsense D435i camera located at the arm of a CANOPIES two-manipulator robot developed in the project. The detection results on the WGISD dataset show that the use of RGB-D information (mAP=0.949) leads to superior performance compared to the use of RGB data alone (mAP=0.891). This trend is also evident in the CANOPIES Grape Bunch and Peduncle dataset, where the mAP for RGB-D images (mAP=0.767) outperforms that of RGB data (mAP=0.725). Regarding depth estimation, our method achieves a mean squared error of 2.66 cm within a distance of 1 m in the CANOPIES dataset. •New bunch and peduncle detection method based on a combination of instance segmentation and monocular depth estimation.•New methods to estimate the peduncle depth.•Combination of depth estimation methods and direct stereo camera measurement.•Improvement of bunch and peduncle detection with respect to the state-of-art.•Tests in WGISD dataset and a new dataset created within the EU CANOPIES project.•Accurate detection of bunches and peduncles in real life experiments.</description><identifier>ISSN: 0168-1699</identifier><identifier>EISSN: 1872-7107</identifier><identifier>DOI: 10.1016/j.compag.2023.108362</identifier><language>eng</language><publisher>Elsevier B.V</publisher><subject>Grape bunch and peduncle depth estimation ; Grape bunch and peduncle detection ; Image segmentation ; Monocular depth ; Robot harvesting</subject><ispartof>Computers and electronics in agriculture, 2023-12, Vol.215, p.108362, Article 108362</ispartof><rights>2023 The Author(s)</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c352t-d4dfa27746ff066bf84c4930b5be350dc2591d47ad73df7a58208b1008e57b433</citedby><cites>FETCH-LOGICAL-c352t-d4dfa27746ff066bf84c4930b5be350dc2591d47ad73df7a58208b1008e57b433</cites><orcidid>0000-0003-3868-9678</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27924,27925</link.rule.ids></links><search><creatorcontrib>Coll-Ribes, Gabriel</creatorcontrib><creatorcontrib>Torres-Rodríguez, Iván J.</creatorcontrib><creatorcontrib>Grau, Antoni</creatorcontrib><creatorcontrib>Guerra, Edmundo</creatorcontrib><creatorcontrib>Sanfeliu, Alberto</creatorcontrib><title>Accurate detection and depth estimation of table grapes and peduncles for robot harvesting, combining monocular depth estimation and CNN methods</title><title>Computers and electronics in agriculture</title><description>Precision agriculture is a growing field in the agricultural industry and it holds great potential in fruit and vegetable harvesting. In this work, we present a robust accurate method for the detection and localization of the peduncle of table grapes, with direct implementation in automatic grape harvesting with robots. The bunch and peduncle detection methods presented in this work rely on a combination of instance segmentation and monocular depth estimation using Convolutional Neural Networks (CNN). Regarding depth estimation, we propose a combination of different depth techniques that allow precise localization of the peduncle using traditional stereo cameras, even with the particular complexity of grape peduncles. The methods proposed in this work have been tested on the WGISD (Embrapa Wine Grape Instance Segmentation) dataset, improving the results of state-of-the-art techniques. Furthermore, within the context of the EU project CANOPIES, the methods have also been tested on a dataset of 1,326 RGB-D images of table grapes, recorded at the Corsira Agricultural Cooperative Society (Aprilia, Italy), using a Realsense D435i camera located at the arm of a CANOPIES two-manipulator robot developed in the project. The detection results on the WGISD dataset show that the use of RGB-D information (mAP=0.949) leads to superior performance compared to the use of RGB data alone (mAP=0.891). This trend is also evident in the CANOPIES Grape Bunch and Peduncle dataset, where the mAP for RGB-D images (mAP=0.767) outperforms that of RGB data (mAP=0.725). Regarding depth estimation, our method achieves a mean squared error of 2.66 cm within a distance of 1 m in the CANOPIES dataset. •New bunch and peduncle detection method based on a combination of instance segmentation and monocular depth estimation.•New methods to estimate the peduncle depth.•Combination of depth estimation methods and direct stereo camera measurement.•Improvement of bunch and peduncle detection with respect to the state-of-art.•Tests in WGISD dataset and a new dataset created within the EU CANOPIES project.•Accurate detection of bunches and peduncles in real life experiments.</description><subject>Grape bunch and peduncle depth estimation</subject><subject>Grape bunch and peduncle detection</subject><subject>Image segmentation</subject><subject>Monocular depth</subject><subject>Robot harvesting</subject><issn>0168-1699</issn><issn>1872-7107</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNp9kM1KAzEUhYMoWKtv4CIP4NT8zEwyG6EU_6DUja5DftspM5MhSQu-hY9sxroTXN17LvccDh8AtxgtMML1_X6hfT_K7YIgQvOJ05qcgRnmjBQMI3YOZvmNF7humktwFeMeZd1wNgNfS60PQSYLjU1Wp9YPUA4mqzHtoI2p7eXP0TuYpOos3AY52vjzNFpzGHSXlfMBBq98gjsZjpNt2N7B3Eq1Q15h7wevD50Mf4OnoNVmA3ubdt7Ea3DhZBftze-cg4-nx_fVS7F-e35dLdeFphVJhSmNk4SxsnYO1bVyvNRlQ5GqlKUVMppUDTYlk4ZR45isOEFcYYS4rZgqKZ2D8pSrg48xWCfGkCuFT4GRmKiKvThRFRNVcaKabQ8nm83djq0NIurWDtqaNmR8wvj2_4BvVVOFkg</recordid><startdate>202312</startdate><enddate>202312</enddate><creator>Coll-Ribes, Gabriel</creator><creator>Torres-Rodríguez, Iván J.</creator><creator>Grau, Antoni</creator><creator>Guerra, Edmundo</creator><creator>Sanfeliu, Alberto</creator><general>Elsevier B.V</general><scope>6I.</scope><scope>AAFTH</scope><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/0000-0003-3868-9678</orcidid></search><sort><creationdate>202312</creationdate><title>Accurate detection and depth estimation of table grapes and peduncles for robot harvesting, combining monocular depth estimation and CNN methods</title><author>Coll-Ribes, Gabriel ; Torres-Rodríguez, Iván J. ; Grau, Antoni ; Guerra, Edmundo ; Sanfeliu, Alberto</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c352t-d4dfa27746ff066bf84c4930b5be350dc2591d47ad73df7a58208b1008e57b433</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Grape bunch and peduncle depth estimation</topic><topic>Grape bunch and peduncle detection</topic><topic>Image segmentation</topic><topic>Monocular depth</topic><topic>Robot harvesting</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Coll-Ribes, Gabriel</creatorcontrib><creatorcontrib>Torres-Rodríguez, Iván J.</creatorcontrib><creatorcontrib>Grau, Antoni</creatorcontrib><creatorcontrib>Guerra, Edmundo</creatorcontrib><creatorcontrib>Sanfeliu, Alberto</creatorcontrib><collection>ScienceDirect Open Access Titles</collection><collection>Elsevier:ScienceDirect:Open Access</collection><collection>CrossRef</collection><jtitle>Computers and electronics in agriculture</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Coll-Ribes, Gabriel</au><au>Torres-Rodríguez, Iván J.</au><au>Grau, Antoni</au><au>Guerra, Edmundo</au><au>Sanfeliu, Alberto</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Accurate detection and depth estimation of table grapes and peduncles for robot harvesting, combining monocular depth estimation and CNN methods</atitle><jtitle>Computers and electronics in agriculture</jtitle><date>2023-12</date><risdate>2023</risdate><volume>215</volume><spage>108362</spage><pages>108362-</pages><artnum>108362</artnum><issn>0168-1699</issn><eissn>1872-7107</eissn><abstract>Precision agriculture is a growing field in the agricultural industry and it holds great potential in fruit and vegetable harvesting. In this work, we present a robust accurate method for the detection and localization of the peduncle of table grapes, with direct implementation in automatic grape harvesting with robots. The bunch and peduncle detection methods presented in this work rely on a combination of instance segmentation and monocular depth estimation using Convolutional Neural Networks (CNN). Regarding depth estimation, we propose a combination of different depth techniques that allow precise localization of the peduncle using traditional stereo cameras, even with the particular complexity of grape peduncles. The methods proposed in this work have been tested on the WGISD (Embrapa Wine Grape Instance Segmentation) dataset, improving the results of state-of-the-art techniques. Furthermore, within the context of the EU project CANOPIES, the methods have also been tested on a dataset of 1,326 RGB-D images of table grapes, recorded at the Corsira Agricultural Cooperative Society (Aprilia, Italy), using a Realsense D435i camera located at the arm of a CANOPIES two-manipulator robot developed in the project. The detection results on the WGISD dataset show that the use of RGB-D information (mAP=0.949) leads to superior performance compared to the use of RGB data alone (mAP=0.891). This trend is also evident in the CANOPIES Grape Bunch and Peduncle dataset, where the mAP for RGB-D images (mAP=0.767) outperforms that of RGB data (mAP=0.725). Regarding depth estimation, our method achieves a mean squared error of 2.66 cm within a distance of 1 m in the CANOPIES dataset. •New bunch and peduncle detection method based on a combination of instance segmentation and monocular depth estimation.•New methods to estimate the peduncle depth.•Combination of depth estimation methods and direct stereo camera measurement.•Improvement of bunch and peduncle detection with respect to the state-of-art.•Tests in WGISD dataset and a new dataset created within the EU CANOPIES project.•Accurate detection of bunches and peduncles in real life experiments.</abstract><pub>Elsevier B.V</pub><doi>10.1016/j.compag.2023.108362</doi><orcidid>https://orcid.org/0000-0003-3868-9678</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0168-1699
ispartof Computers and electronics in agriculture, 2023-12, Vol.215, p.108362, Article 108362
issn 0168-1699
1872-7107
language eng
recordid cdi_crossref_primary_10_1016_j_compag_2023_108362
source ScienceDirect Freedom Collection 2022-2024
subjects Grape bunch and peduncle depth estimation
Grape bunch and peduncle detection
Image segmentation
Monocular depth
Robot harvesting
title Accurate detection and depth estimation of table grapes and peduncles for robot harvesting, combining monocular depth estimation and CNN methods
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-02T21%3A56%3A25IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-elsevier_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Accurate%20detection%20and%20depth%20estimation%20of%20table%20grapes%20and%20peduncles%20for%20robot%20harvesting,%20combining%20monocular%20depth%20estimation%20and%20CNN%20methods&rft.jtitle=Computers%20and%20electronics%20in%20agriculture&rft.au=Coll-Ribes,%20Gabriel&rft.date=2023-12&rft.volume=215&rft.spage=108362&rft.pages=108362-&rft.artnum=108362&rft.issn=0168-1699&rft.eissn=1872-7107&rft_id=info:doi/10.1016/j.compag.2023.108362&rft_dat=%3Celsevier_cross%3ES0168169923007500%3C/elsevier_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c352t-d4dfa27746ff066bf84c4930b5be350dc2591d47ad73df7a58208b1008e57b433%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true