Dataset and Performance Comparison of Deep Learning Architectures for
Plum Detection and Robotic Harvesting
- URL: http://arxiv.org/abs/2105.03832v1
- Date: Sun, 9 May 2021 04:18:58 GMT
- Title: Dataset and Performance Comparison of Deep Learning Architectures for
Plum Detection and Robotic Harvesting
- Authors: Jasper Brown, Salah Sukkarieh
- Abstract summary: Two new datasets are gathered during day and night operation of an actual robotic plum harvesting system.
A range of current generation deep learning object detectors are benchmarked against these.
Two methods for fusing depth and image information are tested for their impact on detector performance.
- Score: 3.7692411550925673
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Many automated operations in agriculture, such as weeding and plant counting,
require robust and accurate object detectors. Robotic fruit harvesting is one
of these, and is an important technology to address the increasing labour
shortages and uncertainty suffered by tree crop growers. An eye-in-hand sensing
setup is commonly used in harvesting systems and provides benefits to sensing
accuracy and flexibility. However, as the hand and camera move from viewing the
entire trellis to picking a specific fruit, large changes in lighting, colour,
obscuration and exposure occur. Object detection algorithms used in harvesting
should be robust to these challenges, but few datasets for assessing this
currently exist. In this work, two new datasets are gathered during day and
night operation of an actual robotic plum harvesting system. A range of current
generation deep learning object detectors are benchmarked against these.
Additionally, two methods for fusing depth and image information are tested for
their impact on detector performance. Significant differences between day and
night accuracy of different detectors is found, transfer learning is identified
as essential in all cases, and depth information fusion is assessed as only
marginally effective. The dataset and benchmark models are made available
online.
Related papers
- Assaying on the Robustness of Zero-Shot Machine-Generated Text Detectors [57.7003399760813]
We explore advanced Large Language Models (LLMs) and their specialized variants, contributing to this field in several ways.
We uncover a significant correlation between topics and detection performance.
These investigations shed light on the adaptability and robustness of these detection methods across diverse topics.
arXiv Detail & Related papers (2023-12-20T10:53:53Z) - Exploring the Effectiveness of Dataset Synthesis: An application of
Apple Detection in Orchards [68.95806641664713]
We explore the usability of Stable Diffusion 2.1-base for generating synthetic datasets of apple trees for object detection.
We train a YOLOv5m object detection model to predict apples in a real-world apple detection dataset.
Results demonstrate that the model trained on generated data is slightly underperforming compared to a baseline model trained on real-world images.
arXiv Detail & Related papers (2023-06-20T09:46:01Z) - Target-aware Dual Adversarial Learning and a Multi-scenario
Multi-Modality Benchmark to Fuse Infrared and Visible for Object Detection [65.30079184700755]
This study addresses the issue of fusing infrared and visible images that appear differently for object detection.
Previous approaches discover commons underlying the two modalities and fuse upon the common space either by iterative optimization or deep networks.
This paper proposes a bilevel optimization formulation for the joint problem of fusion and detection, and then unrolls to a target-aware Dual Adversarial Learning (TarDAL) network for fusion and a commonly used detection network.
arXiv Detail & Related papers (2022-03-30T11:44:56Z) - Agricultural Plant Cataloging and Establishment of a Data Framework from
UAV-based Crop Images by Computer Vision [4.0382342610484425]
We present a hands-on workflow for the automatized temporal and spatial identification and individualization of crop images from UAVs.
The presented approach improves analysis and interpretation of UAV data in agriculture significantly.
arXiv Detail & Related papers (2022-01-08T21:14:07Z) - MetaGraspNet: A Large-Scale Benchmark Dataset for Vision-driven Robotic
Grasping via Physics-based Metaverse Synthesis [78.26022688167133]
We present a large-scale benchmark dataset for vision-driven robotic grasping via physics-based metaverse synthesis.
The proposed dataset contains 100,000 images and 25 different object types.
We also propose a new layout-weighted performance metric alongside the dataset for evaluating object detection and segmentation performance.
arXiv Detail & Related papers (2021-12-29T17:23:24Z) - Potato Crop Stress Identification in Aerial Images using Deep
Learning-based Object Detection [60.83360138070649]
The paper presents an approach for analyzing aerial images of a potato crop using deep neural networks.
The main objective is to demonstrate automated spatial recognition of a healthy versus stressed crop at a plant level.
Experimental validation demonstrated the ability for distinguishing healthy and stressed plants in field images, achieving an average Dice coefficient of 0.74.
arXiv Detail & Related papers (2021-06-14T21:57:40Z) - REGRAD: A Large-Scale Relational Grasp Dataset for Safe and
Object-Specific Robotic Grasping in Clutter [52.117388513480435]
We present a new dataset named regrad to sustain the modeling of relationships among objects and grasps.
Our dataset is collected in both forms of 2D images and 3D point clouds.
Users are free to import their own object models for the generation of as many data as they want.
arXiv Detail & Related papers (2021-04-29T05:31:21Z) - On the Role of Sensor Fusion for Object Detection in Future Vehicular
Networks [25.838878314196375]
We evaluate how using a combination of different sensors affects the detection of the environment in which the vehicles move and operate.
The final objective is to identify the optimal setup that would minimize the amount of data to be distributed over the channel.
arXiv Detail & Related papers (2021-04-23T18:58:37Z) - A Robust Illumination-Invariant Camera System for Agricultural
Applications [7.349727826230863]
Object detection and semantic segmentation are two of the most widely adopted deep learning algorithms in agricultural applications.
We present a high throughput robust active lighting-based camera system that generates consistent images in all lighting conditions.
On average, deep nets for object detection trained on consistent data required nearly four times less data to achieve similar level of accuracy.
arXiv Detail & Related papers (2021-01-06T18:50:53Z) - Few-Shot Object Detection in Real Life: Case Study on Auto-Harvest [26.119977579845614]
Confinement during COVID-19 has caused serious effects on agriculture all over the world.
Within the auto-harvest system, robust few-shot object detection model is one of the bottlenecks.
We present a novel cucumber dataset and propose two data augmentation strategies that help to bridge the context-gap.
arXiv Detail & Related papers (2020-11-05T09:24:33Z) - Comparison of object detection methods for crop damage assessment using
deep learning [0.0]
The goal of this study was a proof-of-concept to detect damaged crop areas from aerial imagery using computer vision and deep learning techniques.
An unmanned aerial system (UAS) equipped with a RGB camera was used for image acquisition.
Three popular object detectors (Faster R-CNN, YOLOv2, and RetinaNet) were assessed for their ability to detect damaged regions in a field.
YOLOv2 and RetinaNet were able to detect crop damage across multiple late-season growth stages. Faster R-CNN was not successful as the other two advanced detectors.
arXiv Detail & Related papers (2019-12-31T06:54:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.