An RGB-D Image Dataset for Lychee Detection and Maturity Classification for Robotic Harvesting
- URL: http://arxiv.org/abs/2510.16800v1
- Date: Sun, 19 Oct 2025 11:47:20 GMT
- Title: An RGB-D Image Dataset for Lychee Detection and Maturity Classification for Robotic Harvesting
- Authors: Zhenpeng Zhang, Yi Wang, Shanglei Chai, Yingying Liu, Zekai Xie, Wenhao Huang, Pengyu Li, Zipei Luo, Dajiang Lu, Yibin Tian,
- Abstract summary: Lychee is a high-value subtropical fruit.<n>The adoption of vision-based harvesting robots can significantly improve productivity.<n>There are currently no consistently and comprehensively annotated open-source lychee datasets.
- Score: 20.028328464541456
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Lychee is a high-value subtropical fruit. The adoption of vision-based harvesting robots can significantly improve productivity while reduce reliance on labor. High-quality data are essential for developing such harvesting robots. However, there are currently no consistently and comprehensively annotated open-source lychee datasets featuring fruits in natural growing environments. To address this, we constructed a dataset to facilitate lychee detection and maturity classification. Color (RGB) images were acquired under diverse weather conditions, and at different times of the day, across multiple lychee varieties, such as Nuomici, Feizixiao, Heiye, and Huaizhi. The dataset encompasses three different ripeness stages and contains 11,414 images, consisting of 878 raw RGB images, 8,780 augmented RGB images, and 1,756 depth images. The images are annotated with 9,658 pairs of lables for lychee detection and maturity classification. To improve annotation consistency, three individuals independently labeled the data, and their results were then aggregated and verified by a fourth reviewer. Detailed statistical analyses were done to examine the dataset. Finally, we performed experiments using three representative deep learning models to evaluate the dataset. It is publicly available for academic
Related papers
- Healthy Harvests: A Comparative Look at Guava Disease Classification Using InceptionV3 [0.0]
This dataset contains 473 original images of Guava.<n>The augmented dataset consists of 3784 images using advanced preprocessing techniques.<n>The InceptionV3 model achieved the impressive accuracy of 98.15%, and ResNet50got 94.46% accuracy.
arXiv Detail & Related papers (2026-02-11T15:59:49Z) - A Generative Data Framework with Authentic Supervision for Underwater Image Restoration and Enhancement [51.382274157144714]
We develop a generative data framework based on unpaired image-to-image translation.<n>The framework constructs synthetic datasets with precise ground-truth labels.<n>Experiments show that models trained on our synthetic data achieve comparable or superior color restoration and generalization performance to those trained on existing benchmarks.
arXiv Detail & Related papers (2025-11-18T14:20:17Z) - Bridging the Gap Between Ideal and Real-world Evaluation: Benchmarking AI-Generated Image Detection in Challenging Scenarios [54.07895223545793]
This paper introduces the Real-World Robustness dataset (RRDataset) for comprehensive evaluation of detection models across three dimensions.<n>RRDataset includes high-quality images from seven major scenarios.<n>We benchmarked 17 detectors and 10 vision-language models (VLMs) on RRDataset and conducted a large-scale human study.
arXiv Detail & Related papers (2025-09-11T06:15:52Z) - Tomato Multi-Angle Multi-Pose Dataset for Fine-Grained Phenotyping [10.807010511060042]
TomatoMAP is a comprehensive dataset for Solanum lycopersicum.<n>Our dataset contains 64,464 RGB images that capture 12 different plant poses from four camera elevation angles.<n>We provide 3,616 high-resolution image subset with pixel-wise semantic and instance segmentation annotations for fine-grained phenotyping.
arXiv Detail & Related papers (2025-07-15T12:56:13Z) - Attention-Based Ensemble Learning for Crop Classification Using Landsat 8-9 Fusion [3.342115421418929]
The study focuses on crop cover identification for irrigated regions of Central Punjab.<n>Data collection was executed in two stages: the first involved identifying and geocoding six target crops.<n>The second stage involved acquiring Landsat 8-9 imagery for each geocoded field to construct a labelled dataset.
arXiv Detail & Related papers (2025-06-23T06:09:39Z) - Merging synthetic and real embryo data for advanced AI predictions [69.07284335967019]
We train two generative models using two datasets-one we created and made publicly available, and one existing public dataset-to generate synthetic embryo images at various cell stages.<n>These were combined with real images to train classification models for embryo cell stage prediction.<n>Our results demonstrate that incorporating synthetic images alongside real data improved classification performance, with the model achieving 97% accuracy compared to 94.5% when trained solely on real data.
arXiv Detail & Related papers (2024-12-02T08:24:49Z) - SatSynth: Augmenting Image-Mask Pairs through Diffusion Models for Aerial Semantic Segmentation [69.42764583465508]
We explore the potential of generative image diffusion to address the scarcity of annotated data in earth observation tasks.
To the best of our knowledge, we are the first to generate both images and corresponding masks for satellite segmentation.
arXiv Detail & Related papers (2024-03-25T10:30:22Z) - Learning from Synthetic Data for Visual Grounding [55.21937116752679]
We show that SynGround can improve the localization capabilities of off-the-shelf vision-and-language models.<n>Data generated with SynGround improves the pointing game accuracy of a pretrained ALBEF and BLIP models by 4.81% and 17.11% absolute percentage points, respectively.
arXiv Detail & Related papers (2024-03-20T17:59:43Z) - Detection of Spider Mites on Labrador Beans through Machine Learning
Approaches Using Custom Datasets [0.0]
This study proposes a visual machine learning approach for plant disease detection, harnessing RGB and NIR data collected in real-world conditions through a JAI FS-1600D-10GE camera to build an RGBN dataset.
A two-stage early plant disease detection model with YOLOv8 and a sequential CNN was used to train on a dataset with partial labels, which showed a 3.6% increase in mAP compared to a single-stage end-to-end segmentation model.
An average of 6.25% validation accuracy increase is found using RGBN in classification compared to RGB using ResNet15 and the sequential CNN models.
arXiv Detail & Related papers (2024-02-12T18:57:06Z) - Stable Diffusion for Data Augmentation in COCO and Weed Datasets [5.81198182644659]
Generative models have increasingly impacted various tasks, from computer vision to interior design and beyond. Stable Diffusion, a powerful diffusion model, enables the creation of high-resolution images with intricate details from text prompts or reference images.<n>This study explores the effectiveness of Stable Diffusion by evaluating seven common categories and three widespread weed species.<n> Promising results were achieved for certain classes, demonstrating the potential of Stable Diffusion in enhancing image-sparse datasets.
arXiv Detail & Related papers (2023-12-07T02:23:32Z) - End-to-end deep learning for directly estimating grape yield from
ground-based imagery [53.086864957064876]
This study demonstrates the application of proximal imaging combined with deep learning for yield estimation in vineyards.
Three model architectures were tested: object detection, CNN regression, and transformer models.
The study showed the applicability of proximal imaging and deep learning for prediction of grapevine yield on a large scale.
arXiv Detail & Related papers (2022-08-04T01:34:46Z) - Classification of Seeds using Domain Randomization on Self-Supervised
Learning Frameworks [0.0]
Key bottleneck is the need for an extensive amount of labelled data to train the convolutional neural networks (CNN)
The work leverages the concepts of Contrastive Learning and Domain Randomi-zation in order to achieve the same.
The use of synthetic images generated from a representational sample crop of real-world images alleviates the need for a large volume of test subjects.
arXiv Detail & Related papers (2021-03-29T12:50:06Z) - Random Network Distillation as a Diversity Metric for Both Image and
Text Generation [62.13444904851029]
We develop a new diversity metric that can be applied to data, both synthetic and natural, of any type.
We validate and deploy this metric on both images and text.
arXiv Detail & Related papers (2020-10-13T22:03:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.