How to rewrite the stars: Mapping your orchard over time through constellations of fruits
- URL: http://arxiv.org/abs/2602.04722v1
- Date: Wed, 04 Feb 2026 16:31:32 GMT
- Title: How to rewrite the stars: Mapping your orchard over time through constellations of fruits
- Authors: Gonçalo P. Matos, Carlos Santiago, João P. Costeira, Ricardo L. Saldanha, Ernesto M. Morgado,
- Abstract summary: We propose a new paradigm to tackle the problem of matching fruits across videos.<n>The proposed method can be successfully used to match fruits across videos and through time.<n>It can also be used to build an orchard map and later use it to locate the camera pose in 6DoF.
- Score: 8.064400168497373
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Following crop growth through the vegetative cycle allows farmers to predict fruit setting and yield in early stages, but it is a laborious and non-scalable task if performed by a human who has to manually measure fruit sizes with a caliper or dendrometers. In recent years, computer vision has been used to automate several tasks in precision agriculture, such as detecting and counting fruits, and estimating their size. However, the fundamental problem of matching the exact same fruits from one video, collected on a given date, to the fruits visible in another video, collected on a later date, which is needed to track fruits' growth through time, remains to be solved. Few attempts were made, but they either assume that the camera always starts from the same known position and that there are sufficiently distinct features to match, or they used other sources of data like GPS. Here we propose a new paradigm to tackle this problem, based on constellations of 3D centroids, and introduce a descriptor for very sparse 3D point clouds that can be used to match fruits across videos. Matching constellations instead of individual fruits is key to deal with non-rigidity, occlusions and challenging imagery with few distinct visual features to track. The results show that the proposed method can be successfully used to match fruits across videos and through time, and also to build an orchard map and later use it to locate the camera pose in 6DoF, thus providing a method for autonomous navigation of robots in the orchard and for selective fruit picking, for example.
Related papers
- Cherry Yield Forecast: Harvest Prediction for Individual Sweet Cherry Trees [0.0]
This paper is part of a publication series from the For5G project that has the goal of creating digital twins of sweet cherry trees.<n>It is concluded that accurate yield prediction for sweet cherry trees is possible when objects are manually counted and that automated features extraction with similar accuracy remains an open problem yet to be solved.
arXiv Detail & Related papers (2025-03-26T10:50:02Z) - FG$^2$: Fine-Grained Cross-View Localization by Fine-Grained Feature Matching [69.81167130510333]
We propose a novel fine-grained cross-view localization method that estimates the 3 Degrees of Freedom pose of a ground-level image in an aerial image of the surroundings.<n>The pose is estimated by aligning a point plane generated from the ground image with a point plane sampled from the aerial image.<n>Compared to the previous state-of-the-art, our method reduces the mean localization error by 28% on the VIGOR cross-area test set.
arXiv Detail & Related papers (2025-03-24T14:34:20Z) - Horticultural Temporal Fruit Monitoring via 3D Instance Segmentation and Re-Identification using Colored Point Clouds [29.23207854514898]
We propose a novel method for fruit instance segmentation and re-identification on 3D terrestrial point clouds collected over time.<n>Our approach directly operates on dense colored point clouds, capturing fine-grained 3D spatial detail.<n>We evaluate our approach on real-world datasets of strawberries and apples, demonstrating that it outperforms existing methods in both instance segmentation and temporal re-identification.
arXiv Detail & Related papers (2024-11-12T13:53:22Z) - A pipeline for multiple orange detection and tracking with 3-D fruit
relocalization and neural-net based yield regression in commercial citrus
orchards [0.0]
We propose a non-invasive alternative that utilizes fruit counting from videos, implemented as a pipeline.
To handle occluded and re-appeared fruit, we introduce a relocalization component that employs 3-D estimation of fruit locations.
By ensuring that at least 30% of the fruit is accurately detected, tracked, and counted, our yield regressor achieves an impressive coefficient of determination of 0.85.
arXiv Detail & Related papers (2023-12-27T21:22:43Z) - Panoptic Mapping with Fruit Completion and Pose Estimation for
Horticultural Robots [33.21287030243106]
Monitoring plants and fruits at high resolution play a key role in the future of agriculture.
Accurate 3D information can pave the way to a diverse number of robotic applications in agriculture ranging from autonomous harvesting to precise yield estimation.
We address the problem of jointly estimating complete 3D shapes of fruit and their pose in a 3D multi-resolution map built by a mobile robot.
arXiv Detail & Related papers (2023-03-15T20:41:24Z) - Fruit Ripeness Classification: a Survey [59.11160990637616]
Many automatic methods have been proposed that employ a variety of feature descriptors for the food item to be graded.
Machine learning and deep learning techniques dominate the top-performing methods.
Deep learning can operate on raw data and thus relieve the users from having to compute complex engineered features.
arXiv Detail & Related papers (2022-12-29T19:32:20Z) - End-to-end deep learning for directly estimating grape yield from
ground-based imagery [53.086864957064876]
This study demonstrates the application of proximal imaging combined with deep learning for yield estimation in vineyards.
Three model architectures were tested: object detection, CNN regression, and transformer models.
The study showed the applicability of proximal imaging and deep learning for prediction of grapevine yield on a large scale.
arXiv Detail & Related papers (2022-08-04T01:34:46Z) - A methodology for detection and localization of fruits in apples
orchards from aerial images [0.0]
This work presents a methodology for automated fruit counting employing aerial-images.
It includes algorithms based on multiple view geometry to perform fruits tracking.
Preliminary assessments show correlations above 0.8 between fruit counting and true yield for apples.
arXiv Detail & Related papers (2021-10-24T01:57:52Z) - MetaPose: Fast 3D Pose from Multiple Views without 3D Supervision [72.5863451123577]
We show how to train a neural model that can perform accurate 3D pose and camera estimation.
Our method outperforms both classical bundle adjustment and weakly-supervised monocular 3D baselines.
arXiv Detail & Related papers (2021-08-10T18:39:56Z) - Wide-Area Crowd Counting: Multi-View Fusion Networks for Counting in
Large Scenes [50.744452135300115]
We propose a deep neural network framework for multi-view crowd counting.
Our methods achieve state-of-the-art results compared to other multi-view counting baselines.
arXiv Detail & Related papers (2020-12-02T03:20:30Z) - Shape and Viewpoint without Keypoints [63.26977130704171]
We present a learning framework that learns to recover the 3D shape, pose and texture from a single image.
We trained on an image collection without any ground truth 3D shape, multi-view, camera viewpoints or keypoint supervision.
We obtain state-of-the-art camera prediction results and show that we can learn to predict diverse shapes and textures across objects.
arXiv Detail & Related papers (2020-07-21T17:58:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.