Leaving the Lines Behind: Vision-Based Crop Row Exit for Agricultural
Robot Navigation
- URL: http://arxiv.org/abs/2306.05869v1
- Date: Fri, 9 Jun 2023 13:02:31 GMT
- Title: Leaving the Lines Behind: Vision-Based Crop Row Exit for Agricultural
Robot Navigation
- Authors: Rajitha de Silva, Grzegorz Cielniak, Junfeng Gao
- Abstract summary: The proposed method could reach the end of the crop row and then navigate into the headland completely leaving behind the crop row with an error margin of 50 cm.
- Score: 6.088167023055281
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Usage of purely vision based solutions for row switching is not well explored
in existing vision based crop row navigation frameworks. This method only uses
RGB images for local feature matching based visual feedback to exit crop row.
Depth images were used at crop row end to estimate the navigation distance
within headland. The algorithm was tested on diverse headland areas with soil
and vegetation. The proposed method could reach the end of the crop row and
then navigate into the headland completely leaving behind the crop row with an
error margin of 50 cm.
Related papers
- Breaking the Frame: Visual Place Recognition by Overlap Prediction [53.17564423756082]
We propose a novel visual place recognition approach based on overlap prediction, called VOP.
VOP proceeds co-visible image sections by obtaining patch-level embeddings using a Vision Transformer backbone.
Our approach uses a voting mechanism to assess overlap scores for potential database images.
arXiv Detail & Related papers (2024-06-23T20:00:20Z) - A Vision-Based Navigation System for Arable Fields [7.338061223686544]
Vision-based navigation systems in arable fields are an underexplored area in agricultural robot navigation.
Current solutions are often crop-specific and aimed to address limited individual conditions such as illumination or weed density.
This paper proposes a suite of deep learning-based perception algorithms using affordable vision sensors for vision-based navigation in arable fields.
arXiv Detail & Related papers (2023-09-21T12:01:59Z) - Geometric-aware Pretraining for Vision-centric 3D Object Detection [77.7979088689944]
We propose a novel geometric-aware pretraining framework called GAPretrain.
GAPretrain serves as a plug-and-play solution that can be flexibly applied to multiple state-of-the-art detectors.
We achieve 46.2 mAP and 55.5 NDS on the nuScenes val set using the BEVFormer method, with a gain of 2.7 and 2.1 points, respectively.
arXiv Detail & Related papers (2023-04-06T14:33:05Z) - Vision-based Vineyard Navigation Solution with Automatic Annotation [2.6013566739979463]
We introduce a vision-based autonomous navigation framework for agriculture robots in trellised cropping systems such as vineyards.
We propose a novel learning-based method to estimate the path traversibility heatmap directly from an RGB-D image.
A trained path detection model was used to develop a full navigation framework consisting of row tracking and row switching modules.
arXiv Detail & Related papers (2023-03-25T03:37:17Z) - Vision based Crop Row Navigation under Varying Field Conditions in
Arable Fields [6.088167023055281]
We present a dataset for crop row detection with 11 field variations from Sugar Beet and Maize crops.
We also present a novel crop row detection algorithm for visual servoing in crop row fields.
arXiv Detail & Related papers (2022-09-28T11:23:34Z) - Deep learning-based Crop Row Detection for Infield Navigation of
Agri-Robots [10.386591972977207]
This paper presents a robust crop row detection algorithm that withstands field variations using inexpensive cameras.
A dataset of sugar beet images was created representing 11 field variations comprised of multiple grow stages, light levels, varying weed densities, curved crop rows and discontinuous crop rows.
Our algorithm demonstrated robust vision-based crop row detection in challenging field conditions outperforming the baseline.
arXiv Detail & Related papers (2022-09-09T12:47:24Z) - Beyond Visual Field of View: Perceiving 3D Environment with Echoes and
Vision [51.385731364529306]
This paper focuses on perceiving and navigating 3D environments using echoes and RGB image.
In particular, we perform depth estimation by fusing RGB image with echoes, received from multiple orientations.
We show that the echoes provide holistic and in-expensive information about the 3D structures complementing the RGB image.
arXiv Detail & Related papers (2022-07-03T22:31:47Z) - Beyond Cross-view Image Retrieval: Highly Accurate Vehicle Localization
Using Satellite Image [91.29546868637911]
This paper addresses the problem of vehicle-mounted camera localization by matching a ground-level image with an overhead-view satellite map.
The key idea is to formulate the task as pose estimation and solve it by neural-net based optimization.
Experiments on standard autonomous vehicle localization datasets have confirmed the superiority of the proposed method.
arXiv Detail & Related papers (2022-04-10T19:16:58Z) - Towards agricultural autonomy: crop row detection under varying field
conditions using deep learning [4.252146169134215]
This paper presents a novel metric to evaluate the robustness of deep learning based semantic segmentation approaches for crop row detection.
A dataset with ten main categories encountered under various field conditions was used for testing.
The effect on these conditions on the angular accuracy of crop row detection was compared.
arXiv Detail & Related papers (2021-09-16T23:12:08Z) - Refer-it-in-RGBD: A Bottom-up Approach for 3D Visual Grounding in RGBD
Images [69.5662419067878]
Grounding referring expressions in RGBD image has been an emerging field.
We present a novel task of 3D visual grounding in single-view RGBD image where the referred objects are often only partially scanned due to occlusion.
Our approach first fuses the language and the visual features at the bottom level to generate a heatmap that localizes the relevant regions in the RGBD image.
Then our approach conducts an adaptive feature learning based on the heatmap and performs the object-level matching with another visio-linguistic fusion to finally ground the referred object.
arXiv Detail & Related papers (2021-03-14T11:18:50Z) - Single Image Depth Estimation Trained via Depth from Defocus Cues [105.67073923825842]
Estimating depth from a single RGB image is a fundamental task in computer vision.
In this work, we rely, instead of different views, on depth from focus cues.
We present results that are on par with supervised methods on KITTI and Make3D datasets and outperform unsupervised learning approaches.
arXiv Detail & Related papers (2020-01-14T20:22:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.