Efficient automated U-Net based tree crown delineation using UAV
multi-spectral imagery on embedded devices
- URL: http://arxiv.org/abs/2107.07826v1
- Date: Fri, 16 Jul 2021 11:17:36 GMT
- Title: Efficient automated U-Net based tree crown delineation using UAV
multi-spectral imagery on embedded devices
- Authors: Kostas Blekos, Stavros Nousias, Aris S Lalos
- Abstract summary: Delineation approaches provide significant benefits to various domains, including agriculture, environmental and natural disasters monitoring.
Deep learning has transformed computer vision and dramatically improved machine translation, though it requires massive dataset for training and resources for inference.
We propose a U-Net based tree delineation method, which is effectively trained using multi-spectral imagery but can then delineate single-spectrum images.
- Score: 2.7393821783237184
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Delineation approaches provide significant benefits to various domains,
including agriculture, environmental and natural disasters monitoring. Most of
the work in the literature utilize traditional segmentation methods that
require a large amount of computational and storage resources. Deep learning
has transformed computer vision and dramatically improved machine translation,
though it requires massive dataset for training and significant resources for
inference. More importantly, energy-efficient embedded vision hardware
delivering real-time and robust performance is crucial in the aforementioned
application. In this work, we propose a U-Net based tree delineation method,
which is effectively trained using multi-spectral imagery but can then
delineate single-spectrum images. The deep architecture that also performs
localization, i.e., a class label corresponds to each pixel, has been
successfully used to allow training with a small set of segmented images. The
ground truth data were generated using traditional image denoising and
segmentation approaches. To be able to execute the proposed DNN efficiently in
embedded platforms designed for deep learning approaches, we employ traditional
model compression and acceleration methods. Extensive evaluation studies using
data collected from UAVs equipped with multi-spectral cameras demonstrate the
effectiveness of the proposed methods in terms of delineation accuracy and
execution efficiency.
Related papers
- Large-Scale Data-Free Knowledge Distillation for ImageNet via Multi-Resolution Data Generation [53.95204595640208]
Data-Free Knowledge Distillation (DFKD) is an advanced technique that enables knowledge transfer from a teacher model to a student model without relying on original training data.
Previous approaches have generated synthetic images at high resolutions without leveraging information from real images.
MUSE generates images at lower resolutions while using Class Activation Maps (CAMs) to ensure that the generated images retain critical, class-specific features.
arXiv Detail & Related papers (2024-11-26T02:23:31Z) - Resource Efficient Perception for Vision Systems [0.0]
Our study introduces a framework aimed at mitigating these challenges by leveraging memory efficient patch based processing for high resolution images.
It incorporates a global context representation alongside local patch information, enabling a comprehensive understanding of the image content.
We demonstrate the effectiveness of our method through superior performance on 7 different benchmarks across classification, object detection, and segmentation.
arXiv Detail & Related papers (2024-05-12T05:33:00Z) - Deep Domain Adaptation: A Sim2Real Neural Approach for Improving Eye-Tracking Systems [80.62854148838359]
Eye image segmentation is a critical step in eye tracking that has great influence over the final gaze estimate.
We use dimensionality-reduction techniques to measure the overlap between the target eye images and synthetic training data.
Our methods result in robust, improved performance when tackling the discrepancy between simulation and real-world data samples.
arXiv Detail & Related papers (2024-03-23T22:32:06Z) - MOCA: Self-supervised Representation Learning by Predicting Masked Online Codebook Assignments [72.6405488990753]
Self-supervised learning can be used for mitigating the greedy needs of Vision Transformer networks.
We propose a single-stage and standalone method, MOCA, which unifies both desired properties.
We achieve new state-of-the-art results on low-shot settings and strong experimental results in various evaluation protocols.
arXiv Detail & Related papers (2023-07-18T15:46:20Z) - Domain Generalization for Mammographic Image Analysis with Contrastive
Learning [62.25104935889111]
The training of an efficacious deep learning model requires large data with diverse styles and qualities.
A novel contrastive learning is developed to equip the deep learning models with better style generalization capability.
The proposed method has been evaluated extensively and rigorously with mammograms from various vendor style domains and several public datasets.
arXiv Detail & Related papers (2023-04-20T11:40:21Z) - Multi-Task Self-Supervised Learning for Image Segmentation Task [0.0]
The paper presents 1. Self-supervised techniques to boost semantic segmentation performance using multi-task learning with Depth prediction and Surface Normalization.
2. Performance evaluation of the different types of weighing techniques (UW, Nash-MTL) used for Multi-task learning.
arXiv Detail & Related papers (2023-02-05T21:25:59Z) - Deep Learning Computer Vision Algorithms for Real-time UAVs On-board
Camera Image Processing [77.34726150561087]
This paper describes how advanced deep learning based computer vision algorithms are applied to enable real-time on-board sensor processing for small UAVs.
All algorithms have been developed using state-of-the-art image processing methods based on deep neural networks.
arXiv Detail & Related papers (2022-11-02T11:10:42Z) - Object Localization Through a Single Multiple-Model Convolutional Neural
Network with a Specific Training Approach [0.0]
A special training approach is proposed for a light convolutional neural network (CNN) to determine the region of interest in an image.
Almost all CNN-based detectors utilize a fixed input size image, which may yield poor performance when dealing with various object sizes.
arXiv Detail & Related papers (2021-03-24T16:52:01Z) - Learning Collision-Free Space Detection from Stereo Images: Homography
Matrix Brings Better Data Augmentation [16.99302954185652]
It remains an open challenge to train deep convolutional neural networks (DCNNs) using only a small quantity of training samples.
This paper explores an effective training data augmentation approach that can be employed to improve the overall DCNN performance.
arXiv Detail & Related papers (2020-12-14T19:14:35Z) - Visual Object Tracking by Segmentation with Graph Convolutional Network [7.729569666460712]
We propose to utilize graph convolutional network (GCN) model for superpixel based object tracking.
The proposed model provides a general end-to-end framework which integrates i) label linear prediction, and ii) structure-aware feature information of each superpixel together.
arXiv Detail & Related papers (2020-09-05T12:43:21Z) - Two-shot Spatially-varying BRDF and Shape Estimation [89.29020624201708]
We propose a novel deep learning architecture with a stage-wise estimation of shape and SVBRDF.
We create a large-scale synthetic training dataset with domain-randomized geometry and realistic materials.
Experiments on both synthetic and real-world datasets show that our network trained on a synthetic dataset can generalize well to real-world images.
arXiv Detail & Related papers (2020-04-01T12:56:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.