A two-step machine learning approach for crop disease detection: an
application of GAN and UAV technology
- URL: http://arxiv.org/abs/2109.11066v1
- Date: Sun, 19 Sep 2021 03:51:20 GMT
- Title: A two-step machine learning approach for crop disease detection: an
application of GAN and UAV technology
- Authors: Aaditya Prasad (1), Nikhil Mehta (1), Matthew Horak (2), Wan D. Bae
(3) ((1) Tesla STEM High School, (2) Lockheed Martin Corporation, (3) Seattle
University)
- Abstract summary: This paper presents a two-step machine learning approach that analyzes low-fidelity and high-fidelity images in sequence.
The results show an accuracy of 96.3% for the high-fidelity system and a 75.5% confidence level for our low-fidelity system.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Automated plant diagnosis is a technology that promises large increases in
cost-efficiency for agriculture. However, multiple problems reduce the
effectiveness of drones, including the inverse relationship between resolution
and speed and the lack of adequate labeled training data. This paper presents a
two-step machine learning approach that analyzes low-fidelity and high-fidelity
images in sequence, preserving efficiency as well as accuracy. Two
data-generators are also used to minimize class imbalance in the high-fidelity
dataset and to produce low-fidelity data that is representative of UAV images.
The analysis of applications and methods is conducted on a database of
high-fidelity apple tree images which are corrupted with class imbalance. The
application begins by generating high-fidelity data using generative networks
and then uses this novel data alongside the original high-fidelity data to
produce low-fidelity images. A machine-learning identifier identifies plants
and labels them as potentially diseased or not. A machine learning classifier
is then given the potentially diseased plant images and returns actual
diagnoses for these plants. The results show an accuracy of 96.3% for the
high-fidelity system and a 75.5% confidence level for our low-fidelity system.
Our drone technology shows promising results in accuracy when compared to
labor-based methods of diagnosis.
Related papers
- Small data deep learning methodology for in-field disease detection [6.2747249113031325]
We present the first machine learning model capable of detecting mild symptoms of late blight in potato crops.
Our proposal exploits the availability of high-resolution images via the concept of patching, and is based on deep convolutional neural networks with a focal loss function.
Our model correctly detects all cases of late blight in the test dataset, demonstrating a high level of accuracy and effectiveness in identifying early symptoms.
arXiv Detail & Related papers (2024-09-25T17:31:17Z) - StealthDiffusion: Towards Evading Diffusion Forensic Detection through Diffusion Model [62.25424831998405]
StealthDiffusion is a framework that modifies AI-generated images into high-quality, imperceptible adversarial examples.
It is effective in both white-box and black-box settings, transforming AI-generated images into high-quality adversarial forgeries.
arXiv Detail & Related papers (2024-08-11T01:22:29Z) - Accelerating Domain-Aware Electron Microscopy Analysis Using Deep Learning Models with Synthetic Data and Image-Wide Confidence Scoring [0.0]
We create a physics-based synthetic image and data generator, resulting in a machine learning model that achieves comparable precision (0.86), recall (0.63), F1 scores (0.71), and engineering property predictions (R2=0.82)
Our study demonstrates that synthetic data can eliminate human reliance in ML and provides a means for domain awareness in cases where many feature detections per image are needed.
arXiv Detail & Related papers (2024-08-02T20:15:15Z) - RIGID: A Training-free and Model-Agnostic Framework for Robust AI-Generated Image Detection [60.960988614701414]
RIGID is a training-free and model-agnostic method for robust AI-generated image detection.
RIGID significantly outperforms existing trainingbased and training-free detectors.
arXiv Detail & Related papers (2024-05-30T14:49:54Z) - Enhanced Droplet Analysis Using Generative Adversarial Networks [0.0]
This work develops an image generator named DropletGAN to generate images of droplets.
It is also used to develop a light droplet detector using the synthetic dataset.
To the best of our knowledge, this work stands as the first to employ a generative model for augmenting droplet detection.
arXiv Detail & Related papers (2024-02-24T21:20:53Z) - Generative models-based data labeling for deep networks regression:
application to seed maturity estimation from UAV multispectral images [3.6868861317674524]
Monitoring seed maturity is an increasing challenge in agriculture due to climate change and more restrictive practices.
Traditional methods are based on limited sampling in the field and analysis in laboratory.
We propose a method for estimating parsley seed maturity using multispectral UAV imagery, with a new approach for automatic data labeling.
arXiv Detail & Related papers (2022-08-09T09:06:51Z) - High-Resolution UAV Image Generation for Sorghum Panicle Detection [23.88932181375298]
We present an approach that uses synthetic training images from generative adversarial networks (GANs) for data augmentation to enhance the performance of Sorghum panicle detection and counting.
Our method can generate synthetic high-resolution UAV RGB images with panicle labels by using image-to-image translation GANs with a limited ground truth dataset of real UAV RGB images.
arXiv Detail & Related papers (2022-05-08T20:26:56Z) - Potato Crop Stress Identification in Aerial Images using Deep
Learning-based Object Detection [60.83360138070649]
The paper presents an approach for analyzing aerial images of a potato crop using deep neural networks.
The main objective is to demonstrate automated spatial recognition of a healthy versus stressed crop at a plant level.
Experimental validation demonstrated the ability for distinguishing healthy and stressed plants in field images, achieving an average Dice coefficient of 0.74.
arXiv Detail & Related papers (2021-06-14T21:57:40Z) - Many-to-One Distribution Learning and K-Nearest Neighbor Smoothing for
Thoracic Disease Identification [83.6017225363714]
deep learning has become the most powerful computer-aided diagnosis technology for improving disease identification performance.
For chest X-ray imaging, annotating large-scale data requires professional domain knowledge and is time-consuming.
In this paper, we propose many-to-one distribution learning (MODL) and K-nearest neighbor smoothing (KNNS) methods to improve a single model's disease identification performance.
arXiv Detail & Related papers (2021-02-26T02:29:30Z) - Circumventing Outliers of AutoAugment with Knowledge Distillation [102.25991455094832]
AutoAugment has been a powerful algorithm that improves the accuracy of many vision tasks.
This paper delves deep into the working mechanism, and reveals that AutoAugment may remove part of discriminative information from the training image.
To relieve the inaccuracy of supervision, we make use of knowledge distillation that refers to the output of a teacher model to guide network training.
arXiv Detail & Related papers (2020-03-25T11:51:41Z) - Rectified Meta-Learning from Noisy Labels for Robust Image-based Plant
Disease Diagnosis [64.82680813427054]
Plant diseases serve as one of main threats to food security and crop production.
One popular approach is to transform this problem as a leaf image classification task, which can be addressed by the powerful convolutional neural networks (CNNs)
We propose a novel framework that incorporates rectified meta-learning module into common CNN paradigm to train a noise-robust deep network without using extra supervision information.
arXiv Detail & Related papers (2020-03-17T09:51:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.