Standardizing and Centralizing Datasets to Enable Efficient Training of
Agricultural Deep Learning Models
- URL: http://arxiv.org/abs/2208.02707v1
- Date: Thu, 4 Aug 2022 15:10:36 GMT
- Title: Standardizing and Centralizing Datasets to Enable Efficient Training of
Agricultural Deep Learning Models
- Authors: Amogh Joshi, Dario Guevara, Mason Earles
- Abstract summary: Deep learning models are typically fine-tuned to agricultural tasks using model weights originally fit to more general, non-agricultural datasets.
We collect a wide range of existing public datasets for three distinct tasks, standardize them, and construct standard training and evaluation pipelines.
We conduct a number of experiments using methods which are commonly used in deep learning tasks, but unexplored in their domain-specific applications for agriculture.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In recent years, deep learning models have become the standard for
agricultural computer vision. Such models are typically fine-tuned to
agricultural tasks using model weights that were originally fit to more
general, non-agricultural datasets. This lack of agriculture-specific
fine-tuning potentially increases training time and resource use, and decreases
model performance, leading an overall decrease in data efficiency. To overcome
this limitation, we collect a wide range of existing public datasets for three
distinct tasks, standardize them, and construct standard training and
evaluation pipelines, providing us with a set of benchmarks and pretrained
models. We then conduct a number of experiments using methods which are
commonly used in deep learning tasks, but unexplored in their domain-specific
applications for agriculture. Our experiments guide us in developing a number
of approaches to improve data efficiency when training agricultural deep
learning models, without large-scale modifications to existing pipelines. Our
results demonstrate that even slight training modifications, such as using
agricultural pretrained model weights, or adopting specific spatial
augmentations into data processing pipelines, can significantly boost model
performance and result in shorter convergence time, saving training resources.
Furthermore, we find that even models trained on low-quality annotations can
produce comparable levels of performance to their high-quality equivalents,
suggesting that datasets with poor annotations can still be used for training,
expanding the pool of currently available datasets. Our methods are broadly
applicable throughout agricultural deep learning, and present high potential
for significant data efficiency improvements.
Related papers
- Transfer Learning with Foundational Models for Time Series Forecasting using Low-Rank Adaptations [0.0]
This study proposes LLIAM, the Llama Lora-Integrated Autorregresive Model.
Low-Rank Adaptations are used to enhance the knowledge of the model with diverse time series datasets, known as the fine-tuning phase.
arXiv Detail & Related papers (2024-10-15T12:14:01Z) - Scaling Laws Do Not Scale [54.72120385955072]
Recent work has argued that as the size of a dataset increases, the performance of a model trained on that dataset will increase.
We argue that this scaling law relationship depends on metrics used to measure performance that may not correspond with how different groups of people perceive the quality of models' output.
Different communities may also have values in tension with each other, leading to difficult, potentially irreconcilable choices about metrics used for model evaluations.
arXiv Detail & Related papers (2023-07-05T15:32:21Z) - Phased Data Augmentation for Training a Likelihood-Based Generative Model with Limited Data [0.0]
Generative models excel in creating realistic images, yet their dependency on extensive datasets for training presents significant challenges.
Current data-efficient methods largely focus on GAN architectures, leaving a gap in training other types of generative models.
"phased data augmentation" is a novel technique that addresses this gap by optimizing training in limited data scenarios without altering the inherent data distribution.
arXiv Detail & Related papers (2023-05-22T03:38:59Z) - On the Costs and Benefits of Adopting Lifelong Learning for Software
Analytics -- Empirical Study on Brown Build and Risk Prediction [17.502553991799832]
This paper evaluates the use of lifelong learning (LL) for industrial use cases at Ubisoft.
LL is used to continuously build and maintain ML-based software analytics tools using an incremental learner that progressively updates the old model using new data.
arXiv Detail & Related papers (2023-05-16T21:57:16Z) - INGENIOUS: Using Informative Data Subsets for Efficient Pre-Training of
Language Models [40.54353850357839]
We show how we can employ submodular optimization to select highly representative subsets of the training corpora.
We show that the resulting models achieve up to $sim99%$ of the performance of the fully-trained models.
arXiv Detail & Related papers (2023-05-11T09:24:41Z) - Evaluating and Crafting Datasets Effective for Deep Learning With Data
Maps [0.0]
Training on large datasets often requires excessive system resources and an infeasible amount of time.
For supervised learning, large datasets require more time for manually labeling samples.
We propose a method of curating smaller datasets with comparable out-of-distribution model accuracy after an initial training session.
arXiv Detail & Related papers (2022-08-22T03:30:18Z) - Teacher Guided Training: An Efficient Framework for Knowledge Transfer [86.6784627427194]
We propose the teacher-guided training (TGT) framework for training a high-quality compact model.
TGT exploits the fact that the teacher has acquired a good representation of the underlying data domain.
We find that TGT can improve accuracy on several image classification benchmarks and a range of text classification and retrieval tasks.
arXiv Detail & Related papers (2022-08-14T10:33:58Z) - Beyond Transfer Learning: Co-finetuning for Action Localisation [64.07196901012153]
We propose co-finetuning -- simultaneously training a single model on multiple upstream'' and downstream'' tasks.
We demonstrate that co-finetuning outperforms traditional transfer learning when using the same total amount of data.
We also show how we can easily extend our approach to multiple upstream'' datasets to further improve performance.
arXiv Detail & Related papers (2022-07-08T10:25:47Z) - CHALLENGER: Training with Attribution Maps [63.736435657236505]
We show that utilizing attribution maps for training neural networks can improve regularization of models and thus increase performance.
In particular, we show that our generic domain-independent approach yields state-of-the-art results in vision, natural language processing and on time series tasks.
arXiv Detail & Related papers (2022-05-30T13:34:46Z) - BERT WEAVER: Using WEight AVERaging to enable lifelong learning for
transformer-based models in biomedical semantic search engines [49.75878234192369]
We present WEAVER, a simple, yet efficient post-processing method that infuses old knowledge into the new model.
We show that applying WEAVER in a sequential manner results in similar word embedding distributions as doing a combined training on all data at once.
arXiv Detail & Related papers (2022-02-21T10:34:41Z) - Understanding the World Through Action [91.3755431537592]
I will argue that a general, principled, and powerful framework for utilizing unlabeled data can be derived from reinforcement learning.
I will discuss how such a procedure is more closely aligned with potential downstream tasks.
arXiv Detail & Related papers (2021-10-24T22:33:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.