Scalable learning for bridging the species gap in image-based plant
phenotyping
- URL: http://arxiv.org/abs/2003.10757v2
- Date: Fri, 24 Apr 2020 02:50:45 GMT
- Title: Scalable learning for bridging the species gap in image-based plant
phenotyping
- Authors: Daniel Ward, Peyman Moghadam
- Abstract summary: The traditional paradigm of applying deep learning -- collect, annotate and train on data -- is not applicable to image-based plant phenotyping.
Data costs include growing physical samples, imaging and labelling them.
Model performance is impacted by the species gap between the domain of each plant species.
- Score: 2.208242292882514
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The traditional paradigm of applying deep learning -- collect, annotate and
train on data -- is not applicable to image-based plant phenotyping as almost
400,000 different plant species exists. Data costs include growing physical
samples, imaging and labelling them. Model performance is impacted by the
species gap between the domain of each plant species, it is not generalisable
and may not transfer to unseen plant species. In this paper, we investigate the
use of synthetic data for leaf instance segmentation. We study multiple
synthetic data training regimes using Mask-RCNN when few or no annotated real
data is available. We also present UPGen: a Universal Plant Generator for
bridging the species gap. UPGen leverages domain randomisation to produce
widely distributed data samples and models stochastic biological variation. Our
methods outperform standard practices, such as transfer learning from publicly
available plant data, by 26.6% and 51.46% on two unseen plant species
respectively. We benchmark UPGen by competing in the CVPPP Leaf Segmentation
Challenge and set a new state-of-the-art, a mean of 88% across A1-4 test
datasets. This study is applicable to use of synthetic data for automating the
measurement of phenotypic traits. Our synthetic dataset and pretrained model
are available at https://csiro-robotics.github.io/UPGen_Webpage/.
Related papers
- DreamDA: Generative Data Augmentation with Diffusion Models [68.22440150419003]
This paper proposes a new classification-oriented framework DreamDA.
DreamDA generates diverse samples that adhere to the original data distribution by considering training images in the original data as seeds.
In addition, since the labels of the generated data may not align with the labels of their corresponding seed images, we introduce a self-training paradigm for generating pseudo labels.
arXiv Detail & Related papers (2024-03-19T15:04:35Z) - Training Class-Imbalanced Diffusion Model Via Overlap Optimization [55.96820607533968]
Diffusion models trained on real-world datasets often yield inferior fidelity for tail classes.
Deep generative models, including diffusion models, are biased towards classes with abundant training images.
We propose a method based on contrastive learning to minimize the overlap between distributions of synthetic images for different classes.
arXiv Detail & Related papers (2024-02-16T16:47:21Z) - Spatial Implicit Neural Representations for Global-Scale Species Mapping [72.92028508757281]
Given a set of locations where a species has been observed, the goal is to build a model to predict whether the species is present or absent at any location.
Traditional methods struggle to take advantage of emerging large-scale crowdsourced datasets.
We use Spatial Implicit Neural Representations (SINRs) to jointly estimate the geographical range of 47k species simultaneously.
arXiv Detail & Related papers (2023-06-05T03:36:01Z) - Few-Shot Learning Enables Population-Scale Analysis of Leaf Traits in
Populus trichocarpa [1.9089478605920305]
This work is designed to provide the plant phenotyping community with (i) methods for fast and accurate image-based feature extraction that require minimal training data, and (ii) a new population-scale data set, including 68 different leaf phenotypes, for domain scientists and machine learning researchers.
All of the few-shot learning code, data, and results are made publicly available.
arXiv Detail & Related papers (2023-01-24T23:40:01Z) - Inside Out: Transforming Images of Lab-Grown Plants for Machine Learning
Applications in Agriculture [0.0]
We employ a contrastive unpaired translation (CUT) generative adversarial network (GAN) to translate indoor plant images to appear as field images.
While we train our network to translate an image containing only a single plant, we show that our method is easily extendable to produce multiple-plant field images.
We also use our synthetic multi-plant images to train several YoloV5 nano object detection models to perform the task of plant detection.
arXiv Detail & Related papers (2022-11-05T20:51:45Z) - Multi-Domain Few-Shot Learning and Dataset for Agricultural Applications [0.0]
We propose a method to learn from a few samples to automatically classify different pests, plants, and their diseases.
We learn a feature extractor to generate embeddings and then update the embeddings using Transformers.
We conduct 42 experiments in total to comprehensively analyze the model and it achieves up to 14% and 24% performance gains on few-shot image classification benchmarks.
arXiv Detail & Related papers (2021-09-21T04:20:18Z) - Mitigating Generation Shifts for Generalized Zero-Shot Learning [52.98182124310114]
Generalized Zero-Shot Learning (GZSL) is the task of leveraging semantic information (e.g., attributes) to recognize the seen and unseen samples, where unseen classes are not observable during training.
We propose a novel Generation Shifts Mitigating Flow framework for learning unseen data synthesis efficiently and effectively.
Experimental results demonstrate that GSMFlow achieves state-of-the-art recognition performance in both conventional and generalized zero-shot settings.
arXiv Detail & Related papers (2021-07-07T11:43:59Z) - Classification of Seeds using Domain Randomization on Self-Supervised
Learning Frameworks [0.0]
Key bottleneck is the need for an extensive amount of labelled data to train the convolutional neural networks (CNN)
The work leverages the concepts of Contrastive Learning and Domain Randomi-zation in order to achieve the same.
The use of synthetic images generated from a representational sample crop of real-world images alleviates the need for a large volume of test subjects.
arXiv Detail & Related papers (2021-03-29T12:50:06Z) - Pre-Trained Models for Heterogeneous Information Networks [57.78194356302626]
We propose a self-supervised pre-training and fine-tuning framework, PF-HIN, to capture the features of a heterogeneous information network.
PF-HIN consistently and significantly outperforms state-of-the-art alternatives on each of these tasks, on four datasets.
arXiv Detail & Related papers (2020-07-07T03:36:28Z) - Two-View Fine-grained Classification of Plant Species [66.75915278733197]
We propose a novel method based on a two-view leaf image representation and a hierarchical classification strategy for fine-grained recognition of plant species.
A deep metric based on Siamese convolutional neural networks is used to reduce the dependence on a large number of training samples and make the method scalable to new plant species.
arXiv Detail & Related papers (2020-05-18T21:57:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.