Mushroom image recognition and distance generation based on
attention-mechanism model and genetic information
- URL: http://arxiv.org/abs/2206.13383v1
- Date: Mon, 27 Jun 2022 15:43:03 GMT
- Title: Mushroom image recognition and distance generation based on
attention-mechanism model and genetic information
- Authors: Wenbin Liao, Jiewen Xiao, Chengbo Zhao, Yonggong Han, ZhiJie Geng,
Jianxin Wang, Yihua Yang
- Abstract summary: We propose a new model based on attention-mechanism, MushroomNet, which applies the lightweight network MobileNetV3 as the backbone model.
On the public dataset, the test accuracy of the MushroomNet model has reached 83.9%, and on the local dataset, the test accuracy has reached 77.4%.
We found that using the MES activation function can predict the genetic distance of mushrooms very well, but the accuracy is lower than that of SoftMax.
- Score: 4.845860279763184
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The species identification of Macrofungi, i.e. mushrooms, has always been a
challenging task. There are still a large number of poisonous mushrooms that
have not been found, which poses a risk to people's life. However, the
traditional identification method requires a large number of experts with
knowledge in the field of taxonomy for manual identification, it is not only
inefficient but also consumes a lot of manpower and capital costs. In this
paper, we propose a new model based on attention-mechanism, MushroomNet, which
applies the lightweight network MobileNetV3 as the backbone model, combined
with the attention structure proposed by us, and has achieved excellent
performance in the mushroom recognition task. On the public dataset, the test
accuracy of the MushroomNet model has reached 83.9%, and on the local dataset,
the test accuracy has reached 77.4%. The proposed attention mechanisms well
focused attention on the bodies of mushroom image for mixed channel attention
and the attention heat maps visualized by Grad-CAM. Further, in this study,
genetic distance was added to the mushroom image recognition task, the genetic
distance was used as the representation space, and the genetic distance between
each pair of mushroom species in the dataset was used as the embedding of the
genetic distance representation space, so as to predict the image distance and
species. identify. We found that using the MES activation function can predict
the genetic distance of mushrooms very well, but the accuracy is lower than
that of SoftMax. The proposed MushroomNet was demonstrated it shows great
potential for automatic and online mushroom image and the proposed automatic
procedure would assist and be a reference to traditional mushroom
classification.
Related papers
- High-Throughput Phenotyping using Computer Vision and Machine Learning [0.0]
We used a dataset provided by Oak Ridge National Laboratory with 1,672 images of Populus Trichocarpa with white labels displaying treatment.
Optical character recognition (OCR) was used to read these labels on the plants.
Machine learning models were used to predict treatment based on those classifications, and analyzed encoded EXIF tags were used for the purpose of finding leaf size and correlations between phenotypes.
arXiv Detail & Related papers (2024-07-08T19:46:31Z) - Genetic InfoMax: Exploring Mutual Information Maximization in
High-Dimensional Imaging Genetics Studies [50.11449968854487]
Genome-wide association studies (GWAS) are used to identify relationships between genetic variations and specific traits.
Representation learning for imaging genetics is largely under-explored due to the unique challenges posed by GWAS.
We introduce a trans-modal learning framework Genetic InfoMax (GIM) to address the specific challenges of GWAS.
arXiv Detail & Related papers (2023-09-26T03:59:21Z) - Spatial Implicit Neural Representations for Global-Scale Species Mapping [72.92028508757281]
Given a set of locations where a species has been observed, the goal is to build a model to predict whether the species is present or absent at any location.
Traditional methods struggle to take advantage of emerging large-scale crowdsourced datasets.
We use Spatial Implicit Neural Representations (SINRs) to jointly estimate the geographical range of 47k species simultaneously.
arXiv Detail & Related papers (2023-06-05T03:36:01Z) - Few-Shot Learning Enables Population-Scale Analysis of Leaf Traits in
Populus trichocarpa [1.9089478605920305]
This work is designed to provide the plant phenotyping community with (i) methods for fast and accurate image-based feature extraction that require minimal training data, and (ii) a new population-scale data set, including 68 different leaf phenotypes, for domain scientists and machine learning researchers.
All of the few-shot learning code, data, and results are made publicly available.
arXiv Detail & Related papers (2023-01-24T23:40:01Z) - Semantic Image Segmentation with Deep Learning for Vine Leaf Phenotyping [59.0626764544669]
In this study, we use Deep Learning methods to semantically segment grapevine leaves images in order to develop an automated object detection system for leaf phenotyping.
Our work contributes to plant lifecycle monitoring through which dynamic traits such as growth and development can be captured and quantified.
arXiv Detail & Related papers (2022-10-24T14:37:09Z) - Taxonomy and evolution predicting using deep learning in images [9.98733710208427]
This study creates a novel recognition framework by systematically studying the mushroom image recognition problem.
We present the first method to map images to DNA, namely used an encoder mapping image to genetic distances, and then decoded DNA through a pre-trained decoder.
arXiv Detail & Related papers (2022-06-28T13:54:14Z) - Towards a Deeper Understanding of Skeleton-based Gait Recognition [4.812321790984493]
In recent years, most gait recognition methods used the person's silhouette to extract the gait features.
Model-based methods do not suffer from these problems and are able to represent the temporal motion of body joints.
In this work, we propose an approach based on Graph Convolutional Networks (GCNs) that combines higher-order inputs, and residual networks.
arXiv Detail & Related papers (2022-04-16T18:23:37Z) - A deep convolutional neural network for classification of Aedes
albopictus mosquitoes [1.6758573326215689]
We introduce the application of two Deep Convolutional Neural Networks in a comparative study to automate the classification task.
We use the transfer learning principle to train two state-of-the-art architectures on the data provided by the Mosquito Alert project.
In addition, we applied explainable models based on the Grad-CAM algorithm to visualise the most discriminant regions of the classified images.
arXiv Detail & Related papers (2021-10-29T17:58:32Z) - Deep Low-Shot Learning for Biological Image Classification and
Visualization from Limited Training Samples [52.549928980694695]
In situ hybridization (ISH) gene expression pattern images from the same developmental stage are compared.
labeling training data with precise stages is very time-consuming even for biologists.
We propose a deep two-step low-shot learning framework to accurately classify ISH images using limited training images.
arXiv Detail & Related papers (2020-10-20T06:06:06Z) - Two-View Fine-grained Classification of Plant Species [66.75915278733197]
We propose a novel method based on a two-view leaf image representation and a hierarchical classification strategy for fine-grained recognition of plant species.
A deep metric based on Siamese convolutional neural networks is used to reduce the dependence on a large number of training samples and make the method scalable to new plant species.
arXiv Detail & Related papers (2020-05-18T21:57:47Z) - Automatic image-based identification and biomass estimation of
invertebrates [70.08255822611812]
Time-consuming sorting and identification of taxa pose strong limitations on how many insect samples can be processed.
We propose to replace the standard manual approach of human expert-based sorting and identification with an automatic image-based technology.
We use state-of-the-art Resnet-50 and InceptionV3 CNNs for the classification task.
arXiv Detail & Related papers (2020-02-05T21:38:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.