Computer-aided Interpretable Features for Leaf Image Classification
- URL: http://arxiv.org/abs/2106.08077v1
- Date: Tue, 15 Jun 2021 12:11:10 GMT
- Title: Computer-aided Interpretable Features for Leaf Image Classification
- Authors: Jayani P. G. Lakshika, Thiyanga S. Talagala
- Abstract summary: We introduce 52 computationally efficient features to classify plant species.
Length, width, area, texture correlation, monotonicity and scagnostics are to name few of them.
The results show that the features are sufficient to discriminate the classes of interest under both supervised and unsupervised learning settings.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Plant species identification is time consuming, costly, and requires lots of
efforts, and expertise knowledge. In recent, many researchers use deep learning
methods to classify plants directly using plant images. While deep learning
models have achieved a great success, the lack of interpretability limit their
widespread application. To overcome this, we explore the use of interpretable,
measurable and computer-aided features extracted from plant leaf images. Image
processing is one of the most challenging, and crucial steps in
feature-extraction. The purpose of image processing is to improve the leaf
image by removing undesired distortion. The main image processing steps of our
algorithm involves: i) Convert original image to RGB (Red-Green-Blue) image,
ii) Gray scaling, iii) Gaussian smoothing, iv) Binary thresholding, v) Remove
stalk, vi) Closing holes, and vii) Resize image. The next step after image
processing is to extract features from plant leaf images. We introduced 52
computationally efficient features to classify plant species. These features
are mainly classified into four groups as: i) shape-based features, ii)
color-based features, iii) texture-based features, and iv) scagnostic features.
Length, width, area, texture correlation, monotonicity and scagnostics are to
name few of them. We explore the ability of features to discriminate the
classes of interest under supervised learning and unsupervised learning
settings. For that, supervised dimensionality reduction technique, Linear
Discriminant Analysis (LDA), and unsupervised dimensionality reduction
technique, Principal Component Analysis (PCA) are used to convert and visualize
the images from digital-image space to feature space. The results show that the
features are sufficient to discriminate the classes of interest under both
supervised and unsupervised learning settings.
Related papers
- Advancements in Feature Extraction Recognition of Medical Imaging Systems Through Deep Learning Technique [0.36651088217486427]
An objective function based on weight is proposed to achieve the purpose of fast image recognition.
A technique for threshold optimization utilizing a simplex algorithm is presented.
It is found that different types of objects are independent of each other and compact in image processing.
arXiv Detail & Related papers (2024-05-23T04:46:51Z) - Grad-CAMO: Learning Interpretable Single-Cell Morphological Profiles from 3D Cell Painting Images [0.0]
We introduce Grad-CAMO, a novel single-cell interpretability score for supervised feature extractors.
Grad-CAMO measures the proportion of a model's attention that is concentrated on the cell of interest versus the background.
arXiv Detail & Related papers (2024-03-26T11:48:37Z) - Ensembling with Deep Generative Views [72.70801582346344]
generative models can synthesize "views" of artificial images that mimic real-world variations, such as changes in color or pose.
Here, we investigate whether such views can be applied to real images to benefit downstream analysis tasks such as image classification.
We use StyleGAN2 as the source of generative augmentations and investigate this setup on classification tasks involving facial attributes, cat faces, and cars.
arXiv Detail & Related papers (2021-04-29T17:58:35Z) - Scalable Visual Attribute Extraction through Hidden Layers of a Residual
ConvNet [7.6702700993064115]
We propose an approach for extracting visual attributes from images, leveraging the learned capability of the hidden layers of a general convolutional network.
We run experiments with a resnet-50 trained on Imagenet, on which we evaluate the output of its different blocks to discriminate between colors and textures.
arXiv Detail & Related papers (2021-03-31T23:39:20Z) - Factors of Influence for Transfer Learning across Diverse Appearance
Domains and Task Types [50.1843146606122]
A simple form of transfer learning is common in current state-of-the-art computer vision models.
Previous systematic studies of transfer learning have been limited and the circumstances in which it is expected to work are not fully understood.
In this paper we carry out an extensive experimental exploration of transfer learning across vastly different image domains.
arXiv Detail & Related papers (2021-03-24T16:24:20Z) - Automatic Radish Wilt Detection Using Image Processing Based Techniques
and Machine Learning Algorithm [3.4392739159262145]
We propose a segmentation and extraction-based technique to detect fusarium wilt in radish crops.
Recent wilt detection algorithms are either based on image processing techniques or conventional machine learning algorithms.
Our methodology is based on a hybrid algorithm, which combines image processing and machine learning.
arXiv Detail & Related papers (2020-09-01T01:37:01Z) - Saliency-driven Class Impressions for Feature Visualization of Deep
Neural Networks [55.11806035788036]
It is advantageous to visualize the features considered to be essential for classification.
Existing visualization methods develop high confidence images consisting of both background and foreground features.
In this work, we propose a saliency-driven approach to visualize discriminative features that are considered most important for a given task.
arXiv Detail & Related papers (2020-07-31T06:11:06Z) - Learning RGB-D Feature Embeddings for Unseen Object Instance
Segmentation [67.88276573341734]
We propose a new method for unseen object instance segmentation by learning RGB-D feature embeddings from synthetic data.
A metric learning loss function is utilized to learn to produce pixel-wise feature embeddings.
We further improve the segmentation accuracy with a new two-stage clustering algorithm.
arXiv Detail & Related papers (2020-07-30T00:23:07Z) - Distilling Localization for Self-Supervised Representation Learning [82.79808902674282]
Contrastive learning has revolutionized unsupervised representation learning.
Current contrastive models are ineffective at localizing the foreground object.
We propose a data-driven approach for learning in variance to backgrounds.
arXiv Detail & Related papers (2020-04-14T16:29:42Z) - Geometrically Mappable Image Features [85.81073893916414]
Vision-based localization of an agent in a map is an important problem in robotics and computer vision.
We propose a method that learns image features targeted for image-retrieval-based localization.
arXiv Detail & Related papers (2020-03-21T15:36:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.