Radon cumulative distribution transform subspace modeling for image
classification
- URL: http://arxiv.org/abs/2004.03669v3
- Date: Wed, 2 Mar 2022 20:43:03 GMT
- Title: Radon cumulative distribution transform subspace modeling for image
classification
- Authors: Mohammad Shifat-E-Rabbi, Xuwang Yin, Abu Hasnat Mohammad Rubaiyat,
Shiying Li, Soheil Kolouri, Akram Aldroubi, Jonathan M. Nichols, and Gustavo
K. Rohde
- Abstract summary: We present a new supervised image classification method applicable to a broad class of image deformation models.
The method makes use of the previously described Radon Cumulative Distribution Transform (R-CDT) for image data.
In addition to the test accuracy performances, we show improvements in terms of computational efficiency.
- Score: 18.709734704950804
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a new supervised image classification method applicable to a broad
class of image deformation models. The method makes use of the previously
described Radon Cumulative Distribution Transform (R-CDT) for image data, whose
mathematical properties are exploited to express the image data in a form that
is more suitable for machine learning. While certain operations such as
translation, scaling, and higher-order transformations are challenging to model
in native image space, we show the R-CDT can capture some of these variations
and thus render the associated image classification problems easier to solve.
The method -- utilizing a nearest-subspace algorithm in R-CDT space -- is
simple to implement, non-iterative, has no hyper-parameters to tune, is
computationally efficient, label efficient, and provides competitive accuracies
to state-of-the-art neural networks for many types of classification problems.
In addition to the test accuracy performances, we show improvements (with
respect to neural network-based methods) in terms of computational efficiency
(it can be implemented without the use of GPUs), number of training samples
needed for training, as well as out-of-distribution generalization. The Python
code for reproducing our results is available at
https://github.com/rohdelab/rcdt_ns_classifier.
Related papers
- Performance of GAN-based augmentation for deep learning COVID-19 image
classification [57.1795052451257]
The biggest challenge in the application of deep learning to the medical domain is the availability of training data.
Data augmentation is a typical methodology used in machine learning when confronted with a limited data set.
In this work, a StyleGAN2-ADA model of Generative Adversarial Networks is trained on the limited COVID-19 chest X-ray image set.
arXiv Detail & Related papers (2023-04-18T15:39:58Z) - Traditional Classification Neural Networks are Good Generators: They are
Competitive with DDPMs and GANs [104.72108627191041]
We show that conventional neural network classifiers can generate high-quality images comparable to state-of-the-art generative models.
We propose a mask-based reconstruction module to make semantic gradients-aware to synthesize plausible images.
We show that our method is also applicable to text-to-image generation by regarding image-text foundation models.
arXiv Detail & Related papers (2022-11-27T11:25:35Z) - Large-Margin Representation Learning for Texture Classification [67.94823375350433]
This paper presents a novel approach combining convolutional layers (CLs) and large-margin metric learning for training supervised models on small datasets for texture classification.
The experimental results on texture and histopathologic image datasets have shown that the proposed approach achieves competitive accuracy with lower computational cost and faster convergence when compared to equivalent CNNs.
arXiv Detail & Related papers (2022-06-17T04:07:45Z) - TransformNet: Self-supervised representation learning through predicting
geometric transformations [0.8098097078441623]
We describe the unsupervised semantic feature learning approach for recognition of the geometric transformation applied to the input data.
The basic concept of our approach is that if someone is unaware of the objects in the images, he/she would not be able to quantitatively predict the geometric transformation that was applied to them.
arXiv Detail & Related papers (2022-02-08T22:41:01Z) - Feature transforms for image data augmentation [74.12025519234153]
In image classification, many augmentation approaches utilize simple image manipulation algorithms.
In this work, we build ensembles on the data level by adding images generated by combining fourteen augmentation approaches.
Pretrained ResNet50 networks are finetuned on training sets that include images derived from each augmentation method.
arXiv Detail & Related papers (2022-01-24T14:12:29Z) - Invariance encoding in sliced-Wasserstein space for image classification
with limited training data [10.435990685398595]
We propose to mathematically augment a nearest subspace classification model in sliced-Wasserstein space by exploiting certain mathematical properties of the Radon Cumulative Distribution Transform (R-CDT)
We demonstrate that for a particular type of learning problem, our mathematical solution has advantages over data augmentation with deep CNNs in terms of classification accuracy and computational complexity.
arXiv Detail & Related papers (2022-01-09T10:25:27Z) - FeatMatch: Feature-Based Augmentation for Semi-Supervised Learning [64.32306537419498]
We propose a novel learned feature-based refinement and augmentation method that produces a varied set of complex transformations.
These transformations also use information from both within-class and across-class representations that we extract through clustering.
We demonstrate that our method is comparable to current state of art for smaller datasets while being able to scale up to larger datasets.
arXiv Detail & Related papers (2020-07-16T17:55:31Z) - Learning to Learn Parameterized Classification Networks for Scalable
Input Images [76.44375136492827]
Convolutional Neural Networks (CNNs) do not have a predictable recognition behavior with respect to the input resolution change.
We employ meta learners to generate convolutional weights of main networks for various input scales.
We further utilize knowledge distillation on the fly over model predictions based on different input resolutions.
arXiv Detail & Related papers (2020-07-13T04:27:25Z) - Improving accuracy and speeding up Document Image Classification through
parallel systems [4.102028235659611]
We show in the RVL-CDIP dataset that we can improve previous results with a much lighter model.
We present an ensemble pipeline which is able to boost solely image input.
Lastly, we expose the training performance differences between PyTorch and Deep Learning frameworks.
arXiv Detail & Related papers (2020-06-16T13:36:07Z) - Adaptive Fractional Dilated Convolution Network for Image Aesthetics
Assessment [33.945579916184364]
An adaptive fractional dilated convolution (AFDC) is developed to tackle this issue in convolutional kernel level.
We provide a concise formulation for mini-batch training and utilize a grouping strategy to reduce computational overhead.
Our experimental results demonstrate that our proposed method achieves state-of-the-art performance on image aesthetics assessment over the AVA dataset.
arXiv Detail & Related papers (2020-04-06T21:56:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.