E-PixelHop: An Enhanced PixelHop Method for Object Classification
- URL: http://arxiv.org/abs/2107.02966v1
- Date: Wed, 7 Jul 2021 01:22:12 GMT
- Title: E-PixelHop: An Enhanced PixelHop Method for Object Classification
- Authors: Yijing Yang, Vasileios Magoulianitis and C.-C. Jay Kuo
- Abstract summary: We propose an enhanced solution for object classification, called E-PixelHop, in this work.
E-PixelHop consists of the following steps.
We conduct pixel-level classification at each hop with various receptive fields.
Forth, pixel-level decisions from each hop and from each color subspace are fused together for image-level decision.
- Score: 29.062926577155885
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Based on PixelHop and PixelHop++, which are recently developed using the
successive subspace learning (SSL) framework, we propose an enhanced solution
for object classification, called E-PixelHop, in this work. E-PixelHop consists
of the following steps. First, to decouple the color channels for a color
image, we apply principle component analysis and project RGB three color
channels onto two principle subspaces which are processed separately for
classification. Second, to address the importance of multi-scale features, we
conduct pixel-level classification at each hop with various receptive fields.
Third, to further improve pixel-level classification accuracy, we develop a
supervised label smoothing (SLS) scheme to ensure prediction consistency.
Forth, pixel-level decisions from each hop and from each color subspace are
fused together for image-level decision. Fifth, to resolve confusing classes
for further performance boosting, we formulate E-PixelHop as a two-stage
pipeline. In the first stage, multi-class classification is performed to get a
soft decision for each class, where the top 2 classes with the highest
probabilities are called confusing classes. Then,we conduct a binary
classification in the second stage. The main contributions lie in Steps 1, 3
and 5.We use the classification of the CIFAR-10 dataset as an example to
demonstrate the effectiveness of the above-mentioned key components of
E-PixelHop.
Related papers
- DFEN: Dual Feature Equalization Network for Medical Image Segmentation [9.091452460153672]
We propose a dual feature equalization network based on the hybrid architecture of Swin Transformer and Convolutional Neural Network.<n>Swin Transformer is utilized as both the encoder and decoder, thereby bolstering the ability of the model to capture long-range dependencies and spatial correlations.
arXiv Detail & Related papers (2025-05-09T09:38:43Z) - Dual-level Fuzzy Learning with Patch Guidance for Image Ordinal Regression [8.538034422744005]
Ordinal regression bridges regression and classification by assigning objects to ordered classes.<n>Current approaches are limited by the availability of only image-level ordinal labels.<n>We propose a Dual-level Fuzzy Learning with Patch Guidance framework, named DFPG.
arXiv Detail & Related papers (2025-05-09T07:01:14Z) - Dual-stage Hyperspectral Image Classification Model with Spectral Supertoken [15.426635239291729]
We introduce the Dual-stage Spectral Supertoken (DSTC), inspired by superpixel concepts.
DSTC employs spectrum-derivative-based pixel clustering to group pixels with similar spectral characteristics into spectral supertokens.
We also propose a class-proportion-based soft label, which adaptively assigns weights to different categories based on their prevalence.
arXiv Detail & Related papers (2024-07-10T01:58:30Z) - Differentiable Registration of Images and LiDAR Point Clouds with
VoxelPoint-to-Pixel Matching [58.10418136917358]
Cross-modality registration between 2D images from cameras and 3D point clouds from LiDARs is a crucial task in computer vision and robotic training.
Previous methods estimate 2D-3D correspondences by matching point and pixel patterns learned by neural networks.
We learn a structured cross-modality matching solver to represent 3D features via a different latent pixel space.
arXiv Detail & Related papers (2023-12-07T05:46:10Z) - Information-guided pixel augmentation for pixel-wise contrastive
learning [22.00687816406677]
pixel-wise contrastive learning helps with pixel-wise tasks such as medical landmark detection.
We propose a pixel augmentation method with a pixel granularity for enhancing unsupervised pixel-wise contrastive learning.
arXiv Detail & Related papers (2022-11-14T05:12:23Z) - SePiCo: Semantic-Guided Pixel Contrast for Domain Adaptive Semantic
Segmentation [52.62441404064957]
Domain adaptive semantic segmentation attempts to make satisfactory dense predictions on an unlabeled target domain by utilizing the model trained on a labeled source domain.
Many methods tend to alleviate noisy pseudo labels, however, they ignore intrinsic connections among cross-domain pixels with similar semantic concepts.
We propose Semantic-Guided Pixel Contrast (SePiCo), a novel one-stage adaptation framework that highlights the semantic concepts of individual pixel.
arXiv Detail & Related papers (2022-04-19T11:16:29Z) - Multi-level Second-order Few-shot Learning [111.0648869396828]
We propose a Multi-level Second-order (MlSo) few-shot learning network for supervised or unsupervised few-shot image classification and few-shot action recognition.
We leverage so-called power-normalized second-order base learner streams combined with features that express multiple levels of visual abstraction.
We demonstrate respectable results on standard datasets such as Omniglot, mini-ImageNet, tiered-ImageNet, Open MIC, fine-grained datasets such as CUB Birds, Stanford Dogs and Cars, and action recognition datasets such as HMDB51, UCF101, and mini-MIT.
arXiv Detail & Related papers (2022-01-15T19:49:00Z) - Saliency Enhancement using Superpixel Similarity [77.34726150561087]
Saliency Object Detection (SOD) has several applications in image analysis.
Deep-learning-based SOD methods are among the most effective, but they may miss foreground parts with similar colors.
We introduce a post-processing method, named textitSaliency Enhancement over Superpixel Similarity (SESS)
We demonstrate that SESS can consistently and considerably improve the results of three deep-learning-based SOD methods on five image datasets.
arXiv Detail & Related papers (2021-12-01T17:22:54Z) - Isometric Propagation Network for Generalized Zero-shot Learning [72.02404519815663]
A popular strategy is to learn a mapping between the semantic space of class attributes and the visual space of images based on the seen classes and their data.
We propose Isometric propagation Network (IPN), which learns to strengthen the relation between classes within each space and align the class dependency in the two spaces.
IPN achieves state-of-the-art performance on three popular Zero-shot learning benchmarks.
arXiv Detail & Related papers (2021-02-03T12:45:38Z) - Learning Category- and Instance-Aware Pixel Embedding for Fast Panoptic
Segmentation [47.26296379603003]
Panoptic segmentation (PS) is a complex scene understanding task.
PS results are simply derived by assigning each pixel to a detected instance or a stuff class.
Our method not only demonstrates fast inference speed but also the first one-stage method to achieve comparable performance to two-stage methods.
arXiv Detail & Related papers (2020-09-28T14:07:50Z) - PixelHop++: A Small Successive-Subspace-Learning-Based (SSL-based) Model
for Image Classification [30.49387075658641]
We propose an improved PixelHop method and call it PixelHop++.
In PixelHop++, one can control the learning model size of fine-granularity,offering a flexible tradeoff between the model size and the classification performance.
arXiv Detail & Related papers (2020-02-08T11:08:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.