HB-net: Holistic bursting cell cluster integrated network for occluded
multi-objects recognition
- URL: http://arxiv.org/abs/2310.11834v1
- Date: Wed, 18 Oct 2023 09:38:51 GMT
- Title: HB-net: Holistic bursting cell cluster integrated network for occluded
multi-objects recognition
- Authors: Xudong Gao, Xiao Guang Gao, Jia Rong, Xiaowei Chen, Xiang Liao, Jun
Chen
- Abstract summary: Multi-label classification (MLC) challenges arise when objects within the visual field may occlude one another.
Traditional convolutional neural networks (CNNs) can tackle these challenges; however, those models tend to be bulky and can only attain modest levels of accuracy.
This paper introduces a pioneering integrated network framework named HB-net.
- Score: 10.113801578414881
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Within the realm of image recognition, a specific category of multi-label
classification (MLC) challenges arises when objects within the visual field may
occlude one another, demanding simultaneous identification of both occluded and
occluding objects. Traditional convolutional neural networks (CNNs) can tackle
these challenges; however, those models tend to be bulky and can only attain
modest levels of accuracy. Leveraging insights from cutting-edge neural science
research, specifically the Holistic Bursting (HB) cell, this paper introduces a
pioneering integrated network framework named HB-net. Built upon the foundation
of HB cell clusters, HB-net is designed to address the intricate task of
simultaneously recognizing multiple occluded objects within images. Various
Bursting cell cluster structures are introduced, complemented by an evidence
accumulation mechanism. Testing is conducted on multiple datasets comprising
digits and letters. The results demonstrate that models incorporating the HB
framework exhibit a significant $2.98\%$ enhancement in recognition accuracy
compared to models without the HB framework ($1.0298$ times, $p=0.0499$).
Although in high-noise settings, standard CNNs exhibit slightly greater
robustness when compared to HB-net models, the models that combine the HB
framework and EA mechanism achieve a comparable level of accuracy and
resilience to ResNet50, despite having only three convolutional layers and
approximately $1/30$ of the parameters. The findings of this study offer
valuable insights for improving computer vision algorithms. The essential code
is provided at https://github.com/d-lab438/hb-net.git.
Related papers
- A Hybrid CNN-VSSM model for Multi-View, Multi-Task Mammography Analysis: Robust Diagnosis with Attention-Based Fusion [5.15423063632115]
Early and accurate interpretation of screening mammograms is essential for effective breast cancer detection.<n>Existing AI approaches fall short by focusing on single view inputs or single-task outputs.<n>We propose a novel multi-view, multitask hybrid deep learning framework that processes all four standard mammography views.
arXiv Detail & Related papers (2025-07-22T18:52:18Z) - A Hybrid Multi-Well Hopfield-CNN with Feature Extraction and K-Means for MNIST Classification [0.0]
This study presents a hybrid model for classifying handwritten digits in the MNIST dataset.<n>It combines convolutional neural networks (CNNs) with a multi-well Hopfield network.<n>The model achieves a high test accuracy of 99.2% on 10,000 MNIST images.
arXiv Detail & Related papers (2025-07-11T17:26:06Z) - Soybean Disease Detection via Interpretable Hybrid CNN-GNN: Integrating MobileNetV2 and GraphSAGE with Cross-Modal Attention [1.927711700724334]
Soybean leaf disease detection is critical for agricultural productivity but faces challenges due to visually similar symptoms and limited interpretability.
This paper proposes an interpretable hybrid Sequential CNN-Graph Neural Network (GNN) framework that synergizes MobileNetV2 for localized feature extraction and GraphSAGE for relational modeling.
Cross-modal interpretability is achieved via Grad-CAM and Eigen-CAM visualizations, generating heatmaps to highlight disease-influential regions.
arXiv Detail & Related papers (2025-03-03T08:12:09Z) - DCENWCNet: A Deep CNN Ensemble Network for White Blood Cell Classification with LIME-Based Explainability [0.0]
White blood cells (WBC) are important parts of our immune system.
Traditional convolutional neural networks (CNN) can classify the blood cell from a part of an object and perform object recognition.
We propose a novel ensemble approach that integrates three CNN architectures, each uniquely configured with different dropout and max-pooling layer settings.
arXiv Detail & Related papers (2025-02-08T05:53:20Z) - Semi-supervised Semantic Segmentation for Remote Sensing Images via Multi-scale Uncertainty Consistency and Cross-Teacher-Student Attention [59.19580789952102]
This paper proposes a novel semi-supervised Multi-Scale Uncertainty and Cross-Teacher-Student Attention (MUCA) model for RS image semantic segmentation tasks.
MUCA constrains the consistency among feature maps at different layers of the network by introducing a multi-scale uncertainty consistency regularization.
MUCA utilizes a Cross-Teacher-Student attention mechanism to guide the student network, guiding the student network to construct more discriminative feature representations.
arXiv Detail & Related papers (2025-01-18T11:57:20Z) - CAVE: Classifying Abnormalities in Video Capsule Endoscopy [0.1937002985471497]
In this study, we explore an ensemble-based approach to improve classification accuracy in complex image datasets.
We leverage the unique feature-extraction capabilities of each model to enhance the overall accuracy.
Experimental evaluations demonstrate that the ensemble achieves higher accuracy and robustness across challenging and imbalanced classes.
arXiv Detail & Related papers (2024-10-26T17:25:08Z) - Mew: Multiplexed Immunofluorescence Image Analysis through an Efficient Multiplex Network [84.88767228835928]
We introduce Mew, a novel framework designed to efficiently process mIF images through the lens of multiplex network.
Mew innovatively constructs a multiplex network comprising two distinct layers: a Voronoi network for geometric information and a Cell-type network for capturing cell-wise homogeneity.
This framework equips a scalable and efficient Graph Neural Network (GNN), capable of processing the entire graph during training.
arXiv Detail & Related papers (2024-07-25T08:22:30Z) - CDIMC-net: Cognitive Deep Incomplete Multi-view Clustering Network [53.72046586512026]
We propose a novel incomplete multi-view clustering network, called Cognitive Deep Incomplete Multi-view Clustering Network (CDIMC-net)
It captures the high-level features and local structure of each view by incorporating the view-specific deep encoders and graph embedding strategy into a framework.
Based on the human cognition, i.e., learning from easy to hard, it introduces a self-paced strategy to select the most confident samples for model training.
arXiv Detail & Related papers (2024-03-28T15:45:03Z) - Segment Anything for comprehensive analysis of grapevine cluster architecture and berry properties [0.0]
This study demonstrates out-of-the-box SAM's high accuracy in identifying individual berries in 2D cluster images.
The correlation between human-identified berries and SAM predictions was very strong.
arXiv Detail & Related papers (2024-03-19T17:37:18Z) - Edge Detectors Can Make Deep Convolutional Neural Networks More Robust [25.871767605100636]
This paper first employs the edge detectors as layer kernels and designs a binary edge feature branch (BEFB) to learn the binary edge features.
The accuracy of the BEFB integrated models is better than the original ones on all datasets when facing FGSM, PGD, and C&W attacks.
The work in this paper for the first time shows it is feasible to enhance the robustness of DCNNs through combining both shape-like features and texture features.
arXiv Detail & Related papers (2024-02-26T10:54:26Z) - Unveiling the Unseen: Identifiable Clusters in Trained Depthwise
Convolutional Kernels [56.69755544814834]
Recent advances in depthwise-separable convolutional neural networks (DS-CNNs) have led to novel architectures.
This paper reveals another striking property of DS-CNN architectures: discernible and explainable patterns emerge in their trained depthwise convolutional kernels in all layers.
arXiv Detail & Related papers (2024-01-25T19:05:53Z) - A heterogeneous group CNN for image super-resolution [127.2132400582117]
Convolutional neural networks (CNNs) have obtained remarkable performance via deep architectures.
We present a heterogeneous group SR CNN (HGSRCNN) via leveraging structure information of different types to obtain a high-quality image.
arXiv Detail & Related papers (2022-09-26T04:14:59Z) - Two-Stream Graph Convolutional Network for Intra-oral Scanner Image
Segmentation [133.02190910009384]
We propose a two-stream graph convolutional network (i.e., TSGCN) to handle inter-view confusion between different raw attributes.
Our TSGCN significantly outperforms state-of-the-art methods in 3D tooth (surface) segmentation.
arXiv Detail & Related papers (2022-04-19T10:41:09Z) - A Hierarchical Block Distance Model for Ultra Low-Dimensional Graph
Representations [0.0]
This paper proposes a novel scalable graph representation learning method named the Block Distance Model (HBDM)
HBDM accounts for homophily and transitivity by accurately approximating the latent distance model (LDM) throughout the hierarchy.
We evaluate the performance of the HBDM on massive networks consisting of millions of nodes.
arXiv Detail & Related papers (2022-04-12T15:23:12Z) - An Efficient Cervical Whole Slide Image Analysis Framework Based on
Multi-scale Semantic and Spatial Features using Deep Learning [2.7218168309244652]
This study designs a novel inline connection network (InCNet) by enriching the multi-scale connectivity to build the lightweight model named You Only Look Cytopathology Once (YOLCO)
The proposed model allows the input size enlarged to megapixel that can stitch the WSI without any overlap by the average repeats.
Based on Transformer for classifying the integrated multi-scale multi-task features, the experimental results appear $0.872$ AUC score better and $2.51times$ faster than the best conventional method in WSI classification.
arXiv Detail & Related papers (2021-06-29T06:24:55Z) - CARAFE++: Unified Content-Aware ReAssembly of FEatures [132.49582482421246]
We propose unified Content-Aware ReAssembly of FEatures (CARAFE++), a universal, lightweight and highly effective operator to fulfill this goal.
CARAFE++ generates adaptive kernels on-the-fly to enable instance-specific content-aware handling.
It shows consistent and substantial gains across all the tasks with negligible computational overhead.
arXiv Detail & Related papers (2020-12-07T07:34:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.