Adinkra Symbol Recognition using Classical Machine Learning and Deep
Learning
- URL: http://arxiv.org/abs/2311.15728v1
- Date: Mon, 27 Nov 2023 11:26:41 GMT
- Title: Adinkra Symbol Recognition using Classical Machine Learning and Deep
Learning
- Authors: Michael Adjeisah, Kwame Omono Asamoah, Martha Asamoah Yeboah, Raji
Rafiu King, Godwin Ferguson Achaab and Kingsley Adjei
- Abstract summary: We build a CNN model for classification and recognition using six convolutional layers, three fully connected layers, and optional dropout regularization.
We assess the model's performance by measuring its accuracy and convergence rate.
We hope this application inspires ideas on the various uses of AI in organizing our traditional and modern lives.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Artificial intelligence (AI) has emerged as a transformative influence,
engendering paradigm shifts in global societies, spanning academia and
industry. However, in light of these rapid advances, addressing the
underrepresentation of black communities and African countries in AI is
crucial. Boosting enthusiasm for AI can be effectively accomplished by
showcasing straightforward applications around tasks like identifying and
categorizing traditional symbols, such as Adinkra symbols, or familiar objects
within the community. In this research endeavor, we dived into classical
machine learning and harnessed the power of deep learning models to tackle the
intricate task of classifying and recognizing Adinkra symbols. The idea led to
a newly constructed ADINKRA dataset comprising 174,338 images meticulously
organized into 62 distinct classes, each representing a singular and emblematic
symbol. We constructed a CNN model for classification and recognition using six
convolutional layers, three fully connected (FC) layers, and optional dropout
regularization. The model is a simpler and smaller version of VGG, with fewer
layers, smaller channel sizes, and a fixed kernel size. Additionally, we tap
into the transfer learning capabilities provided by pre-trained models like VGG
and ResNet. These models assist us in both classifying images and extracting
features that can be used with classical machine learning models. We assess the
model's performance by measuring its accuracy and convergence rate and
visualizing the areas that significantly influence its predictions. These
evaluations serve as a foundational benchmark for future assessments of the
ADINKRA dataset. We hope this application exemplar inspires ideas on the
various uses of AI in organizing our traditional and modern lives.
Related papers
- Implicit Neural Representation Facilitates Unified Universal Vision Encoding [11.947746726150001]
A first-of-its-kind model learns representations which are simultaneously useful for recognition and generation.<n>We train our model as a hyper-network for implicit neural representation, which learns to map images to model weights for fast, accurate reconstruction.<n>The model also learns an unprecedented compressed embedding space with outstanding performance for various visual tasks.
arXiv Detail & Related papers (2026-01-20T18:59:57Z) - UNIFORM: Unifying Knowledge from Large-scale and Diverse Pre-trained Models [62.76435672183968]
We introduce a novel framework, namely UNIFORM, for knowledge transfer from a diverse set of off-the-shelf models into one student model.<n>We propose a dedicated voting mechanism to capture the consensus of knowledge both at the logit level and at the feature level.<n>Experiments demonstrate that UNIFORM effectively enhances unsupervised object recognition performance compared to strong knowledge transfer baselines.
arXiv Detail & Related papers (2025-08-27T00:56:11Z) - Few-Shot Class-Incremental Model Attribution Using Learnable Representation From CLIP-ViT Features [1.534667887016089]
This work proposes a new strategy to deal with persistently emerging generative models.
We adapt few-shot class-incremental learning (FSCIL) mechanisms for MA problem to uncover novel generative AI models.
To learn an effective representation, we propose Adaptive Integration Module (AIM) to calculate a weighted sum of CLIP-ViT block features for each image.
arXiv Detail & Related papers (2025-03-11T08:05:26Z) - Shortcut Learning Susceptibility in Vision Classifiers [3.004632712148892]
Shortcut learning is where machine learning models exploit spurious correlations in data instead of capturing meaningful features.
This phenomenon is prevalent across various machine learning applications, including vision, natural language processing, and speech recognition.
We systematically evaluate these architectures by introducing deliberate shortcuts into the dataset that are positionally correlated with class labels.
arXiv Detail & Related papers (2025-02-13T10:25:52Z) - A Survey on All-in-One Image Restoration: Taxonomy, Evaluation and Future Trends [67.43992456058541]
Image restoration (IR) refers to the process of improving visual quality of images while removing degradation, such as noise, blur, weather effects, and so on.
Traditional IR methods typically target specific types of degradation, which limits their effectiveness in real-world scenarios with complex distortions.
The all-in-one image restoration (AiOIR) paradigm has emerged, offering a unified framework that adeptly addresses multiple degradation types.
arXiv Detail & Related papers (2024-10-19T11:11:09Z) - Generalized Robot 3D Vision-Language Model with Fast Rendering and Pre-Training Vision-Language Alignment [55.11291053011696]
This work presents a framework for dealing with 3D scene understanding when the labeled scenes are quite limited.
To extract knowledge for novel categories from the pre-trained vision-language models, we propose a hierarchical feature-aligned pre-training and knowledge distillation strategy.
In the limited reconstruction case, our proposed approach, termed WS3D++, ranks 1st on the large-scale ScanNet benchmark.
arXiv Detail & Related papers (2023-12-01T15:47:04Z) - Knowledge-Aware Prompt Tuning for Generalizable Vision-Language Models [64.24227572048075]
We propose a Knowledge-Aware Prompt Tuning (KAPT) framework for vision-language models.
Our approach takes inspiration from human intelligence in which external knowledge is usually incorporated into recognizing novel categories of objects.
arXiv Detail & Related papers (2023-08-22T04:24:45Z) - Dense Sample Deep Learning [0.0]
Despite the growing use of Deep Learning (DL) networks, little is actually understood about the learning mechanisms and representations.
In this paper we explore these questions with a large (1.24M weights; VGG) DL in a novel high density sample task.
We harvest some basic observations of the learning dynamics of DL and propose a new theory of complex feature construction based on our results.
arXiv Detail & Related papers (2023-07-20T16:21:14Z) - Recognizing Unseen Objects via Multimodal Intensive Knowledge Graph
Propagation [68.13453771001522]
We propose a multimodal intensive ZSL framework that matches regions of images with corresponding semantic embeddings.
We conduct extensive experiments and evaluate our model on large-scale real-world data.
arXiv Detail & Related papers (2023-06-14T13:07:48Z) - Few-Shot Class-Incremental Learning from an Open-Set Perspective [10.898784938875702]
We explore the important task of Few-Shot Class-Incremental Learning (FSCIL) and its extreme data scarcity condition of one-shot.
In ALICE, instead of the commonly used cross-entropy loss, we propose to use the angular penalty loss to obtain well-clustered features.
Experiments on benchmark datasets, including CIFAR100, miniImageNet, and CUB200, demonstrate the improved performance of ALICE.
arXiv Detail & Related papers (2022-07-30T05:42:48Z) - Class-Specific Semantic Reconstruction for Open Set Recognition [101.24781422480406]
Open set recognition enables deep neural networks (DNNs) to identify samples of unknown classes.
We propose a novel method, called Class-Specific Semantic Reconstruction (CSSR), that integrates the power of auto-encoder (AE) and prototype learning.
Results of experiments conducted on multiple datasets show that the proposed method achieves outstanding performance in both close and open set recognition.
arXiv Detail & Related papers (2022-07-05T16:25:34Z) - WenLan 2.0: Make AI Imagine via a Multimodal Foundation Model [74.4875156387271]
We develop a novel foundation model pre-trained with huge multimodal (visual and textual) data.
We show that state-of-the-art results can be obtained on a wide range of downstream tasks.
arXiv Detail & Related papers (2021-10-27T12:25:21Z) - Multi-Branch Deep Radial Basis Function Networks for Facial Emotion
Recognition [80.35852245488043]
We propose a CNN based architecture enhanced with multiple branches formed by radial basis function (RBF) units.
RBF units capture local patterns shared by similar instances using an intermediate representation.
We show it is the incorporation of local information what makes the proposed model competitive.
arXiv Detail & Related papers (2021-09-07T21:05:56Z) - BENDR: using transformers and a contrastive self-supervised learning
task to learn from massive amounts of EEG data [15.71234837305808]
We consider how to adapt techniques and architectures used for language modelling (LM) to encephalography modelling (EM)
We find that a single pre-trained model is capable of modelling completely novel raw EEG sequences recorded with differing hardware.
Both the internal representations of this model and the entire architecture can be fine-tuned to a variety of downstream BCI and EEG classification tasks.
arXiv Detail & Related papers (2021-01-28T14:54:01Z) - Eigen-CAM: Class Activation Map using Principal Components [1.2691047660244335]
This paper builds on previous ideas to cope with the increasing demand for interpretable, robust, and transparent models.
The proposed Eigen-CAM computes and visualizes the principle components of the learned features/representations from the convolutional layers.
arXiv Detail & Related papers (2020-08-01T17:14:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.