Reinforcement Learning Based Handwritten Digit Recognition with
Two-State Q-Learning
- URL: http://arxiv.org/abs/2007.01193v2
- Date: Mon, 10 Aug 2020 10:17:30 GMT
- Title: Reinforcement Learning Based Handwritten Digit Recognition with
Two-State Q-Learning
- Authors: Abdul Mueed Hafiz, Ghulam Mohiuddin Bhat
- Abstract summary: We present a Hybrid approach based on Deep Learning and Reinforcement Learning.
Q-Learning is used with two Q-states and four actions.
Our approach outperforms other contemporary techniques like AlexNet, CNN-Nearest Neighbor and CNNSupport Vector Machine.
- Score: 1.8782750537161614
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a simple yet efficient Hybrid Classifier based on Deep Learning
and Reinforcement Learning. Q-Learning is used with two Q-states and four
actions. Conventional techniques use feature maps extracted from Convolutional
Neural Networks (CNNs) and include them in the Qstates along with past history.
This leads to difficulties with these approaches as the number of states is
very large number due to high dimensions of the feature maps. Since our method
uses only two Q-states it is simple and has much lesser number of parameters to
optimize and also thus has a straightforward reward function. Also, the
approach uses unexplored actions for image processing vis-a-vis other
contemporary techniques. Three datasets have been used for benchmarking of the
approach. These are the MNIST Digit Image Dataset, the USPS Digit Image Dataset
and the MATLAB Digit Image Dataset. The performance of the proposed hybrid
classifier has been compared with other contemporary techniques like a
well-established Reinforcement Learning Technique, AlexNet, CNN-Nearest
Neighbor Classifier and CNNSupport Vector Machine Classifier. Our approach
outperforms these contemporary hybrid classifiers on all the three datasets
used.
Related papers
- Hybrid CNN Bi-LSTM neural network for Hyperspectral image classification [1.2691047660244332]
This paper proposes a neural network combining 3-D CNN, 2-D CNN and Bi-LSTM.
It could achieve 99.83, 99.98 and 100 percent accuracy using only 30 percent trainable parameters of the state-of-art model in IP, PU and SA datasets respectively.
arXiv Detail & Related papers (2024-02-15T15:46:13Z) - MOCA: Self-supervised Representation Learning by Predicting Masked Online Codebook Assignments [72.6405488990753]
Self-supervised learning can be used for mitigating the greedy needs of Vision Transformer networks.
We propose a single-stage and standalone method, MOCA, which unifies both desired properties.
We achieve new state-of-the-art results on low-shot settings and strong experimental results in various evaluation protocols.
arXiv Detail & Related papers (2023-07-18T15:46:20Z) - Boosting Low-Data Instance Segmentation by Unsupervised Pre-training
with Saliency Prompt [103.58323875748427]
This work offers a novel unsupervised pre-training solution for low-data regimes.
Inspired by the recent success of the Prompting technique, we introduce a new pre-training method that boosts QEIS models.
Experimental results show that our method significantly boosts several QEIS models on three datasets.
arXiv Detail & Related papers (2023-02-02T15:49:03Z) - RankDNN: Learning to Rank for Few-shot Learning [70.49494297554537]
This paper introduces a new few-shot learning pipeline that casts relevance ranking for image retrieval as binary ranking relation classification.
It provides a new perspective on few-shot learning and is complementary to state-of-the-art methods.
arXiv Detail & Related papers (2022-11-28T13:59:31Z) - Focal Sparse Convolutional Networks for 3D Object Detection [121.45950754511021]
We introduce two new modules to enhance the capability of Sparse CNNs.
They are focal sparse convolution (Focals Conv) and its multi-modal variant of focal sparse convolution with fusion.
For the first time, we show that spatially learnable sparsity in sparse convolution is essential for sophisticated 3D object detection.
arXiv Detail & Related papers (2022-04-26T17:34:10Z) - Comparison Analysis of Traditional Machine Learning and Deep Learning
Techniques for Data and Image Classification [62.997667081978825]
The purpose of the study is to analyse and compare the most common machine learning and deep learning techniques used for computer vision 2D object classification tasks.
Firstly, we will present the theoretical background of the Bag of Visual words model and Deep Convolutional Neural Networks (DCNN)
Secondly, we will implement a Bag of Visual Words model, the VGG16 CNN Architecture.
arXiv Detail & Related papers (2022-04-11T11:34:43Z) - A Novel Hand Gesture Detection and Recognition system based on
ensemble-based Convolutional Neural Network [3.5665681694253903]
Detection of hand portion has become a challenging task in computer vision and pattern recognition communities.
Deep learning algorithm like convolutional neural network (CNN) architecture has become a very popular choice for classification tasks.
In this paper, an ensemble of CNN-based approaches is presented to overcome some problems like high variance during prediction, overfitting problem and also prediction errors.
arXiv Detail & Related papers (2022-02-25T06:46:58Z) - Overhead-MNIST: Machine Learning Baselines for Image Classification [0.0]
Twenty-three machine learning algorithms were trained then scored to establish baseline comparison metrics.
The Overhead-MNIST dataset is a collection of satellite images similar in style to the ubiquitous MNIST hand-written digits.
We present results for the overall best performing algorithm as a baseline for edge deployability and future performance improvement.
arXiv Detail & Related papers (2021-07-01T13:30:39Z) - Segment as Points for Efficient Online Multi-Object Tracking and
Segmentation [66.03023110058464]
We propose a highly effective method for learning instance embeddings based on segments by converting the compact image representation to un-ordered 2D point cloud representation.
Our method generates a new tracking-by-points paradigm where discriminative instance embeddings are learned from randomly selected points rather than images.
The resulting online MOTS framework, named PointTrack, surpasses all the state-of-the-art methods by large margins.
arXiv Detail & Related papers (2020-07-03T08:29:35Z) - Image Classification by Reinforcement Learning with Two-State Q-Learning [0.0]
Hybridception is presented which is based on deep learning and reinforcement learning.
The proposed technique uses only two Q-states it is straightforward and has much lesser number of optimization parameters.
arXiv Detail & Related papers (2020-06-28T14:54:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.