CNN-based Methods for Object Recognition with High-Resolution Tactile
Sensors
- URL: http://arxiv.org/abs/2305.12417v1
- Date: Sun, 21 May 2023 09:54:12 GMT
- Title: CNN-based Methods for Object Recognition with High-Resolution Tactile
Sensors
- Authors: Juan M. Gandarias (1), Alfonso J. Garc\'ia-Cerezo (1), Jes\'us M.
G\'omez-de-Gabriel (1) ((1) Robotics and Mechatronics, Systems Engineering
and Automation Department, University of Malaga)
- Abstract summary: A high-resolution tactile sensor has been attached to a robotic end-effector to identify contacted objects.
Two CNN-based approaches have been employed to classify pressure images.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Novel high-resolution pressure-sensor arrays allow treating pressure readings
as standard images. Computer vision algorithms and methods such as
Convolutional Neural Networks (CNN) can be used to identify contact objects. In
this paper, a high-resolution tactile sensor has been attached to a robotic
end-effector to identify contacted objects. Two CNN-based approaches have been
employed to classify pressure images. These methods include a transfer learning
approach using a pre-trained CNN on an RGB-images dataset and a custom-made CNN
(TactNet) trained from scratch with tactile information. The transfer learning
approach can be carried out by retraining the classification layers of the
network or replacing these layers with an SVM. Overall, 11 configurations based
on these methods have been tested: 8 transfer learning-based, and 3
TactNet-based. Moreover, a study of the performance of the methods and a
comparative discussion with the current state-of-the-art on tactile object
recognition is presented.
Related papers
- Classification and regression of trajectories rendered as images via 2D Convolutional Neural Networks [0.0]
Recent advances in computer vision have facilitated the processing of trajectories rendered as images via artificial neural networks with 2d convolutional layers (CNNs)
In this study, we investigate the effectiveness of CNNs for solving classification and regression problems from synthetic trajectories rendered as images using different modalities.
Results highlight the importance of choosing an appropriate image resolution according to model depth and motion history in applications where movement direction is critical.
arXiv Detail & Related papers (2024-09-27T15:27:04Z) - Alleviating Catastrophic Forgetting in Facial Expression Recognition with Emotion-Centered Models [49.3179290313959]
The proposed method, emotion-centered generative replay (ECgr), tackles this challenge by integrating synthetic images from generative adversarial networks.
ECgr incorporates a quality assurance algorithm to ensure the fidelity of generated images.
The experimental results on four diverse facial expression datasets demonstrate that incorporating images generated by our pseudo-rehearsal method enhances training on the targeted dataset and the source dataset.
arXiv Detail & Related papers (2024-04-18T15:28:34Z) - T-TAME: Trainable Attention Mechanism for Explaining Convolutional
Networks and Vision Transformers [9.284740716447342]
"Black box" nature of neural networks is a barrier to adoption in applications where explainability is essential.
This paper presents T-TAME, Transformer-compatible Trainable Attention Mechanism for Explanations.
Proposed architecture and training technique can be easily applied to any convolutional or Vision Transformer-like neural network.
arXiv Detail & Related papers (2024-03-07T14:25:03Z) - Visual Recognition with Deep Nearest Centroids [57.35144702563746]
We devise deep nearest centroids (DNC), a conceptually elegant yet surprisingly effective network for large-scale visual recognition.
Compared with parametric counterparts, DNC performs better on image classification (CIFAR-10, ImageNet) and greatly boots pixel recognition (ADE20K, Cityscapes)
arXiv Detail & Related papers (2022-09-15T15:47:31Z) - ECLAD: Extracting Concepts with Local Aggregated Descriptors [6.470466745237234]
We propose a novel method for automatic concept extraction and localization based on representations obtained through pixel-wise aggregations of CNN activation maps.
We introduce a process for the validation of concept-extraction techniques based on synthetic datasets with pixel-wise annotations of their main components.
arXiv Detail & Related papers (2022-06-09T14:25:23Z) - Classification of EEG Motor Imagery Using Deep Learning for
Brain-Computer Interface Systems [79.58173794910631]
A trained T1 class Convolutional Neural Network (CNN) model will be used to examine its ability to successfully identify motor imagery.
In theory, and if the model has been trained accurately, it should be able to identify a class and label it accordingly.
The CNN model will then be restored and used to try and identify the same class of motor imagery data using much smaller sampled data.
arXiv Detail & Related papers (2022-05-31T17:09:46Z) - Learning to Synthesize Volumetric Meshes from Vision-based Tactile
Imprints [26.118805500471066]
Vision-based tactile sensors typically utilize a deformable elastomer and a camera mounted above to provide high-resolution image observations of contacts.
This paper focuses on learning to synthesize the mesh of the elastomer based on the image imprints acquired from vision-based tactile sensors.
A graph neural network (GNN) is introduced to learn the image-to-mesh mappings with supervised learning.
arXiv Detail & Related papers (2022-03-29T00:24:10Z) - Knowledge Distillation By Sparse Representation Matching [107.87219371697063]
We propose Sparse Representation Matching (SRM) to transfer intermediate knowledge from one Convolutional Network (CNN) to another by utilizing sparse representation.
We formulate as a neural processing block, which can be efficiently optimized using gradient descent and integrated into any CNN in a plug-and-play manner.
Our experiments demonstrate that is robust to architectural differences between the teacher and student networks, and outperforms other KD techniques across several datasets.
arXiv Detail & Related papers (2021-03-31T11:47:47Z) - Improving Object Detection in Art Images Using Only Style Transfer [5.156484100374058]
We propose and evaluate a process for training neural networks to localize objects - specifically people - in art images.
We generate a large dataset for training and validation by modifying the images in the COCO dataset using AdaIn style transfer.
The result is a significant improvement on the state of the art and a new way forward for creating datasets to train neural networks to process art images.
arXiv Detail & Related papers (2021-02-12T13:48:46Z) - Emotional EEG Classification using Connectivity Features and
Convolutional Neural Networks [81.74442855155843]
We introduce a new classification system that utilizes brain connectivity with a CNN and validate its effectiveness via the emotional video classification.
The level of concentration of the brain connectivity related to the emotional property of the target video is correlated with classification performance.
arXiv Detail & Related papers (2021-01-18T13:28:08Z) - Ventral-Dorsal Neural Networks: Object Detection via Selective Attention [51.79577908317031]
We propose a new framework called Ventral-Dorsal Networks (VDNets)
Inspired by the structure of the human visual system, we propose the integration of a "Ventral Network" and a "Dorsal Network"
Our experimental results reveal that the proposed method outperforms state-of-the-art object detection approaches.
arXiv Detail & Related papers (2020-05-15T23:57:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.