Distinguishing artefacts: evaluating the saturation point of
convolutional neural networks
- URL: http://arxiv.org/abs/2105.10448v1
- Date: Fri, 21 May 2021 16:33:20 GMT
- Title: Distinguishing artefacts: evaluating the saturation point of
convolutional neural networks
- Authors: Ric Real, James Gopsill, David Jones, Chris Snider, Ben Hicks
- Abstract summary: This paper presents a method for generating synthetic image data sets from online CAD model repositories.
1,000 CAD models were curated and processed to generate large scale surrogate data sets, featuring model coverage at steps of 10$circ$, 30$circ$, 60$circ$, and 120$circ$ degrees.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Prior work has shown Convolutional Neural Networks (CNNs) trained on
surrogate Computer Aided Design (CAD) models are able to detect and classify
real-world artefacts from photographs. The applications of which support
twinning of digital and physical assets in design, including rapid extraction
of part geometry from model repositories, information search \& retrieval and
identifying components in the field for maintenance, repair, and recording. The
performance of CNNs in classification tasks have been shown dependent on
training data set size and number of classes. Where prior works have used
relatively small surrogate model data sets ($<100$ models), the question
remains as to the ability of a CNN to differentiate between models in
increasingly large model repositories. This paper presents a method for
generating synthetic image data sets from online CAD model repositories, and
further investigates the capacity of an off-the-shelf CNN architecture trained
on synthetic data to classify models as class size increases. 1,000 CAD models
were curated and processed to generate large scale surrogate data sets,
featuring model coverage at steps of 10$^{\circ}$, 30$^{\circ}$, 60$^{\circ}$,
and 120$^{\circ}$ degrees. The findings demonstrate the capability of computer
vision algorithms to classify artefacts in model repositories of up to 200,
beyond this point the CNN's performance is observed to deteriorate
significantly, limiting its present ability for automated twinning of physical
to digital artefacts. Although, a match is more often found in the top-5
results showing potential for information search and retrieval on large
repositories of surrogate models.
Related papers
- Modeling & Evaluating the Performance of Convolutional Neural Networks for Classifying Steel Surface Defects [0.0]
Recently, outstanding identification rates in image classification tasks were achieved by convolutional neural networks (CNNs)
DenseNet201 had the greatest detection rate on the NEU dataset, falling in at 98.37 percent.
arXiv Detail & Related papers (2024-06-19T08:14:50Z) - Towards Scalable and Versatile Weight Space Learning [51.78426981947659]
This paper introduces the SANE approach to weight-space learning.
Our method extends the idea of hyper-representations towards sequential processing of subsets of neural network weights.
arXiv Detail & Related papers (2024-06-14T13:12:07Z) - Stealing the Invisible: Unveiling Pre-Trained CNN Models through
Adversarial Examples and Timing Side-Channels [14.222432788661914]
We present an approach based on the observation that the classification patterns of adversarial images can be used as a means to steal the models.
Our approach exploits varying misclassifications of adversarial images across different models to fingerprint several renowned Convolutional Neural Network (CNN) and Vision Transformer (ViT) architectures.
arXiv Detail & Related papers (2024-02-19T08:47:20Z) - NLLB-CLIP -- train performant multilingual image retrieval model on a
budget [65.268245109828]
We present NLLB-CLIP - CLIP model with a text encoder from the NLLB model.
We used an automatically created dataset of 106,246 good-quality images with captions in 201 languages.
We show that NLLB-CLIP is comparable in quality to state-of-the-art models and significantly outperforms them on low-resource languages.
arXiv Detail & Related papers (2023-09-04T23:26:11Z) - Revisiting Classifier: Transferring Vision-Language Models for Video
Recognition [102.93524173258487]
Transferring knowledge from task-agnostic pre-trained deep models for downstream tasks is an important topic in computer vision research.
In this study, we focus on transferring knowledge for video classification tasks.
We utilize the well-pretrained language model to generate good semantic target for efficient transferring learning.
arXiv Detail & Related papers (2022-07-04T10:00:47Z) - CONVIQT: Contrastive Video Quality Estimator [63.749184706461826]
Perceptual video quality assessment (VQA) is an integral component of many streaming and video sharing platforms.
Here we consider the problem of learning perceptually relevant video quality representations in a self-supervised manner.
Our results indicate that compelling representations with perceptual bearing can be obtained using self-supervised learning.
arXiv Detail & Related papers (2022-06-29T15:22:01Z) - Classification of EEG Motor Imagery Using Deep Learning for
Brain-Computer Interface Systems [79.58173794910631]
A trained T1 class Convolutional Neural Network (CNN) model will be used to examine its ability to successfully identify motor imagery.
In theory, and if the model has been trained accurately, it should be able to identify a class and label it accordingly.
The CNN model will then be restored and used to try and identify the same class of motor imagery data using much smaller sampled data.
arXiv Detail & Related papers (2022-05-31T17:09:46Z) - Real-time Human Detection Model for Edge Devices [0.0]
Convolutional Neural Networks (CNNs) have replaced traditional feature extraction and machine learning models in detection and classification tasks.
Lightweight CNN models have been recently introduced for real-time tasks.
This paper suggests a CNN-based lightweight model that can fit on a limited edge device such as Raspberry Pi.
arXiv Detail & Related papers (2021-11-20T18:42:17Z) - A Convolutional Neural Network Approach to the Classification of
Engineering Models [0.9558392439655015]
This paper presents a deep learning approach for the classification of Engineering (CAD) models using Convolutional Neural Networks (CNNs)
It is proposed to use a residual network architecture for CADNET, inspired by the popular ResNet.
The LFD-based CNN approach using the proposed network architecture, along with gradient boosting yielded the best classification accuracy on CADNET.
arXiv Detail & Related papers (2021-07-14T04:33:50Z) - Closed-form Continuous-Depth Models [99.40335716948101]
Continuous-depth neural models rely on advanced numerical differential equation solvers.
We present a new family of models, termed Closed-form Continuous-depth (CfC) networks, that are simple to describe and at least one order of magnitude faster.
arXiv Detail & Related papers (2021-06-25T22:08:51Z) - Multi-task pre-training of deep neural networks for digital pathology [8.74883469030132]
We first assemble and transform many digital pathology datasets into a pool of 22 classification tasks and almost 900k images.
We show that our models used as feature extractors either improve significantly over ImageNet pre-trained models or provide comparable performance.
arXiv Detail & Related papers (2020-05-05T08:50:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.