Efficient Neural Net Approaches in Metal Casting Defect Detection
- URL: http://arxiv.org/abs/2208.04150v1
- Date: Mon, 8 Aug 2022 13:54:36 GMT
- Title: Efficient Neural Net Approaches in Metal Casting Defect Detection
- Authors: Rohit Lal, Bharath Kumar Bolla, Sabeesh Ethiraj
- Abstract summary: This research proposes a lightweight architecture that is efficient in terms of accuracy and inference time.
Our results indicate that a custom model of 590K parameters with depth-wise separable convolutions outperformed pretrained architectures.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: One of the most pressing challenges prevalent in the steel manufacturing
industry is the identification of surface defects. Early identification of
casting defects can help boost performance, including streamlining production
processes. Though, deep learning models have helped bridge this gap and
automate most of these processes, there is a dire need to come up with
lightweight models that can be deployed easily with faster inference times.
This research proposes a lightweight architecture that is efficient in terms of
accuracy and inference time compared with sophisticated pre-trained CNN
architectures like MobileNet, Inception, and ResNet, including vision
transformers. Methodologies to minimize computational requirements such as
depth-wise separable convolution and global average pooling (GAP) layer,
including techniques that improve architectural efficiencies and augmentations,
have been experimented. Our results indicate that a custom model of 590K
parameters with depth-wise separable convolutions outperformed pretrained
architectures such as Resnet and Vision transformers in terms of accuracy
(81.87%) and comfortably outdid architectures such as Resnet, Inception, and
Vision transformers in terms of faster inference times (12 ms). Blurpool fared
outperformed other techniques, with an accuracy of 83.98%. Augmentations had a
paradoxical effect on the model performance. No direct correlation between
depth-wise and 3x3 convolutions on inference time, they, however, they played a
direct role in improving model efficiency by enabling the networks to go deeper
and by decreasing the number of trainable parameters. Our work sheds light on
the fact that custom networks with efficient architectures and faster inference
times can be built without the need of relying on pre-trained architectures.
Related papers
- Task-Oriented Real-time Visual Inference for IoVT Systems: A Co-design Framework of Neural Networks and Edge Deployment [61.20689382879937]
Task-oriented edge computing addresses this by shifting data analysis to the edge.
Existing methods struggle to balance high model performance with low resource consumption.
We propose a novel co-design framework to optimize neural network architecture.
arXiv Detail & Related papers (2024-10-29T19:02:54Z) - TurboViT: Generating Fast Vision Transformers via Generative
Architecture Search [74.24393546346974]
Vision transformers have shown unprecedented levels of performance in tackling various visual perception tasks in recent years.
There has been significant research recently on the design of efficient vision transformer architecture.
In this study, we explore the generation of fast vision transformer architecture designs via generative architecture search.
arXiv Detail & Related papers (2023-08-22T13:08:29Z) - Efficient Deep Learning Methods for Identification of Defective Casting
Products [0.0]
In this paper, we have compared and contrasted various pre-trained and custom-built AI architectures.
Our results show that custom architectures are efficient than pre-trained mobile architectures.
Augmentation experimentations have also been carried out on the custom architectures to make the models more robust and generalizable.
arXiv Detail & Related papers (2022-05-14T19:35:05Z) - SmoothNets: Optimizing CNN architecture design for differentially
private deep learning [69.10072367807095]
DPSGD requires clipping and noising of per-sample gradients.
This introduces a reduction in model utility compared to non-private training.
We distilled a new model architecture termed SmoothNet, which is characterised by increased robustness to the challenges of DP-SGD training.
arXiv Detail & Related papers (2022-05-09T07:51:54Z) - The Untapped Potential of Off-the-Shelf Convolutional Neural Networks [29.205446247063673]
We show that existing off-the-shelf models like ResNet-50 are capable of over 95% accuracy on ImageNet.
This level of performance currently exceeds that of models with over 20x more parameters and significantly more complex training procedures.
arXiv Detail & Related papers (2021-03-17T20:04:46Z) - An Efficient Quantitative Approach for Optimizing Convolutional Neural
Networks [16.072287925319806]
We propose 3D-Receptive Field (3DRF) to estimate the quality of a CNN architecture and guide the search process of designs.
Our models can achieve up to 5.47% accuracy improvement and up to 65.38% parameters, compared with state-of-the-art CNN structures like MobileNet and ResNet.
arXiv Detail & Related papers (2020-09-11T05:14:34Z) - A Compact Deep Architecture for Real-time Saliency Prediction [42.58396452892243]
Saliency models aim to imitate the attention mechanism in the human visual system.
Deep models have a high number of parameters which makes them less suitable for real-time applications.
Here we propose a compact yet fast model for real-time saliency prediction.
arXiv Detail & Related papers (2020-08-30T17:47:16Z) - Towards Practical Lipreading with Distilled and Efficient Models [57.41253104365274]
Lipreading has witnessed a lot of progress due to the resurgence of neural networks.
Recent works have placed emphasis on aspects such as improving performance by finding the optimal architecture or improving generalization.
There is still a significant gap between the current methodologies and the requirements for an effective deployment of lipreading in practical scenarios.
We propose a series of innovations that significantly bridge that gap: first, we raise the state-of-the-art performance by a wide margin on LRW and LRW-1000 to 88.5% and 46.6%, respectively using self-distillation.
arXiv Detail & Related papers (2020-07-13T16:56:27Z) - Making DensePose fast and light [78.49552144907513]
Existing neural network models capable of solving this task are heavily parameterized.
To enable Dense Pose inference on the end device with current models, one needs to support an expensive server-side infrastructure and have a stable internet connection.
In this work, we target the problem of redesigning the DensePose R-CNN model's architecture so that the final network retains most of its accuracy but becomes more light-weight and fast.
arXiv Detail & Related papers (2020-06-26T19:42:20Z) - FBNetV3: Joint Architecture-Recipe Search using Predictor Pretraining [65.39532971991778]
We present an accuracy predictor that scores architecture and training recipes jointly, guiding both sample selection and ranking.
We run fast evolutionary searches in just CPU minutes to generate architecture-recipe pairs for a variety of resource constraints.
FBNetV3 makes up a family of state-of-the-art compact neural networks that outperform both automatically and manually-designed competitors.
arXiv Detail & Related papers (2020-06-03T05:20:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.