Lightweight Face Recognition: An Improved MobileFaceNet Model
- URL: http://arxiv.org/abs/2311.15326v1
- Date: Sun, 26 Nov 2023 15:01:00 GMT
- Title: Lightweight Face Recognition: An Improved MobileFaceNet Model
- Authors: Ahmad Hassanpour, Yasamin Kowsari
- Abstract summary: This paper focuses on MobileFaceNet and its modified variant, MMobileFaceNet.
The need for efficient FR models on devices with limited computational resources has led to the development of models with reduced memory footprints and computational demands without sacrificing accuracy.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper presents an extensive exploration and comparative analysis of
lightweight face recognition (FR) models, specifically focusing on
MobileFaceNet and its modified variant, MMobileFaceNet. The need for efficient
FR models on devices with limited computational resources has led to the
development of models with reduced memory footprints and computational demands
without sacrificing accuracy. Our research delves into the impact of dataset
selection, model architecture, and optimization algorithms on the performance
of FR models. We highlight our participation in the EFaR-2023 competition,
where our models showcased exceptional performance, particularly in categories
restricted by the number of parameters. By employing a subset of the Webface42M
dataset and integrating sharpness-aware minimization (SAM) optimization, we
achieved significant improvements in accuracy across various benchmarks,
including those that test for cross-pose, cross-age, and cross-ethnicity
performance. The results underscore the efficacy of our approach in crafting
models that are not only computationally efficient but also maintain high
accuracy in diverse conditions.
Related papers
- Task-Oriented Real-time Visual Inference for IoVT Systems: A Co-design Framework of Neural Networks and Edge Deployment [61.20689382879937]
Task-oriented edge computing addresses this by shifting data analysis to the edge.
Existing methods struggle to balance high model performance with low resource consumption.
We propose a novel co-design framework to optimize neural network architecture.
arXiv Detail & Related papers (2024-10-29T19:02:54Z) - A-SDM: Accelerating Stable Diffusion through Redundancy Removal and
Performance Optimization [54.113083217869516]
In this work, we first explore the computational redundancy part of the network.
We then prune the redundancy blocks of the model and maintain the network performance.
Thirdly, we propose a global-regional interactive (GRI) attention to speed up the computationally intensive attention part.
arXiv Detail & Related papers (2023-12-24T15:37:47Z) - Developing a Resource-Constraint EdgeAI model for Surface Defect
Detection [1.338174941551702]
We propose a lightweight EdgeAI architecture modified from Xception for on-device training in a resource-constraint edge environment.
We evaluate our model on a PCB defect detection task and compare its performance against existing lightweight models.
Our method can be applied to other resource-constraint applications while maintaining significant performance.
arXiv Detail & Related papers (2023-12-04T15:28:31Z) - Fairer and More Accurate Tabular Models Through NAS [14.147928131445852]
We propose using multi-objective Neural Architecture Search (NAS) and Hyperparameter Optimization (HPO) in the first application to the very challenging domain of tabular data.
We show that models optimized solely for accuracy with NAS often fail to inherently address fairness concerns.
We produce architectures that consistently dominate state-of-the-art bias mitigation methods either in fairness, accuracy or both.
arXiv Detail & Related papers (2023-10-18T17:56:24Z) - CATfOOD: Counterfactual Augmented Training for Improving Out-of-Domain
Performance and Calibration [59.48235003469116]
We show that data augmentation consistently enhances OOD performance.
We also show that CF augmented models which are easier to calibrate also exhibit much lower entropy when assigning importance.
arXiv Detail & Related papers (2023-09-14T16:16:40Z) - Slimmable Domain Adaptation [112.19652651687402]
We introduce a simple framework, Slimmable Domain Adaptation, to improve cross-domain generalization with a weight-sharing model bank.
Our framework surpasses other competing approaches by a very large margin on multiple benchmarks.
arXiv Detail & Related papers (2022-06-14T06:28:04Z) - Real-time Human Detection Model for Edge Devices [0.0]
Convolutional Neural Networks (CNNs) have replaced traditional feature extraction and machine learning models in detection and classification tasks.
Lightweight CNN models have been recently introduced for real-time tasks.
This paper suggests a CNN-based lightweight model that can fit on a limited edge device such as Raspberry Pi.
arXiv Detail & Related papers (2021-11-20T18:42:17Z) - Modeling the Second Player in Distributionally Robust Optimization [90.25995710696425]
We argue for the use of neural generative models to characterize the worst-case distribution.
This approach poses a number of implementation and optimization challenges.
We find that the proposed approach yields models that are more robust than comparable baselines.
arXiv Detail & Related papers (2021-03-18T14:26:26Z) - Real-time Denoising and Dereverberation with Tiny Recurrent U-Net [12.533488149023025]
We propose Tiny Recurrent U-Net (TRU-Net), a lightweight online inference model that matches the performance of current state-of-the-art models.
The size of the quantized version of TRU-Net is 362 kilobytes, which is small enough to be deployed on edge devices.
Results of both objective and subjective evaluations have shown that our model can achieve competitive performance with the current state-of-the-art models.
arXiv Detail & Related papers (2021-02-05T14:46:41Z) - Models, Pixels, and Rewards: Evaluating Design Trade-offs in Visual
Model-Based Reinforcement Learning [109.74041512359476]
We study a number of design decisions for the predictive model in visual MBRL algorithms.
We find that a range of design decisions that are often considered crucial, such as the use of latent spaces, have little effect on task performance.
We show how this phenomenon is related to exploration and how some of the lower-scoring models on standard benchmarks will perform the same as the best-performing models when trained on the same training data.
arXiv Detail & Related papers (2020-12-08T18:03:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.