The Larger The Fairer? Small Neural Networks Can Achieve Fairness for
Edge Devices
- URL: http://arxiv.org/abs/2202.11317v1
- Date: Wed, 23 Feb 2022 05:26:22 GMT
- Title: The Larger The Fairer? Small Neural Networks Can Achieve Fairness for
Edge Devices
- Authors: Yi Sheng, Junhuan Yang, Yawen Wu, Kevin Mao, Yiyu Shi, Jingtong Hu,
Weiwen Jiang, Lei Yang
- Abstract summary: Fairness concerns gradually emerge in many applications, such as face recognition and mobile medical.
This work proposes a novel Fairness- and Hardware-aware Neural architecture search framework, namely FaHaNa.
We show that FaHaNa can identify a series of neural networks with higher fairness and accuracy on a dermatology dataset.
- Score: 16.159547410954602
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Along with the progress of AI democratization, neural networks are being
deployed more frequently in edge devices for a wide range of applications.
Fairness concerns gradually emerge in many applications, such as face
recognition and mobile medical. One fundamental question arises: what will be
the fairest neural architecture for edge devices? By examining the existing
neural networks, we observe that larger networks typically are fairer. But,
edge devices call for smaller neural architectures to meet hardware
specifications. To address this challenge, this work proposes a novel Fairness-
and Hardware-aware Neural architecture search framework, namely FaHaNa. Coupled
with a model freezing approach, FaHaNa can efficiently search for neural
networks with balanced fairness and accuracy, while guaranteed to meet hardware
specifications. Results show that FaHaNa can identify a series of neural
networks with higher fairness and accuracy on a dermatology dataset. Target
edge devices, FaHaNa finds a neural architecture with slightly higher accuracy,
5.28x smaller size, 15.14% higher fairness score, compared with MobileNetV2;
meanwhile, on Raspberry PI and Odroid XU-4, it achieves 5.75x and 5.79x
speedup.
Related papers
- SqueezerFaceNet: Reducing a Small Face Recognition CNN Even More Via
Filter Pruning [55.84746218227712]
We develop SqueezerFaceNet, a light face recognition network which less than 1M parameters.
We show that it can be further reduced (up to 40%) without an appreciable loss in performance.
arXiv Detail & Related papers (2023-07-20T08:38:50Z) - Fast GraspNeXt: A Fast Self-Attention Neural Network Architecture for
Multi-task Learning in Computer Vision Tasks for Robotic Grasping on the Edge [80.88063189896718]
High architectural and computational complexity can result in poor suitability for deployment on embedded devices.
Fast GraspNeXt is a fast self-attention neural network architecture tailored for embedded multi-task learning in computer vision tasks for robotic grasping.
arXiv Detail & Related papers (2023-04-21T18:07:14Z) - FPGA-optimized Hardware acceleration for Spiking Neural Networks [69.49429223251178]
This work presents the development of a hardware accelerator for an SNN, with off-line training, applied to an image recognition task.
The design targets a Xilinx Artix-7 FPGA, using in total around the 40% of the available hardware resources.
It reduces the classification time by three orders of magnitude, with a small 4.5% impact on the accuracy, if compared to its software, full precision counterpart.
arXiv Detail & Related papers (2022-01-18T13:59:22Z) - CondenseNeXt: An Ultra-Efficient Deep Neural Network for Embedded
Systems [0.0]
A Convolutional Neural Network (CNN) is a class of Deep Neural Network (DNN) widely used in the analysis of visual images captured by an image sensor.
In this paper, we propose a neoteric variant of deep convolutional neural network architecture to ameliorate the performance of existing CNN architectures for real-time inference on embedded systems.
arXiv Detail & Related papers (2021-12-01T18:20:52Z) - Profiling Neural Blocks and Design Spaces for Mobile Neural Architecture
Search [21.48915618572691]
We analyze the neural blocks used to build Once-for-All (MobileNetV3), ProxylessNAS and ResNet families.
We show that searching in the reduced search space generates better accuracy-latency frontiers than searching in the original search spaces.
arXiv Detail & Related papers (2021-09-25T19:34:45Z) - Binary Neural Network for Speaker Verification [13.472791713805762]
This paper focuses on how to apply binary neural networks to the task of speaker verification.
Experiment results show that, after binarizing the Convolutional Neural Network, the ResNet34-based network achieves an EER of around 5%.
arXiv Detail & Related papers (2021-04-06T06:04:57Z) - Neural Architecture Search on ImageNet in Four GPU Hours: A
Theoretically Inspired Perspective [88.39981851247727]
We propose a novel framework called training-free neural architecture search (TE-NAS)
TE-NAS ranks architectures by analyzing the spectrum of the neural tangent kernel (NTK) and the number of linear regions in the input space.
We show that: (1) these two measurements imply the trainability and expressivity of a neural network; (2) they strongly correlate with the network's test accuracy.
arXiv Detail & Related papers (2021-02-23T07:50:44Z) - Benchmarking Quantized Neural Networks on FPGAs with FINN [0.42439262432068253]
Using lower precision comes at the cost of negligible loss in accuracy.
This article aims to assess the impact of mixed-precision when applied to neural networks deployed on FPGAs.
arXiv Detail & Related papers (2021-02-02T06:42:07Z) - AttendNets: Tiny Deep Image Recognition Neural Networks for the Edge via
Visual Attention Condensers [81.17461895644003]
We introduce AttendNets, low-precision, highly compact deep neural networks tailored for on-device image recognition.
AttendNets possess deep self-attention architectures based on visual attention condensers.
Results show AttendNets have significantly lower architectural and computational complexity when compared to several deep neural networks.
arXiv Detail & Related papers (2020-09-30T01:53:17Z) - MS-RANAS: Multi-Scale Resource-Aware Neural Architecture Search [94.80212602202518]
We propose Multi-Scale Resource-Aware Neural Architecture Search (MS-RANAS)
We employ a one-shot architecture search approach in order to obtain a reduced search cost.
We achieve state-of-the-art results in terms of accuracy-speed trade-off.
arXiv Detail & Related papers (2020-09-29T11:56:01Z) - A Power-Efficient Binary-Weight Spiking Neural Network Architecture for
Real-Time Object Classification [1.5291703721641183]
We propose a binary-weight spiking neural network (BW-SNN) hardware architecture for low-power real-time object classification on edge platforms.
This design stores a full neural network on-chip, and hence requires no off-chip bandwidth.
arXiv Detail & Related papers (2020-03-12T11:25:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.