LRCN-RetailNet: A recurrent neural network architecture for accurate
people counting
- URL: http://arxiv.org/abs/2004.09672v2
- Date: Tue, 12 May 2020 16:21:12 GMT
- Title: LRCN-RetailNet: A recurrent neural network architecture for accurate
people counting
- Authors: Lucas Massa, Adriano Barbosa, Krerley Oliveira, Thales Vieira
- Abstract summary: We introduce LRCN-RetailNet: a recurrent neural network architecture capable of learning a non-linear regression model.
We show that, through a supervised learning approach, the trained models are capable of predicting the people count with high accuracy.
- Score: 4.731404257629232
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Measuring and analyzing the flow of customers in retail stores is essential
for a retailer to better comprehend customers' behavior and support
decision-making. Nevertheless, not much attention has been given to the
development of novel technologies for automatic people counting. We introduce
LRCN-RetailNet: a recurrent neural network architecture capable of learning a
non-linear regression model and accurately predicting the people count from
videos captured by low-cost surveillance cameras. The input video format
follows the recently proposed RGBP image format, which is comprised of color
and people (foreground) information. Our architecture is capable of considering
two relevant aspects: spatial features extracted through convolutional layers
from the RGBP images; and the temporal coherence of the problem, which is
exploited by recurrent layers. We show that, through a supervised learning
approach, the trained models are capable of predicting the people count with
high accuracy. Additionally, we present and demonstrate that a straightforward
modification of the methodology is effective to exclude salespeople from the
people count. Comprehensive experiments were conducted to validate, evaluate
and compare the proposed architecture. Results corroborated that LRCN-RetailNet
remarkably outperforms both the previous RetailNet architecture, which was
limited to evaluating a single image per iteration; and a state-of-the-art
neural network for object detection. Finally, computational performance
experiments confirmed that the entire methodology is effective to estimate
people count in real-time.
Related papers
- Task-Oriented Real-time Visual Inference for IoVT Systems: A Co-design Framework of Neural Networks and Edge Deployment [61.20689382879937]
Task-oriented edge computing addresses this by shifting data analysis to the edge.
Existing methods struggle to balance high model performance with low resource consumption.
We propose a novel co-design framework to optimize neural network architecture.
arXiv Detail & Related papers (2024-10-29T19:02:54Z) - Time Elastic Neural Networks [2.1756081703276]
We introduce and detail an atypical neural network architecture, called time elastic neural network (teNN)
The novelty compared to classical neural network architecture is that it explicitly incorporates time warping ability.
We demonstrate that, during the training process, the teNN succeeds in reducing the number of neurons required within each cell.
arXiv Detail & Related papers (2024-05-27T09:01:30Z) - CAManim: Animating end-to-end network activation maps [0.2509487459755192]
We propose a novel XAI visualization method denoted CAManim that seeks to broaden and focus end-user understanding of CNN predictions.
We additionally propose a novel quantitative assessment that expands upon the Remove and Debias (ROAD) metric.
This builds upon prior research to address the increasing demand for interpretable, robust, and transparent model assessment methodology.
arXiv Detail & Related papers (2023-12-19T01:07:36Z) - Visual Prompting Upgrades Neural Network Sparsification: A Data-Model Perspective [64.04617968947697]
We introduce a novel data-model co-design perspective: to promote superior weight sparsity.
Specifically, customized Visual Prompts are mounted to upgrade neural Network sparsification in our proposed VPNs framework.
arXiv Detail & Related papers (2023-12-03T13:50:24Z) - Learning to Learn with Generative Models of Neural Network Checkpoints [71.06722933442956]
We construct a dataset of neural network checkpoints and train a generative model on the parameters.
We find that our approach successfully generates parameters for a wide range of loss prompts.
We apply our method to different neural network architectures and tasks in supervised and reinforcement learning.
arXiv Detail & Related papers (2022-09-26T17:59:58Z) - ResPerfNet: Deep Residual Learning for Regressional Performance Modeling
of Deep Neural Networks [0.16311150636417257]
We propose a deep learning-based method, ResPerfNet, which trains a residual neural network with representative datasets obtained on the target platform to predict the performance for a deep neural network.
Our experimental results show that ResPerfNet can accurately predict the execution time of individual neural network layers and full network models on a variety of platforms.
arXiv Detail & Related papers (2020-12-03T03:02:42Z) - Neural Representations in Hybrid Recommender Systems: Prediction versus
Regularization [8.384351067134999]
We define the neural representation for prediction (NRP) framework and apply it to the autoencoder-based recommendation systems.
We also apply the NRP framework to a direct neural network structure which predicts the ratings without reconstructing the user and item information.
The results confirm that neural representations are better for prediction than regularization and show that the NRP framework, combined with the direct neural network structure, outperforms the state-of-the-art methods in the prediction task.
arXiv Detail & Related papers (2020-10-12T23:12:49Z) - A Semi-Supervised Assessor of Neural Architectures [157.76189339451565]
We employ an auto-encoder to discover meaningful representations of neural architectures.
A graph convolutional neural network is introduced to predict the performance of architectures.
arXiv Detail & Related papers (2020-05-14T09:02:33Z) - ResNeSt: Split-Attention Networks [86.25490825631763]
We present a modularized architecture, which applies the channel-wise attention on different network branches to leverage their success in capturing cross-feature interactions and learning diverse representations.
Our model, named ResNeSt, outperforms EfficientNet in accuracy and latency trade-off on image classification.
arXiv Detail & Related papers (2020-04-19T20:40:31Z) - Network Adjustment: Channel Search Guided by FLOPs Utilization Ratio [101.84651388520584]
This paper presents a new framework named network adjustment, which considers network accuracy as a function of FLOPs.
Experiments on standard image classification datasets and a wide range of base networks demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2020-04-06T15:51:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.