Designing an Efficient End-to-end Machine Learning Pipeline for
Real-time Empty-shelf Detection
- URL: http://arxiv.org/abs/2205.13060v1
- Date: Wed, 25 May 2022 21:51:20 GMT
- Title: Designing an Efficient End-to-end Machine Learning Pipeline for
Real-time Empty-shelf Detection
- Authors: Dipendra Jha, Ata Mahjoubfar, Anupama Joshi
- Abstract summary: On-shelf availability (OSA) of products in retail stores is a critical business criterion.
Here, we present an elegant approach for designing an end-to-end machine learning pipeline for real-time empty shelf detection.
Our dataset contains 1000 images, collected and annotated by following well-defined guidelines.
- Score: 0.483420384410068
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: On-Shelf Availability (OSA) of products in retail stores is a critical
business criterion in the fast moving consumer goods and retails sector. When a
product is out-of-stock (OOS) and a customer cannot find it on its designed
shelf, this causes a negative impact on the customer's behaviors and future
demands. Several methods are being adopted by retailers today to detect empty
shelves and ensure high OSA of products; however, such methods are generally
ineffective and infeasible since they are either manual, expensive or less
accurate. Recently machine learning based solutions have been proposed, but
they suffer from high computation cost and low accuracy problem due to lack of
large annotated datasets of on-shelf products. Here, we present an elegant
approach for designing an end-to-end machine learning (ML) pipeline for
real-time empty shelf detection. Considering the strong dependency between the
quality of ML models and the quality of data, we focus on the importance of
proper data collection, cleaning and correct data annotation before delving
into modeling. Since an empty-shelf detection solution should be
computationally-efficient for real-time predictions, we explore different
run-time optimizations to improve the model performance. Our dataset contains
1000 images, collected and annotated by following well-defined guidelines. Our
low-latency model achieves a mean average F1-score of 68.5%, and can process up
to 67 images/s on Intel Xeon Gold and up to 860 images/s on an A100 GPU. Our
annotated dataset is publicly available along with our optimized models.
Related papers
- A Lightweight and Extensible Cell Segmentation and Classification Model for Whole Slide Images [0.0]
We propose a solution that enhances data quality, model performance, and usability by creating a lightweight, cell segmentation and classification model.
We update data labels through cross-relabeling to refine annotations of PanNuke and MoNuSAC, producing a unified dataset with seven distinct cell types.
Third, to address foundation models' computational demands, we distill knowledge to reduce model size and complexity while maintaining comparable performance.
arXiv Detail & Related papers (2025-02-26T15:19:52Z) - Attribute-to-Delete: Machine Unlearning via Datamodel Matching [65.13151619119782]
Machine unlearning -- efficiently removing a small "forget set" training data on a pre-divertrained machine learning model -- has recently attracted interest.
Recent research shows that machine unlearning techniques do not hold up in such a challenging setting.
arXiv Detail & Related papers (2024-10-30T17:20:10Z) - Scaling Retrieval-Based Language Models with a Trillion-Token Datastore [85.4310806466002]
We find that increasing the size of the datastore used by a retrieval-based LM monotonically improves language modeling and several downstream tasks without obvious saturation.
By plotting compute-optimal scaling curves with varied datastore, model, and pretraining data sizes, we show that using larger datastores can significantly improve model performance for the same training compute budget.
arXiv Detail & Related papers (2024-07-09T08:27:27Z) - Uncertainty Aware Learning for Language Model Alignment [97.36361196793929]
We propose uncertainty-aware learning (UAL) to improve the model alignment of different task scenarios.
We implement UAL in a simple fashion -- adaptively setting the label smoothing value of training according to the uncertainty of individual samples.
Experiments on widely used benchmarks demonstrate that our UAL significantly and consistently outperforms standard supervised fine-tuning.
arXiv Detail & Related papers (2024-06-07T11:37:45Z) - How to Train Data-Efficient LLMs [56.41105687693619]
We study data-efficient approaches for pre-training language models (LLMs)
We find that Ask-LLM and Density sampling are the best methods in their respective categories.
In our comparison of 19 samplers, involving hundreds of evaluation tasks and pre-training runs, we find that Ask-LLM and Density are the best methods in their respective categories.
arXiv Detail & Related papers (2024-02-15T02:27:57Z) - A Simple and Efficient Baseline for Data Attribution on Images [107.12337511216228]
Current state-of-the-art approaches require a large ensemble of as many as 300,000 models to accurately attribute model predictions.
In this work, we focus on a minimalist baseline, utilizing the feature space of a backbone pretrained via self-supervised learning to perform data attribution.
Our method is model-agnostic and scales easily to large datasets.
arXiv Detail & Related papers (2023-11-03T17:29:46Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z) - Complementary Ensemble Learning [1.90365714903665]
We derive a technique to improve performance of state-of-the-art deep learning models.
Specifically, we train auxiliary models which are able to complement state-of-the-art model uncertainty.
arXiv Detail & Related papers (2021-11-09T03:23:05Z) - AI Total: Analyzing Security ML Models with Imperfect Data in Production [2.629585075202626]
Development of new machine learning models is typically done on manually curated data sets.
We develop a web-based visualization system that allows the users to quickly gather headline performance numbers.
It also enables the users to immediately observe the root cause of an issue when something goes wrong.
arXiv Detail & Related papers (2021-10-13T20:56:05Z) - FreaAI: Automated extraction of data slices to test machine learning
models [2.475112368179548]
We show the feasibility of automatically extracting feature models that result in explainable data slices over which the ML solution under-performs.
Our novel technique, IBM FreaAI aka FreaAI, extracts such slices from structured ML test data or any other labeled data.
arXiv Detail & Related papers (2021-08-12T09:21:16Z) - Improving the Performance of Fine-Grain Image Classifiers via Generative
Data Augmentation [0.5161531917413706]
We develop Data Augmentation from Proficient Pre-Training of Robust Generative Adrial Networks (DAPPER GAN)
DAPPER GAN is an ML analytics support tool that automatically generates novel views of training images.
We experimentally evaluate this technique on the Stanford Cars dataset, demonstrating improved vehicle make and model classification accuracy.
arXiv Detail & Related papers (2020-08-12T15:29:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.