Concept-based Anomaly Detection in Retail Stores for Automatic
Correction using Mobile Robots
- URL: http://arxiv.org/abs/2310.14063v1
- Date: Sat, 21 Oct 2023 16:49:23 GMT
- Title: Concept-based Anomaly Detection in Retail Stores for Automatic
Correction using Mobile Robots
- Authors: Aditya Kapoor, Vartika Sengar, Nijil George, Vighnesh Vatsal,
Jayavardhana Gubbi, Balamuralidhar P and Arpan Pal
- Abstract summary: Co-AD is a Concept-based Anomaly Detection approach using a Vision Transformer (ViT)
It is able to flag misplaced objects without using a prior knowledge base such as a planogram.
It has a peak success rate of 89.90% on anomaly detection image sets of retail objects.
- Score: 3.989104441591223
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Tracking of inventory and rearrangement of misplaced items are some of the
most labor-intensive tasks in a retail environment. While there have been
attempts at using vision-based techniques for these tasks, they mostly use
planogram compliance for detection of any anomalies, a technique that has been
found lacking in robustness and scalability. Moreover, existing systems rely on
human intervention to perform corrective actions after detection. In this
paper, we present Co-AD, a Concept-based Anomaly Detection approach using a
Vision Transformer (ViT) that is able to flag misplaced objects without using a
prior knowledge base such as a planogram. It uses an auto-encoder architecture
followed by outlier detection in the latent space. Co-AD has a peak success
rate of 89.90% on anomaly detection image sets of retail objects drawn from the
RP2K dataset, compared to 80.81% on the best-performing baseline of a standard
ViT auto-encoder. To demonstrate its utility, we describe a robotic mobile
manipulation pipeline to autonomously correct the anomalies flagged by Co-AD.
This work is ultimately aimed towards developing autonomous mobile robot
solutions that reduce the need for human intervention in retail store
management.
Related papers
- Uncertainty Estimation for 3D Object Detection via Evidential Learning [63.61283174146648]
We introduce a framework for quantifying uncertainty in 3D object detection by leveraging an evidential learning loss on Bird's Eye View representations in the 3D detector.
We demonstrate both the efficacy and importance of these uncertainty estimates on identifying out-of-distribution scenes, poorly localized objects, and missing (false negative) detections.
arXiv Detail & Related papers (2024-10-31T13:13:32Z) - GeneralAD: Anomaly Detection Across Domains by Attending to Distorted Features [68.14842693208465]
GeneralAD is an anomaly detection framework designed to operate in semantic, near-distribution, and industrial settings.
We propose a novel self-supervised anomaly generation module that employs straightforward operations like noise addition and shuffling to patch features.
We extensively evaluated our approach on ten datasets, achieving state-of-the-art results in six and on-par performance in the remaining.
arXiv Detail & Related papers (2024-07-17T09:27:41Z) - Conformalized Teleoperation: Confidently Mapping Human Inputs to High-Dimensional Robot Actions [4.855534476454559]
We learn a mapping from low-dimensional human inputs to high-dimensional robot actions.
Our key idea is to adapt the assistive map at training time to additionally estimate high-dimensional action quantiles.
We propose an uncertainty-interval-based mechanism for detecting high-uncertainty user inputs and robot states.
arXiv Detail & Related papers (2024-06-11T23:16:46Z) - Multimodal Anomaly Detection based on Deep Auto-Encoder for Object Slip
Perception of Mobile Manipulation Robots [22.63980025871784]
The proposed framework integrates heterogeneous data streams collected from various robot sensors, including RGB and depth cameras, a microphone, and a force-torque sensor.
The integrated data is used to train a deep autoencoder to construct latent representations of the multisensory data that indicate the normal status.
Anomalies can then be identified by error scores measured by the difference between the trained encoder's latent values and the latent values of reconstructed input data.
arXiv Detail & Related papers (2024-03-06T09:15:53Z) - Run-time Introspection of 2D Object Detection in Automated Driving
Systems Using Learning Representations [13.529124221397822]
We introduce a novel introspection solution for 2D object detection based on Deep Neural Networks (DNNs)
We implement several state-of-the-art (SOTA) introspection mechanisms for error detection in 2D object detection, using one-stage and two-stage object detectors evaluated on KITTI and BDD datasets.
Our performance evaluation shows that the proposed introspection solution outperforms SOTA methods, achieving an absolute reduction in the missed error ratio of 9% to 17% in the BDD dataset.
arXiv Detail & Related papers (2024-03-02T10:56:14Z) - Integral Migrating Pre-trained Transformer Encoder-decoders for Visual
Object Detection [78.2325219839805]
imTED improves the state-of-the-art of few-shot object detection by up to 7.6% AP.
Experiments on MS COCO dataset demonstrate that imTED consistently outperforms its counterparts by 2.8%.
arXiv Detail & Related papers (2022-05-19T15:11:20Z) - Proactive Anomaly Detection for Robot Navigation with Multi-Sensor
Fusion [7.293053431456775]
Mobile robots produce anomalous behaviors that can lead to navigation failures.
Reactive anomaly detection methods identify anomalous task executions based on the current robot state.
We propose a proactive anomaly detection network (PAAD) for robot navigation in unstructured and uncertain environments.
arXiv Detail & Related papers (2022-04-03T19:48:40Z) - One-Shot Object Affordance Detection in the Wild [76.46484684007706]
Affordance detection refers to identifying the potential action possibilities of objects in an image.
We devise a One-Shot Affordance Detection Network (OSAD-Net) that estimates the human action purpose and then transfers it to help detect the common affordance from all candidate images.
With complex scenes and rich annotations, our PADv2 dataset can be used as a test bed to benchmark affordance detection methods.
arXiv Detail & Related papers (2021-08-08T14:53:10Z) - SABER: Data-Driven Motion Planner for Autonomously Navigating
Heterogeneous Robots [112.2491765424719]
We present an end-to-end online motion planning framework that uses a data-driven approach to navigate a heterogeneous robot team towards a global goal.
We use model predictive control (SMPC) to calculate control inputs that satisfy robot dynamics, and consider uncertainty during obstacle avoidance with chance constraints.
recurrent neural networks are used to provide a quick estimate of future state uncertainty considered in the SMPC finite-time horizon solution.
A Deep Q-learning agent is employed to serve as a high-level path planner, providing the SMPC with target positions that move the robots towards a desired global goal.
arXiv Detail & Related papers (2021-08-03T02:56:21Z) - Domain and Modality Gaps for LiDAR-based Person Detection on Mobile
Robots [91.01747068273666]
This paper studies existing LiDAR-based person detectors with a particular focus on mobile robot scenarios.
Experiments revolve around the domain gap between driving and mobile robot scenarios, as well as the modality gap between 3D and 2D LiDAR sensors.
Results provide practical insights into LiDAR-based person detection and facilitate informed decisions for relevant mobile robot designs and applications.
arXiv Detail & Related papers (2021-06-21T16:35:49Z) - Unsupervised anomaly detection for a Smart Autonomous Robotic Assistant
Surgeon (SARAS)using a deep residual autoencoder [7.655239948659381]
Anomalous events in a surgical setting are rare, making it difficult to capture data to train a detection model in a supervised fashion.
We propose an unsupervised approach to anomaly detection for robotic-assisted surgery based on deep residual autoencoders.
The system achieves recall and precision equal to 78.4%, 91.5%, respectively, on Cholec80 and of 95.6%, 88.1% on the SARAS phantom dataset.
arXiv Detail & Related papers (2021-04-22T12:10:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.