Optimizing Camera Configurations for Multi-View Pedestrian Detection
- URL: http://arxiv.org/abs/2312.02144v1
- Date: Mon, 4 Dec 2023 18:59:02 GMT
- Title: Optimizing Camera Configurations for Multi-View Pedestrian Detection
- Authors: Yunzhong Hou, Xingjian Leng, Tom Gedeon, Liang Zheng
- Abstract summary: In this work, we present a novel solution that features a transformer-based camera configuration generator.
Using reinforcement learning, this generator autonomously explores vast combinations within the action space and searches for configurations that give the highest detection accuracy.
Across multiple simulation scenarios, the configurations generated by our transformer-based model consistently outperform random search, optimization, and configurations designed by human experts.
- Score: 21.89117952343898
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Jointly considering multiple camera views (multi-view) is very effective for
pedestrian detection under occlusion. For such multi-view systems, it is
critical to have well-designed camera configurations, including camera
locations, directions, and fields-of-view (FoVs). Usually, these configurations
are crafted based on human experience or heuristics. In this work, we present a
novel solution that features a transformer-based camera configuration
generator. Using reinforcement learning, this generator autonomously explores
vast combinations within the action space and searches for configurations that
give the highest detection accuracy according to the training dataset. The
generator learns advanced techniques like maximizing coverage, minimizing
occlusion, and promoting collaboration. Across multiple simulation scenarios,
the configurations generated by our transformer-based model consistently
outperform random search, heuristic-based methods, and configurations designed
by human experts, shedding light on future camera layout optimization.
Related papers
- Redundancy-Aware Camera Selection for Indoor Scene Neural Rendering [54.468355408388675]
We build a similarity matrix that incorporates both the spatial diversity of the cameras and the semantic variation of the images.
We apply a diversity-based sampling algorithm to optimize the camera selection.
We also develop a new dataset, IndoorTraj, which includes long and complex camera movements captured by humans in virtual indoor environments.
arXiv Detail & Related papers (2024-09-11T08:36:49Z) - RETINA: a hardware-in-the-loop optical facility with reduced optical aberrations [0.0]
Vision-based navigation algorithms have established themselves as effective solutions to determine the spacecraft state in orbit with low-cost and versatile sensors.
A dedicated simulation framework must be developed to emulate the orbital environment in a laboratory setup.
This paper presents the design of a low-aberration optical facility called RETINA to perform this task.
arXiv Detail & Related papers (2024-07-02T11:26:37Z) - City-Scale Multi-Camera Vehicle Tracking System with Improved Self-Supervised Camera Link Model [0.0]
This article introduces an innovative multi-camera vehicle tracking system that utilizes a self-supervised camera link model.
The proposed method achieves a new state-of-the-art among automatic camera-link based methods in CityFlow V2 benchmarks with 61.07% IDF1 Score.
arXiv Detail & Related papers (2024-05-18T17:28:35Z) - Toward Global Sensing Quality Maximization: A Configuration Optimization
Scheme for Camera Networks [15.795407587722924]
We investigate the reconfiguration strategy for the parameterized camera network model.
We form a single quantity that measures the sensing quality of the targets by the camera network.
We verify the effectiveness of our approach through extensive simulations and experiments.
arXiv Detail & Related papers (2022-11-28T09:21:47Z) - SurroundDepth: Entangling Surrounding Views for Self-Supervised
Multi-Camera Depth Estimation [101.55622133406446]
We propose a SurroundDepth method to incorporate the information from multiple surrounding views to predict depth maps across cameras.
Specifically, we employ a joint network to process all the surrounding views and propose a cross-view transformer to effectively fuse the information from multiple views.
In experiments, our method achieves the state-of-the-art performance on the challenging multi-camera depth estimation datasets.
arXiv Detail & Related papers (2022-04-07T17:58:47Z) - CNN-based Omnidirectional Object Detection for HermesBot Autonomous
Delivery Robot with Preliminary Frame Classification [53.56290185900837]
We propose an algorithm for optimizing a neural network for object detection using preliminary binary frame classification.
An autonomous mobile robot with 6 rolling-shutter cameras on the perimeter providing a 360-degree field of view was used as the experimental setup.
arXiv Detail & Related papers (2021-10-22T15:05:37Z) - DeepMultiCap: Performance Capture of Multiple Characters Using Sparse
Multiview Cameras [63.186486240525554]
DeepMultiCap is a novel method for multi-person performance capture using sparse multi-view cameras.
Our method can capture time varying surface details without the need of using pre-scanned template models.
arXiv Detail & Related papers (2021-05-01T14:32:13Z) - Infrastructure-based Multi-Camera Calibration using Radial Projections [117.22654577367246]
Pattern-based calibration techniques can be used to calibrate the intrinsics of the cameras individually.
Infrastucture-based calibration techniques are able to estimate the extrinsics using 3D maps pre-built via SLAM or Structure-from-Motion.
We propose to fully calibrate a multi-camera system from scratch using an infrastructure-based approach.
arXiv Detail & Related papers (2020-07-30T09:21:04Z) - Redesigning SLAM for Arbitrary Multi-Camera Systems [51.81798192085111]
Adding more cameras to SLAM systems improves robustness and accuracy but complicates the design of the visual front-end significantly.
In this work, we aim at an adaptive SLAM system that works for arbitrary multi-camera setups.
We adapt a state-of-the-art visual-inertial odometry with these modifications, and experimental results show that the modified pipeline can adapt to a wide range of camera setups.
arXiv Detail & Related papers (2020-03-04T11:44:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.