APT: Adaptive Perceptual quality based camera Tuning using reinforcement
learning
- URL: http://arxiv.org/abs/2211.08504v1
- Date: Tue, 15 Nov 2022 21:02:48 GMT
- Title: APT: Adaptive Perceptual quality based camera Tuning using reinforcement
learning
- Authors: Sibendu Paul, Kunal Rao, Giuseppe Coviello, Murugan Sankaradas, Oliver
Po, Y. Charlie Hu and Srimat Chakradhar
- Abstract summary: Capturing poor-quality video adversely affects the accuracy of analytics.
We propose a novel, reinforcement-learning based system that tunes the camera parameters to ensure a high-quality video capture.
As a result, such tuning restores the accuracy of insights when environmental conditions or scene content change.
- Score: 2.0741583844039915
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Cameras are increasingly being deployed in cities, enterprises and roads
world-wide to enable many applications in public safety, intelligent
transportation, retail, healthcare and manufacturing. Often, after initial
deployment of the cameras, the environmental conditions and the scenes around
these cameras change, and our experiments show that these changes can adversely
impact the accuracy of insights from video analytics. This is because the
camera parameter settings, though optimal at deployment time, are not the best
settings for good-quality video capture as the environmental conditions and
scenes around a camera change during operation. Capturing poor-quality video
adversely affects the accuracy of analytics. To mitigate the loss in accuracy
of insights, we propose a novel, reinforcement-learning based system APT that
dynamically, and remotely (over 5G networks), tunes the camera parameters, to
ensure a high-quality video capture, which mitigates any loss in accuracy of
video analytics. As a result, such tuning restores the accuracy of insights
when environmental conditions or scene content change. APT uses reinforcement
learning, with no-reference perceptual quality estimation as the reward
function. We conducted extensive real-world experiments, where we
simultaneously deployed two cameras side-by-side overlooking an enterprise
parking lot (one camera only has manufacturer-suggested default setting, while
the other camera is dynamically tuned by APT during operation). Our experiments
demonstrated that due to dynamic tuning by APT, the analytics insights are
consistently better at all times of the day: the accuracy of object detection
video analytics application was improved on average by ~ 42%. Since our reward
function is independent of any analytics task, APT can be readily used for
different video analytics tasks.
Related papers
- Redundancy-Aware Camera Selection for Indoor Scene Neural Rendering [54.468355408388675]
We build a similarity matrix that incorporates both the spatial diversity of the cameras and the semantic variation of the images.
We apply a diversity-based sampling algorithm to optimize the camera selection.
We also develop a new dataset, IndoorTraj, which includes long and complex camera movements captured by humans in virtual indoor environments.
arXiv Detail & Related papers (2024-09-11T08:36:49Z) - VICAN: Very Efficient Calibration Algorithm for Large Camera Networks [49.17165360280794]
We introduce a novel methodology that extends Pose Graph Optimization techniques.
We consider the bipartite graph encompassing cameras, object poses evolving dynamically, and camera-object relative transformations at each time step.
Our framework retains compatibility with traditional PGO solvers, but its efficacy benefits from a custom-tailored optimization scheme.
arXiv Detail & Related papers (2024-03-25T17:47:03Z) - Homography Estimation in Complex Topological Scenes [6.023710971800605]
Surveillance videos and images are used for a broad set of applications, ranging from traffic analysis to crime detection.
Extrinsic camera calibration data is important for most analysis applications.
We present an automated camera-calibration process leveraging a dictionary-based approach that does not require prior knowledge on any camera settings.
arXiv Detail & Related papers (2023-08-02T11:31:43Z) - Enhanced Stable View Synthesis [86.69338893753886]
We introduce an approach to enhance the novel view synthesis from images taken from a freely moving camera.
The introduced approach focuses on outdoor scenes where recovering accurate geometric scaffold and camera pose is challenging.
arXiv Detail & Related papers (2023-03-30T01:53:14Z) - PL-EVIO: Robust Monocular Event-based Visual Inertial Odometry with
Point and Line Features [3.6355269783970394]
Event cameras are motion-activated sensors that capture pixel-level illumination changes instead of the intensity image with a fixed frame rate.
We propose a robust, high-accurate, and real-time optimization-based monocular event-based visual-inertial odometry (VIO) method.
arXiv Detail & Related papers (2022-09-25T06:14:12Z) - Why is the video analytics accuracy fluctuating, and what can we do
about it? [2.0741583844039915]
It is a common practice to think of a video as a sequence of images (frames), and re-use deep neural network models that are trained only on images for similar analytics tasks on videos.
In this paper, we show that this leap of faith that deep learning models that work well on images will also work well on videos is actually flawed.
We show that even when a video camera is viewing a scene that is not changing in any human-perceptible way, the accuracy of video analytics application fluctuates noticeably.
arXiv Detail & Related papers (2022-08-23T23:16:24Z) - GPU-accelerated SIFT-aided source identification of stabilized videos [63.084540168532065]
We exploit the parallelization capabilities of Graphics Processing Units (GPUs) in the framework of stabilised frames inversion.
We propose to exploit SIFT features.
to estimate the camera momentum and %to identify less stabilized temporal segments.
Experiments confirm the effectiveness of the proposed approach in reducing the required computational time and improving the source identification accuracy.
arXiv Detail & Related papers (2022-07-29T07:01:31Z) - RADNet: A Deep Neural Network Model for Robust Perception in Moving
Autonomous Systems [8.706086688708014]
We develop a novel ranking method to rank videos based on the degree of global camera motion.
For the high ranking camera videos we show that the accuracy of action detection is decreased.
We propose an action detection pipeline that is robust to the camera motion effect and verify it empirically.
arXiv Detail & Related papers (2022-04-30T23:14:08Z) - CamTuner: Reinforcement-Learning based System for Camera Parameter
Tuning to enhance Analytics [2.637265703777453]
CamTuner is a system to automatically adapt the complex sensor to changing environments.
Our dynamic tuning approach results in up to 12% improvement in the accuracy of insights from several video analytics tasks.
arXiv Detail & Related papers (2021-07-08T16:43:02Z) - A Modified Fourier-Mellin Approach for Source Device Identification on
Stabilized Videos [72.40789387139063]
multimedia forensic tools usually exploit characteristic noise traces left by the camera sensor on the acquired frames.
This analysis requires that the noise pattern characterizing the camera and the noise pattern extracted from video frames under analysis are geometrically aligned.
We propose to overcome this limitation by searching scaling and rotation parameters in the frequency domain.
arXiv Detail & Related papers (2020-05-20T12:06:40Z) - Scene-Adaptive Video Frame Interpolation via Meta-Learning [54.87696619177496]
We propose to adapt the model to each video by making use of additional information that is readily available at test time.
We obtain significant performance gains with only a single gradient update without any additional parameters.
arXiv Detail & Related papers (2020-04-02T02:46:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.