Visual Sensor Pose Optimisation Using Rendering-based Visibility Models
for Robust Cooperative Perception
- URL: http://arxiv.org/abs/2106.05308v1
- Date: Wed, 9 Jun 2021 18:02:32 GMT
- Title: Visual Sensor Pose Optimisation Using Rendering-based Visibility Models
for Robust Cooperative Perception
- Authors: Eduardo Arnold, Sajjad Mozaffari, Mehrdad Dianati, Paul Jennings
- Abstract summary: Visual Sensor Networks can be used in a variety of perception applications such as infrastructure support for autonomous driving in complex road segments.
The pose of the sensors in such networks directly determines the coverage of the environment and objects therein.
This paper proposes two novel sensor pose optimisation methods, based on gradient-ascent and Programming techniques.
- Score: 4.5144287492490625
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Visual Sensor Networks can be used in a variety of perception applications
such as infrastructure support for autonomous driving in complex road segments.
The pose of the sensors in such networks directly determines the coverage of
the environment and objects therein, which impacts the performance of
applications such as object detection and tracking. Existing sensor pose
optimisation methods in the literature either maximise the coverage of ground
surfaces, or consider the visibility of the target objects as binary variables,
which cannot represent various degrees of visibility. Such formulations cannot
guarantee the visibility of the target objects as they fail to consider
occlusions. This paper proposes two novel sensor pose optimisation methods,
based on gradient-ascent and Integer Programming techniques, which maximise the
visibility of multiple target objects in cluttered environments. Both methods
consider a realistic visibility model based on a rendering engine that provides
pixel-level visibility information about the target objects. The proposed
methods are evaluated in a complex environment and compared to existing methods
in the literature. The evaluation results indicate that explicitly modelling
the visibility of target objects is critical to avoid occlusions in cluttered
environments. Furthermore, both methods significantly outperform existing
methods in terms of object visibility.
Related papers
- Perceptual Piercing: Human Visual Cue-based Object Detection in Low Visibility Conditions [2.0409124291940826]
This study proposes a novel deep learning framework inspired by atmospheric scattering and human visual cortex mechanisms to enhance object detection under poor visibility scenarios such as fog, smoke, and haze.
The objective is to enhance the precision and reliability of detection systems under adverse environmental conditions.
arXiv Detail & Related papers (2024-10-02T04:03:07Z) - Visual Forecasting as a Mid-level Representation for Avoidance [8.712750753534532]
The challenge of navigation in environments with dynamic objects continues to be a central issue in the study of autonomous agents.
While predictive methods hold promise, their reliance on precise state information makes them less practical for real-world implementation.
This study presents visual forecasting as an innovative alternative.
arXiv Detail & Related papers (2023-09-17T13:32:03Z) - Weakly-supervised Contrastive Learning for Unsupervised Object Discovery [52.696041556640516]
Unsupervised object discovery is promising due to its ability to discover objects in a generic manner.
We design a semantic-guided self-supervised learning model to extract high-level semantic features from images.
We introduce Principal Component Analysis (PCA) to localize object regions.
arXiv Detail & Related papers (2023-07-07T04:03:48Z) - Analysis of voxel-based 3D object detection methods efficiency for
real-time embedded systems [93.73198973454944]
Two popular voxel-based 3D object detection methods are studied in this paper.
Our experiments show that these methods mostly fail to detect distant small objects due to the sparsity of the input point clouds at large distances.
Our findings suggest that a considerable part of the computations of existing methods is focused on locations of the scene that do not contribute with successful detection.
arXiv Detail & Related papers (2021-05-21T12:40:59Z) - Ensembling object detectors for image and video data analysis [98.26061123111647]
We propose a method for ensembling the outputs of multiple object detectors for improving detection performance and precision of bounding boxes on image data.
We extend it to video data by proposing a two-stage tracking-based scheme for detection refinement.
arXiv Detail & Related papers (2021-02-09T12:38:16Z) - Unadversarial Examples: Designing Objects for Robust Vision [100.4627585672469]
We develop a framework that exploits the sensitivity of modern machine learning algorithms to input perturbations in order to design "robust objects"
We demonstrate the efficacy of the framework on a wide variety of vision-based tasks ranging from standard benchmarks to (in-simulation) robotics.
arXiv Detail & Related papers (2020-12-22T18:26:07Z) - Object-Driven Active Mapping for More Accurate Object Pose Estimation
and Robotic Grasping [5.385583891213281]
The framework is built on an object SLAM system integrated with a simultaneous multi-object pose estimation process.
By combining the mapping module and the exploration strategy, an accurate object map that is compatible with robotic grasping can be generated.
arXiv Detail & Related papers (2020-12-03T09:36:55Z) - Slender Object Detection: Diagnoses and Improvements [74.40792217534]
In this paper, we are concerned with the detection of a particular type of objects with extreme aspect ratios, namely textbfslender objects.
For a classical object detection method, a drastic drop of $18.9%$ mAP on COCO is observed, if solely evaluated on slender objects.
arXiv Detail & Related papers (2020-11-17T09:39:42Z) - Visual Object Recognition in Indoor Environments Using Topologically
Persistent Features [2.2344764434954256]
Object recognition in unseen indoor environments remains a challenging problem for visual perception of mobile robots.
We propose the use of topologically persistent features, which rely on the objects' shape information, to address this challenge.
We implement the proposed method on a real-world robot to demonstrate its usefulness.
arXiv Detail & Related papers (2020-10-07T06:04:17Z) - Benchmarking Unsupervised Object Representations for Video Sequences [111.81492107649889]
We compare the perceptual abilities of four object-centric approaches: ViMON, OP3, TBA and SCALOR.
Our results suggest that the architectures with unconstrained latent representations learn more powerful representations in terms of object detection, segmentation and tracking.
Our benchmark may provide fruitful guidance towards learning more robust object-centric video representations.
arXiv Detail & Related papers (2020-06-12T09:37:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.