Deep Leaning-Based Ultra-Fast Stair Detection
- URL: http://arxiv.org/abs/2201.05275v1
- Date: Fri, 14 Jan 2022 02:05:01 GMT
- Title: Deep Leaning-Based Ultra-Fast Stair Detection
- Authors: Chen Wang, Zhongcai Pei, Shuang Qiu, Zhiyong Tang
- Abstract summary: We propose an end-to-end method for stair line detection based on deep learning.
In experiments, our method can achieve high performance in terms of both speed and accuracy.
A lightweight version can even achieve 300+ frames per second with the same resolution.
- Score: 6.362951673024623
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Staircases are some of the most common building structures in urban
environments. Stair detection is an important task for various applications,
including the environmental perception of exoskeleton robots, humanoid robots,
and rescue robots and the navigation of visually impaired people. Most existing
stair detection algorithms have difficulty dealing with the diversity of stair
structure materials, extreme light and serious occlusion. Inspired by human
perception, we propose an end-to-end method based on deep learning.
Specifically, we treat the process of stair line detection as a multitask
involving coarse-grained semantic segmentation and object detection. The input
images are divided into cells, and a simple neural network is used to judge
whether each cell contains stair lines. For cells containing stair lines, the
locations of the stair lines relative to each cell are regressed. Extensive
experiments on our dataset show that our method can achieve high performance in
terms of both speed and accuracy. A lightweight version can even achieve 300+
frames per second with the same resolution. Our code is available at GitHub.
Related papers
- Staircase Localization for Autonomous Exploration in Urban Environments [7.301415426190581]
A staircase localization method is proposed for robots to explore urban environments autonomously.
The proposed method employs a modular design in the form of a cascade pipeline consisting of three modules of stair detection, line segment detection, and stair localization modules.
arXiv Detail & Related papers (2024-03-26T02:28:49Z) - StairNet: Visual Recognition of Stairs for Human-Robot Locomotion [2.3811618212533663]
StairNet is an initiative to support the development of new deep learning models for visual sensing and recognition of stairs.
We present an overview of the development of our large-scale dataset with over 515,000 manually labeled images.
We show that StairNet can be an effective platform to develop and study new visual perception systems for human-robot locomotion.
arXiv Detail & Related papers (2023-10-31T17:30:57Z) - StairNetV3: Depth-aware Stair Modeling using Deep Learning [6.145334325463317]
Vision-based stair perception can help autonomous mobile robots deal with the challenge of climbing stairs.
Current monocular vision methods are difficult to model stairs accurately without depth information.
This paper proposes a depth-aware stair modeling method for monocular vision.
arXiv Detail & Related papers (2023-08-13T08:11:40Z) - DeepLSD: Line Segment Detection and Refinement with Deep Image Gradients [105.25109274550607]
Line segments are increasingly used in vision tasks.
Traditional line detectors based on the image gradient are extremely fast and accurate, but lack robustness in noisy images and challenging conditions.
We propose to combine traditional and learned approaches to get the best of both worlds: an accurate and robust line detector.
arXiv Detail & Related papers (2022-12-15T12:36:49Z) - RGB-D based Stair Detection using Deep Learning for Autonomous Stair
Climbing [6.362951673024623]
We propose a neural network architecture with inputs of both RGB map and depth map.
Specifically, we design the selective module which can make the network learn the complementary relationship between RGB map and depth map.
Experiments on our dataset show that our method can achieve better accuracy and recall compared with the previous state-of-the-art deep learning method.
arXiv Detail & Related papers (2022-12-02T11:22:52Z) - Deep Hierarchical Planning from Pixels [86.14687388689204]
Director is a method for learning hierarchical behaviors directly from pixels by planning inside the latent space of a learned world model.
Despite operating in latent space, the decisions are interpretable because the world model can decode goals into images for visualization.
Director also learns successful behaviors across a wide range of environments, including visual control, Atari games, and DMLab levels.
arXiv Detail & Related papers (2022-06-08T18:20:15Z) - Learning Perceptual Locomotion on Uneven Terrains using Sparse Visual
Observations [75.60524561611008]
This work aims to exploit the use of sparse visual observations to achieve perceptual locomotion over a range of commonly seen bumps, ramps, and stairs in human-centred environments.
We first formulate the selection of minimal visual input that can represent the uneven surfaces of interest, and propose a learning framework that integrates such exteroceptive and proprioceptive data.
We validate the learned policy in tasks that require omnidirectional walking over flat ground and forward locomotion over terrains with obstacles, showing a high success rate.
arXiv Detail & Related papers (2021-09-28T20:25:10Z) - Rapid Exploration for Open-World Navigation with Latent Goal Models [78.45339342966196]
We describe a robotic learning system for autonomous exploration and navigation in diverse, open-world environments.
At the core of our method is a learned latent variable model of distances and actions, along with a non-parametric topological memory of images.
We use an information bottleneck to regularize the learned policy, giving us (i) a compact visual representation of goals, (ii) improved generalization capabilities, and (iii) a mechanism for sampling feasible goals for exploration.
arXiv Detail & Related papers (2021-04-12T23:14:41Z) - ViNG: Learning Open-World Navigation with Visual Goals [82.84193221280216]
We propose a learning-based navigation system for reaching visually indicated goals.
We show that our system, which we call ViNG, outperforms previously-proposed methods for goal-conditioned reinforcement learning.
We demonstrate ViNG on a number of real-world applications, such as last-mile delivery and warehouse inspection.
arXiv Detail & Related papers (2020-12-17T18:22:32Z) - Anchor-free Small-scale Multispectral Pedestrian Detection [88.7497134369344]
We propose a method for effective and efficient multispectral fusion of the two modalities in an adapted single-stage anchor-free base architecture.
We aim at learning pedestrian representations based on object center and scale rather than direct bounding box predictions.
Results show our method's effectiveness in detecting small-scaled pedestrians.
arXiv Detail & Related papers (2020-08-19T13:13:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.