LaneAF: Robust Multi-Lane Detection with Affinity Fields
- URL: http://arxiv.org/abs/2103.12040v1
- Date: Mon, 22 Mar 2021 17:43:19 GMT
- Title: LaneAF: Robust Multi-Lane Detection with Affinity Fields
- Authors: Hala Abualsaud, Sean Liu, David Lu, Kenny Situ, Akshay Rangesh and
Mohan M. Trivedi
- Abstract summary: This study presents an approach to lane detection involving the prediction of binary segmentation masks and per-pixel affinity fields.
affinity fields, along with the binary masks, can then be used to cluster lane pixels horizontally and vertically into corresponding lane instances in a post-processing step.
Our proposed approach performs on par with state-of-the-art approaches on the limited TuSimple benchmark, and sets a new state-of-the-art on the challenging CULane dataset.
- Score: 9.607267724849766
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This study presents an approach to lane detection involving the prediction of
binary segmentation masks and per-pixel affinity fields. These affinity fields,
along with the binary masks, can then be used to cluster lane pixels
horizontally and vertically into corresponding lane instances in a
post-processing step. This clustering is achieved through a simple row-by-row
decoding process with little overhead; such an approach allows LaneAF to detect
a variable number of lanes without assuming a fixed or maximum number of lanes.
Moreover, this form of clustering is more interpretable in comparison to
previous visual clustering approaches, and can be analyzed to identify and
correct sources of error. Qualitative and quantitative results obtained on
popular lane detection datasets demonstrate the model's ability to detect and
cluster lanes effectively and robustly. Our proposed approach performs on par
with state-of-the-art approaches on the limited TuSimple benchmark, and sets a
new state-of-the-art on the challenging CULane dataset.
Related papers
- One-Step Multi-View Clustering Based on Transition Probability [61.841829428397034]
We introduce the One-Step Multi-View Clustering Based on Transition Probability (OSMVC-TP)
Our method directly learns the transition probabilities from anchor points to categories, and calculates the transition probabilities from samples to categories.
To maintain consistency in labels across different views, we apply a Schatten p-norm constraint on the tensor composed of the soft labels.
arXiv Detail & Related papers (2024-03-03T09:43:23Z) - Masked Gamma-SSL: Learning Uncertainty Estimation via Masked Image
Modeling [19.000718685399935]
This work proposes a semantic segmentation network that produces high-quality uncertainty estimates in a single forward pass.
We exploit general representations from foundation models and unlabelled datasets through a Masked Image Modeling (MIM) approach.
For neural networks used in safety-critical applications, bias in the training data can lead to errors.
arXiv Detail & Related papers (2024-02-27T15:49:54Z) - Applying Unsupervised Semantic Segmentation to High-Resolution UAV Imagery for Enhanced Road Scene Parsing [12.558144256470827]
A novel unsupervised road parsing framework is presented.
The proposed method achieves a mean Intersection over Union (mIoU) of 89.96% on the development dataset without any manual annotation.
arXiv Detail & Related papers (2024-02-05T13:16:12Z) - Towards General Visual-Linguistic Face Forgery Detection [95.73987327101143]
Deepfakes are realistic face manipulations that can pose serious threats to security, privacy, and trust.
Existing methods mostly treat this task as binary classification, which uses digital labels or mask signals to train the detection model.
We propose a novel paradigm named Visual-Linguistic Face Forgery Detection(VLFFD), which uses fine-grained sentence-level prompts as the annotation.
arXiv Detail & Related papers (2023-07-31T10:22:33Z) - Boosting Few-shot Fine-grained Recognition with Background Suppression
and Foreground Alignment [53.401889855278704]
Few-shot fine-grained recognition (FS-FGR) aims to recognize novel fine-grained categories with the help of limited available samples.
We propose a two-stage background suppression and foreground alignment framework, which is composed of a background activation suppression (BAS) module, a foreground object alignment (FOA) module, and a local to local (L2L) similarity metric.
Experiments conducted on multiple popular fine-grained benchmarks demonstrate that our method outperforms the existing state-of-the-art by a large margin.
arXiv Detail & Related papers (2022-10-04T07:54:40Z) - Rethinking Clustering-Based Pseudo-Labeling for Unsupervised
Meta-Learning [146.11600461034746]
Method for unsupervised meta-learning, CACTUs, is a clustering-based approach with pseudo-labeling.
This approach is model-agnostic and can be combined with supervised algorithms to learn from unlabeled data.
We prove that the core reason for this is lack of a clustering-friendly property in the embedding space.
arXiv Detail & Related papers (2022-09-27T19:04:36Z) - On Mitigating Hard Clusters for Face Clustering [48.39472979642971]
Face clustering is a promising way to scale up face recognition systems using large-scale unlabeled face images.
We introduce two novel modules, Neighborhood-Diffusion-based Density (NDDe) and Transition-Probability-based Distance (TPDi)
Our experiments on multiple benchmarks show that each module contributes to the final performance of our method.
arXiv Detail & Related papers (2022-07-25T03:55:15Z) - Rank-Consistency Deep Hashing for Scalable Multi-Label Image Search [90.30623718137244]
We propose a novel deep hashing method for scalable multi-label image search.
A new rank-consistency objective is applied to align the similarity orders from two spaces.
A powerful loss function is designed to penalize the samples whose semantic similarity and hamming distance are mismatched.
arXiv Detail & Related papers (2021-02-02T13:46:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.