Contrastive Learning for Lane Detection via Cross-Similarity
- URL: http://arxiv.org/abs/2308.08242v3
- Date: Fri, 1 Sep 2023 08:12:30 GMT
- Title: Contrastive Learning for Lane Detection via Cross-Similarity
- Authors: Ali Zoljodi, Sadegh Abadijou, Mina Alibeigi, Masoud Daneshtalab
- Abstract summary: Lane markings have strong shape priors, but their visibility is easily compromised.
A large amount of data is required to train a lane detection approach that can withstand natural variations caused by low visibility.
Our solution, Contrastive Learning for Lane Detection via cross-similarity (CLLD), is a self-supervised learning method.
- Score: 0.5735035463793009
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Detecting road lanes is challenging due to intricate markings vulnerable to
unfavorable conditions. Lane markings have strong shape priors, but their
visibility is easily compromised. Factors like lighting, weather, vehicles,
pedestrians, and aging colors challenge the detection. A large amount of data
is required to train a lane detection approach that can withstand natural
variations caused by low visibility. This is because there are numerous lane
shapes and natural variations that exist. Our solution, Contrastive Learning
for Lane Detection via cross-similarity (CLLD), is a self-supervised learning
method that tackles this challenge by enhancing lane detection models
resilience to real-world conditions that cause lane low visibility. CLLD is a
novel multitask contrastive learning that trains lane detection approaches to
detect lane markings even in low visible situations by integrating local
feature contrastive learning (CL) with our new proposed operation
cross-similarity. Local feature CL focuses on extracting features for small
image parts, which is necessary to localize lane segments, while
cross-similarity captures global features to detect obscured lane segments
using their surrounding. We enhance cross-similarity by randomly masking parts
of input images for augmentation. Evaluated on benchmark datasets, CLLD
outperforms state-of-the-art contrastive learning, especially in
visibility-impairing conditions like shadows. Compared to supervised learning,
CLLD excels in scenarios like shadows and crowded scenes.
Related papers
- Context-aware TFL: A Universal Context-aware Contrastive Learning Framework for Temporal Forgery Localization [60.73623588349311]
We propose a universal context-aware contrastive learning framework (UniCaCLF) for temporal forgery localization.<n>Our approach leverages supervised contrastive learning to discover and identify forged instants by means of anomaly detection.<n>An efficient context-aware contrastive coding is introduced to further push the limit of instant feature distinguishability between genuine and forged instants.
arXiv Detail & Related papers (2025-06-10T06:40:43Z) - Cross-Modal Consistency Learning for Sign Language Recognition [92.44927164283641]
Existing pre-training methods solely focus on the compact pose data.
We propose a Cross-modal Consistency Learning framework (CCL- SLR)
CCL- SLR learns from both RGB and pose modalities based on self-supervised pre-training.
arXiv Detail & Related papers (2025-03-16T12:34:07Z) - LDTR: Transformer-based Lane Detection with Anchor-chain Representation [11.184960972042406]
Lane detection scenarios with limited- or no-visual-clue of lanes remain challenging and crucial for automated driving.
Inspired by the DETR architecture, we propose LDTR, a transformer-based model to address these issues.
Experimental results demonstrate that LDTR achieves state-of-the-art performance on well-known datasets.
arXiv Detail & Related papers (2024-03-21T12:29:26Z) - LVLane: Deep Learning for Lane Detection and Classification in
Challenging Conditions [2.5641096293146712]
We present an end-to-end lane detection and classification system based on deep learning methodologies.
In our study, we introduce a unique dataset meticulously curated to encompass scenarios that pose significant challenges for state-of-the-art (SOTA) lane localization models.
We propose a CNN-based classification branch, seamlessly integrated with the detector, facilitating the identification of distinct lane types.
arXiv Detail & Related papers (2023-07-13T16:09:53Z) - Towards Effective Image Manipulation Detection with Proposal Contrastive
Learning [61.5469708038966]
We propose Proposal Contrastive Learning (PCL) for effective image manipulation detection.
Our PCL consists of a two-stream architecture by extracting two types of global features from RGB and noise views respectively.
Our PCL can be easily adapted to unlabeled data in practice, which can reduce manual labeling costs and promote more generalizable features.
arXiv Detail & Related papers (2022-10-16T13:30:13Z) - Repainting and Imitating Learning for Lane Detection [52.5220065495956]
Current lane detection methods are struggling with the invisibility lane issue caused by heavy shadows.
We propose a novel Repainting and Imitating Learning framework containing a pair of teacher and student.
Our method introduces no extra time cost during inference and can be plug-and-play in various cutting-edge lane detection networks.
arXiv Detail & Related papers (2022-10-11T02:26:39Z) - Dual Contrastive Learning for General Face Forgery Detection [64.41970626226221]
We propose a novel face forgery detection framework, named Dual Contrastive Learning (DCL), which constructs positive and negative paired data.
To explore the essential discrepancies, Intra-Instance Contrastive Learning (Intra-ICL) is introduced to focus on the local content inconsistencies prevalent in the forged faces.
arXiv Detail & Related papers (2021-12-27T05:44:40Z) - Semantically Contrastive Learning for Low-light Image Enhancement [48.71522073014808]
Low-light image enhancement (LLE) remains challenging due to the unfavorable prevailing low-contrast and weak-visibility problems of single RGB images.
We propose an effective semantically contrastive learning paradigm for LLE (namely SCL-LLE)
Our method surpasses the state-of-the-arts LLE models over six independent cross-scenes datasets.
arXiv Detail & Related papers (2021-12-13T07:08:33Z) - MC-LCR: Multi-modal contrastive classification by locally correlated
representations for effective face forgery detection [11.124150983521158]
We propose a novel framework named Multi-modal Contrastive Classification by Locally Correlated Representations.
Our MC-LCR aims to amplify implicit local discrepancies between authentic and forged faces from both spatial and frequency domains.
We achieve state-of-the-art performance and demonstrate the robustness and generalization of our method.
arXiv Detail & Related papers (2021-10-07T09:24:12Z) - Local Relation Learning for Face Forgery Detection [73.73130683091154]
We propose a novel perspective of face forgery detection via local relation learning.
Specifically, we propose a Multi-scale Patch Similarity Module (MPSM), which measures the similarity between features of local regions.
We also propose an RGB-Frequency Attention Module (RFAM) to fuse information in both RGB and frequency domains for more comprehensive local feature representation.
arXiv Detail & Related papers (2021-05-06T10:44:32Z) - Pluggable Weakly-Supervised Cross-View Learning for Accurate Vehicle
Re-Identification [53.6218051770131]
Cross-view consistent feature representation is key for accurate vehicle ReID.
Existing approaches resort to supervised cross-view learning using extensive extra viewpoints annotations.
We present a pluggable Weakly-supervised Cross-View Learning (WCVL) module for vehicle ReID.
arXiv Detail & Related papers (2021-03-09T11:51:09Z) - Lane Detection Model Based on Spatio-Temporal Network With Double
Convolutional Gated Recurrent Units [11.968518335236787]
Lane detection will remain an open problem for some time to come.
A-temporal network with double Conal Gated Recurrent Units (ConvGRUs) proposed to address lane detection in challenging scenes.
Our model can outperform the state-of-the-art lane detection models.
arXiv Detail & Related papers (2020-08-10T06:50:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.