PLGAN: Generative Adversarial Networks for Power-Line Segmentation in
Aerial Images
- URL: http://arxiv.org/abs/2204.07243v1
- Date: Thu, 14 Apr 2022 21:43:31 GMT
- Title: PLGAN: Generative Adversarial Networks for Power-Line Segmentation in
Aerial Images
- Authors: Rabab Abdelfattah, Xiaofeng Wang, Song Wang
- Abstract summary: PLGAN is a simple yet effective method to segment power lines from aerial images with different backgrounds.
We exploit the appropriate form of the generated images for high-quality feature embedding.
Our proposed PLGAN outperforms the prior state-of-the-art methods for semantic segmentation and line detection.
- Score: 15.504887854179666
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Accurate segmentation of power lines in various aerial images is very
important for UAV flight safety. The complex background and very thin
structures of power lines, however, make it an inherently difficult task in
computer vision. This paper presents PLGAN, a simple yet effective method based
on generative adversarial networks, to segment power lines from aerial images
with different backgrounds. Instead of directly using the adversarial networks
to generate the segmentation, we take their certain decoding features and embed
them into another semantic segmentation network by considering more context,
geometry, and appearance information of power lines. We further exploit the
appropriate form of the generated images for high-quality feature embedding and
define a new loss function in the Hough-transform parameter space to enhance
the segmentation of very thin power lines. Extensive experiments and
comprehensive analysis demonstrate that our proposed PLGAN outperforms the
prior state-of-the-art methods for semantic segmentation and line detection.
Related papers
- TransResNet: Integrating the Strengths of ViTs and CNNs for High Resolution Medical Image Segmentation via Feature Grafting [6.987177704136503]
High-resolution images are preferable in medical imaging domain as they significantly improve the diagnostic capability of the underlying method.
Most of the existing deep learning-based techniques for medical image segmentation are optimized for input images having small spatial dimensions and perform poorly on high-resolution images.
We propose a parallel-in-branch architecture called TransResNet, which incorporates Transformer and CNN in a parallel manner to extract features from multi-resolution images independently.
arXiv Detail & Related papers (2024-10-01T18:22:34Z) - Boosting Cross-Domain Point Classification via Distilling Relational Priors from 2D Transformers [59.0181939916084]
Traditional 3D networks mainly focus on local geometric details and ignore the topological structure between local geometries.
We propose a novel Priors Distillation (RPD) method to extract priors from the well-trained transformers on massive images.
Experiments on the PointDA-10 and the Sim-to-Real datasets verify that the proposed method consistently achieves the state-of-the-art performance of UDA for point cloud classification.
arXiv Detail & Related papers (2024-07-26T06:29:09Z) - Enhancing Polygonal Building Segmentation via Oriented Corners [0.3749861135832072]
This paper introduces a novel deep convolutional neural network named OriCornerNet, which directly extracts delineated building polygons from input images.
Our approach involves a deep model that predicts building footprint masks, corners, and orientation vectors that indicate directions toward adjacent corners.
Performance evaluations conducted on SpaceNet Vegas and CrowdAI-small datasets demonstrate the competitive efficacy of our approach.
arXiv Detail & Related papers (2024-07-17T01:59:06Z) - Shape-Sensitive Loss for Catheter and Guidewire Segmentation [5.115480059688438]
We introduce a shape-sensitive loss function for catheter and guidewire segmentation.
We utilize it in a vision transformer network to establish a new state-of-the-art result on a large-scale X-ray images dataset.
arXiv Detail & Related papers (2023-11-19T03:05:21Z) - Dual-Augmented Transformer Network for Weakly Supervised Semantic
Segmentation [4.02487511510606]
Weakly supervised semantic segmentation (WSSS) is a fundamental computer vision task, which aims to segment out the object within only class-level labels.
Traditional methods adopt the CNN-based network and utilize the class activation map (CAM) strategy to discover the object regions.
An alternative is to explore vision transformers (ViT) to encode the image to acquire the global semantic information.
We propose a dual network with both CNN-based and transformer networks for mutually complementary learning.
arXiv Detail & Related papers (2023-09-30T08:41:11Z) - Autonomous Point Cloud Segmentation for Power Lines Inspection in Smart
Grid [56.838297900091426]
An unsupervised Machine Learning (ML) framework is proposed, to detect, extract and analyze the characteristics of power lines of both high and low voltage.
The proposed framework can efficiently detect the power lines and perform PLC-based hazard analysis.
arXiv Detail & Related papers (2023-08-14T17:14:58Z) - DUFormer: Solving Power Line Detection Task in Aerial Images using
Semantic Segmentation [17.77548837421917]
Unmanned aerial vehicles (UAVs) are frequently used for inspecting power lines and capturing high-resolution aerial images.
To tackle this problem, we introduce DUFormer, a semantic segmentation algorithm explicitly designed to detect power lines in aerial images.
Our proposed method outperforms all state-of-the-art methods in power line segmentation on the publicly accessible TTPLA dataset.
arXiv Detail & Related papers (2023-04-12T12:59:02Z) - DeepLSD: Line Segment Detection and Refinement with Deep Image Gradients [105.25109274550607]
Line segments are increasingly used in vision tasks.
Traditional line detectors based on the image gradient are extremely fast and accurate, but lack robustness in noisy images and challenging conditions.
We propose to combine traditional and learned approaches to get the best of both worlds: an accurate and robust line detector.
arXiv Detail & Related papers (2022-12-15T12:36:49Z) - Weakly-supervised segmentation of referring expressions [81.73850439141374]
Text grounded semantic SEGmentation learns segmentation masks directly from image-level referring expressions without pixel-level annotations.
Our approach demonstrates promising results for weakly-supervised referring expression segmentation on the PhraseCut and RefCOCO datasets.
arXiv Detail & Related papers (2022-05-10T07:52:24Z) - Weakly-supervised fire segmentation by visualizing intermediate CNN
layers [82.75113406937194]
Fire localization in images and videos is an important step for an autonomous system to combat fire incidents.
We consider weakly supervised segmentation of fire in images, in which only image labels are used to train the network.
We show that in the case of fire segmentation, which is a binary segmentation problem, the mean value of features in a mid-layer of classification CNN can perform better than conventional Class Activation Mapping (CAM) method.
arXiv Detail & Related papers (2021-11-16T11:56:28Z) - Less is More: Pay Less Attention in Vision Transformers [61.05787583247392]
Less attention vIsion Transformer builds upon the fact that convolutions, fully-connected layers, and self-attentions have almost equivalent mathematical expressions for processing image patch sequences.
The proposed LIT achieves promising performance on image recognition tasks, including image classification, object detection and instance segmentation.
arXiv Detail & Related papers (2021-05-29T05:26:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.