A hierarchical semantic segmentation framework for computer vision-based
bridge damage detection
- URL: http://arxiv.org/abs/2207.08878v1
- Date: Mon, 18 Jul 2022 18:42:54 GMT
- Title: A hierarchical semantic segmentation framework for computer vision-based
bridge damage detection
- Authors: Jingxiao Liu, Yujie Wei, Bingqing Chen
- Abstract summary: Computer vision-based damage detection using remote cameras and unmanned aerial vehicles (UAVs) enables efficient and low-cost bridge health monitoring.
This paper introduces a semantic segmentation framework that imposes the hierarchical semantic relationship between component category and damage types.
In this way, the damage detection model could focus on learning features from possible damaged regions only and avoid the effects of other irrelevant regions.
- Score: 3.7642333932730634
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Computer vision-based damage detection using remote cameras and unmanned
aerial vehicles (UAVs) enables efficient and low-cost bridge health monitoring
that reduces labor costs and the needs for sensor installation and maintenance.
By leveraging recent semantic image segmentation approaches, we are able to
find regions of critical structural components and recognize damage at the
pixel level using images as the only input. However, existing methods perform
poorly when detecting small damages (e.g., cracks and exposed rebars) and thin
objects with limited image samples, especially when the components of interest
are highly imbalanced. To this end, this paper introduces a semantic
segmentation framework that imposes the hierarchical semantic relationship
between component category and damage types. For example, certain concrete
cracks only present on bridge columns and therefore the non-column region will
be masked out when detecting such damages. In this way, the damage detection
model could focus on learning features from possible damaged regions only and
avoid the effects of other irrelevant regions. We also utilize multi-scale
augmentation that provides views with different scales that preserves
contextual information of each image without losing the ability of handling
small and thin objects. Furthermore, the proposed framework employs important
sampling that repeatedly samples images containing rare components (e.g.,
railway sleeper and exposed rebars) to provide more data samples, which
addresses the imbalanced data challenge.
Related papers
- Structural damage detection via hierarchical damage information with volumetric assessment [3.1033928913175766]
Post-detection, there is the challenge of reliance on manual assessments of detected damages.
Guided-DetNet is characterized by Generative Attention Module (GAM), Hierarchical Elimination Algorithm (HEA), and Volumetric Contour Visual Assessment (VCVA)
Guided-DetNet outperformed the best-compared models in a triple classification task by a difference of not less than 3% and not less than 2% in a dual detection task under varying metrics.
arXiv Detail & Related papers (2024-07-29T04:33:04Z) - Visual Context-Aware Person Fall Detection [52.49277799455569]
We present a segmentation pipeline to semi-automatically separate individuals and objects in images.
Background objects such as beds, chairs, or wheelchairs can challenge fall detection systems, leading to false positive alarms.
We demonstrate that object-specific contextual transformations during training effectively mitigate this challenge.
arXiv Detail & Related papers (2024-04-11T19:06:36Z) - Few Shot Part Segmentation Reveals Compositional Logic for Industrial Anomaly Detection [11.996050578189056]
We introduce a novel component segmentation model for Logical anomalies (LA) detection.
To ensure consistent segmentation across unlabeled images, we employ a histogram matching loss in conjunction with an entropy loss.
For effective LA detection, we propose an adaptive scaling strategy to standardize anomaly scores from different memory banks in inference.
arXiv Detail & Related papers (2023-12-21T12:14:31Z) - Improving Vision Anomaly Detection with the Guidance of Language
Modality [64.53005837237754]
This paper tackles the challenges for vision modality from a multimodal point of view.
We propose Cross-modal Guidance (CMG) to tackle the redundant information issue and sparse space issue.
To learn a more compact latent space for the vision anomaly detector, CMLE learns a correlation structure matrix from the language modality.
arXiv Detail & Related papers (2023-10-04T13:44:56Z) - Unsupervised Skin Lesion Segmentation via Structural Entropy
Minimization on Multi-Scale Superpixel Graphs [59.19218582436495]
We propose an unsupervised Skin Lesion sEgmentation framework based on structural entropy and isolation forest outlier Detection, namely SLED.
Skin lesions are segmented by minimizing the structural entropy of a superpixel graph constructed from the dermoscopic image.
We characterize the consistency of healthy skin features and devise a novel multi-scale segmentation mechanism by outlier detection, which enhances the segmentation accuracy by leveraging the superpixel features from multiple scales.
arXiv Detail & Related papers (2023-09-05T02:15:51Z) - PAIF: Perception-Aware Infrared-Visible Image Fusion for Attack-Tolerant
Semantic Segmentation [50.556961575275345]
We propose a perception-aware fusion framework to promote segmentation robustness in adversarial scenes.
We show that our scheme substantially enhances the robustness, with gains of 15.3% mIOU, compared with advanced competitors.
arXiv Detail & Related papers (2023-08-08T01:55:44Z) - High-Resolution Vision Transformers for Pixel-Level Identification of
Structural Components and Damage [1.8923948104852863]
We develop a semantic segmentation network based on vision transformers and Laplacian pyramids scaling networks.
The proposed framework has been evaluated through comprehensive experiments on a dataset of bridge inspection report images.
arXiv Detail & Related papers (2023-08-06T03:34:25Z) - High-Fidelity Visual Structural Inspections through Transformers and
Learnable Resizers [2.126862120884775]
Recent advances in unmanned aerial vehicles (UAVs) and artificial intelligence have made the visual inspections faster, safer, and more reliable.
High-resolution segmentation is extremely challenging due to the high computational memory demands.
We propose a hybrid strategy that can adapt to different inspections tasks by managing the global and local semantics trade-off.
arXiv Detail & Related papers (2022-10-21T18:08:26Z) - You Better Look Twice: a new perspective for designing accurate
detectors with reduced computations [56.34005280792013]
BLT-net is a new low-computation two-stage object detection architecture.
It reduces computations by separating objects from background using a very lite first-stage.
Resulting image proposals are then processed in the second-stage by a highly accurate model.
arXiv Detail & Related papers (2021-07-21T12:39:51Z) - Synthetic Image Augmentation for Damage Region Segmentation using
Conditional GAN with Structure Edge [0.0]
We propose a synthetic augmentation procedure to generate damaged images using the image-to-image translation mapping.
We apply popular per-pixel segmentation algorithms such as the FCN-8s, SegNet, and DeepLabv3+Xception-v2.
We demonstrate that re-training a data set added with synthetic augmentation procedure make higher accuracy based on indices the mean IoU, damage region of interest IoU, precision, recall, BF score when we predict test images.
arXiv Detail & Related papers (2020-05-07T06:04:02Z) - RescueNet: Joint Building Segmentation and Damage Assessment from
Satellite Imagery [83.49145695899388]
RescueNet is a unified model that can simultaneously segment buildings and assess the damage levels to individual buildings and can be trained end-to-end.
RescueNet is tested on the large scale and diverse xBD dataset and achieves significantly better building segmentation and damage classification performance than previous methods.
arXiv Detail & Related papers (2020-04-15T19:52:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.