Optimized Deep Encoder-Decoder Methods for Crack Segmentation
- URL: http://arxiv.org/abs/2008.06266v2
- Date: Thu, 26 Aug 2021 09:57:55 GMT
- Title: Optimized Deep Encoder-Decoder Methods for Crack Segmentation
- Authors: Jacob K\"onig, Mark Jenkins, Mike Mannion, Peter Barrie, Gordon
Morison
- Abstract summary: Surface crack segmentation poses a challenging computer vision task as background, shape, colour and size of cracks vary.
In this work we propose optimized deep encoder-decoder methods consisting of a combination of techniques which yield an increase in crack segmentation performance.
- Score: 0.9449650062296824
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Surface crack segmentation poses a challenging computer vision task as
background, shape, colour and size of cracks vary. In this work we propose
optimized deep encoder-decoder methods consisting of a combination of
techniques which yield an increase in crack segmentation performance.
Specifically we propose a decoder-part for an encoder-decoder based deep
learning architecture for semantic segmentation and study its components to
achieve increased performance. We also examine the use of different encoder
strategies and introduce a data augmentation policy to increase the amount of
available training data. The performance evaluation of our method is carried
out on four publicly available crack segmentation datasets. Additionally, we
introduce two techniques into the field of surface crack segmentation,
previously not used there: Generating results using test-time-augmentation and
performing a statistical result analysis over multiple training runs. The
former approach generally yields increased performance results, whereas the
latter allows for more reproducible and better representability of a methods
results. Using those aforementioned strategies with our proposed
encoder-decoder architecture we are able to achieve new state of the art
results in all datasets.
Related papers
- Extreme Encoder Output Frame Rate Reduction: Improving Computational
Latencies of Large End-to-End Models [59.57732929473519]
We apply multiple frame reduction layers in the encoder to compress encoder outputs into a small number of output frames.
We demonstrate that we can generate one encoder output frame for every 2.56 sec of input speech, without significantly affecting word error rate on a large-scale voice search task.
arXiv Detail & Related papers (2024-02-27T03:40:44Z) - Using DUCK-Net for Polyp Image Segmentation [0.0]
"DUCK-Net" is capable of effectively learning and generalizing from small amounts of medical images to perform accurate segmentation tasks.
We demonstrate its capabilities specifically for polyp segmentation in colonoscopy images.
arXiv Detail & Related papers (2023-11-03T20:58:44Z) - TEACHTEXT: CrossModal Generalized Distillation for Text-Video Retrieval [103.85002875155551]
We propose a novel generalized distillation method, TeachText, for exploiting large-scale language pretraining.
We extend our method to video side modalities and show that we can effectively reduce the number of used modalities at test time.
Our approach advances the state of the art on several video retrieval benchmarks by a significant margin and adds no computational overhead at test time.
arXiv Detail & Related papers (2021-04-16T17:55:28Z) - A Holistically-Guided Decoder for Deep Representation Learning with
Applications to Semantic Segmentation and Object Detection [74.88284082187462]
One common strategy is to adopt dilated convolutions in the backbone networks to extract high-resolution feature maps.
We propose one novel holistically-guided decoder which is introduced to obtain the high-resolution semantic-rich feature maps.
arXiv Detail & Related papers (2020-12-18T10:51:49Z) - EfficientFCN: Holistically-guided Decoding for Semantic Segmentation [49.27021844132522]
State-of-the-art semantic segmentation algorithms are mostly based on dilated Fully Convolutional Networks (dilatedFCN)
We propose the EfficientFCN, whose backbone is a common ImageNet pre-trained network without any dilated convolution.
Such a framework achieves comparable or even better performance than state-of-the-art methods with only 1/3 of the computational cost.
arXiv Detail & Related papers (2020-08-24T14:48:23Z) - Beyond Single Stage Encoder-Decoder Networks: Deep Decoders for Semantic
Image Segmentation [56.44853893149365]
Single encoder-decoder methodologies for semantic segmentation are reaching their peak in terms of segmentation quality and efficiency per number of layers.
We propose a new architecture based on a decoder which uses a set of shallow networks for capturing more information content.
In order to further improve the architecture we introduce a weight function which aims to re-balance classes to increase the attention of the networks to under-represented objects.
arXiv Detail & Related papers (2020-07-19T18:44:34Z) - Automatic Crack Detection on Road Pavements Using Encoder Decoder
Architecture [9.34360241512198]
The proposed algorithm considers an encoder-decoder architecture with hierarchical feature learning and dilated convolution, named U-Hierarchical Dilated Network (U-HDN)
Crack characteristics with multiple context information are automatically able to learn and perform end-to-end crack detection.
arXiv Detail & Related papers (2020-07-01T13:32:23Z) - Rethinking and Improving Natural Language Generation with Layer-Wise
Multi-View Decoding [59.48857453699463]
In sequence-to-sequence learning, the decoder relies on the attention mechanism to efficiently extract information from the encoder.
Recent work has proposed to use representations from different encoder layers for diversified levels of information.
We propose layer-wise multi-view decoding, where for each decoder layer, together with the representations from the last encoder layer, which serve as a global view, those from other encoder layers are supplemented for a stereoscopic view of the source sequences.
arXiv Detail & Related papers (2020-05-16T20:00:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.