Solar Potential Assessment using Multi-Class Buildings Segmentation from
Aerial Images
- URL: http://arxiv.org/abs/2111.11397v1
- Date: Mon, 22 Nov 2021 18:16:07 GMT
- Title: Solar Potential Assessment using Multi-Class Buildings Segmentation from
Aerial Images
- Authors: Hasan Nasrallah, Abed Ellatif Samhat, Ghaleb Faour, Yilei Shi and Ali
J. Ghandour
- Abstract summary: We exploit the power of fully convolutional neural networks for an instance segmentation task using extra added classes to the output.
We also show that CutMix mixed data augmentations and the One-Cycle learning rate policy are greater regularization methods to achieve a better fit on the training data.
- Score: 3.180674374101366
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Semantic Segmentation of buildings present in satellite images using
encoder-decoder like convolutional neural networks is being achieved with
relatively high pixel-wise metric scores. In this paper, we aim to exploit the
power of fully convolutional neural networks for an instance segmentation task
using extra added classes to the output along with the watershed processing
technique to leverage better object-wise metric results. We also show that
CutMix mixed data augmentations and the One-Cycle learning rate policy are
greater regularization methods to achieve a better fit on the training data and
increase performance. Furthermore, Mixed Precision Training provided more
flexibility to experiment with bigger networks and batches while maintaining
stability and convergence during training. We compare and show the effect of
these additional changes throughout our whole pipeline to finally provide a set
a tuned hyper-parameters that are proven to perform better.
Related papers
- DynaSeg: A Deep Dynamic Fusion Method for Unsupervised Image Segmentation Incorporating Feature Similarity and Spatial Continuity [0.5755004576310334]
We introduce DynaSeg, an innovative unsupervised image segmentation approach.
Unlike traditional methods, DynaSeg employs a dynamic weighting scheme that adapts flexibly to image characteristics.
DynaSeg prevents undersegmentation failures where the number of predicted clusters might converge to one.
arXiv Detail & Related papers (2024-05-09T00:30:45Z) - Rethinking Transformers Pre-training for Multi-Spectral Satellite
Imagery [78.43828998065071]
Recent advances in unsupervised learning have demonstrated the ability of large vision models to achieve promising results on downstream tasks.
Such pre-training techniques have also been explored recently in the remote sensing domain due to the availability of large amount of unlabelled data.
In this paper, we re-visit transformers pre-training and leverage multi-scale information that is effectively utilized with multiple modalities.
arXiv Detail & Related papers (2024-03-08T16:18:04Z) - Image edge enhancement for effective image classification [7.470763273994321]
We propose an edge enhancement-based method to enhance both accuracy and training speed of neural networks.
Our approach involves extracting high frequency features, such as edges, from images within the available dataset and fusing them with the original images.
arXiv Detail & Related papers (2024-01-13T10:01:34Z) - Online Convolutional Re-parameterization [51.97831675242173]
We present online convolutional re- parameterization (OREPA), a two-stage pipeline, aiming to reduce the huge training overhead by squeezing the complex training-time block into a single convolution.
Compared with the state-of-the-art re-param models, OREPA is able to save the training-time memory cost by about 70% and accelerate the training speed by around 2x.
We also conduct experiments on object detection and semantic segmentation and show consistent improvements on the downstream tasks.
arXiv Detail & Related papers (2022-04-02T09:50:19Z) - Combining Image Features and Patient Metadata to Enhance Transfer
Learning [0.0]
We compare the performance of six state-of-the-art deep neural networks in classification tasks when using only image features, to when these are combined with patient metadata.
Our results indicate that this performance enhancement may be a general property of deep networks and should be explored in other areas.
arXiv Detail & Related papers (2021-10-08T15:43:31Z) - Fusion of CNNs and statistical indicators to improve image
classification [65.51757376525798]
Convolutional Networks have dominated the field of computer vision for the last ten years.
Main strategy to prolong this trend relies on further upscaling networks in size.
We hypothesise that adding heterogeneous sources of information may be more cost-effective to a CNN than building a bigger network.
arXiv Detail & Related papers (2020-12-20T23:24:31Z) - Mixup-Transformer: Dynamic Data Augmentation for NLP Tasks [75.69896269357005]
Mixup is the latest data augmentation technique that linearly interpolates input examples and the corresponding labels.
In this paper, we explore how to apply mixup to natural language processing tasks.
We incorporate mixup to transformer-based pre-trained architecture, named "mixup-transformer", for a wide range of NLP tasks.
arXiv Detail & Related papers (2020-10-05T23:37:30Z) - Memory-efficient training with streaming dimensionality reduction [14.198224213972173]
We introduce streaming batch component analysis as an update algorithm for Deep Neural Network training.
We show that streaming batch component analysis can effectively train convolutional neural networks on a variety of common datasets.
arXiv Detail & Related papers (2020-04-25T02:13:43Z) - Understanding the Effects of Data Parallelism and Sparsity on Neural
Network Training [126.49572353148262]
We study two factors in neural network training: data parallelism and sparsity.
Despite their promising benefits, understanding of their effects on neural network training remains elusive.
arXiv Detail & Related papers (2020-03-25T10:49:22Z) - Curriculum By Smoothing [52.08553521577014]
Convolutional Neural Networks (CNNs) have shown impressive performance in computer vision tasks such as image classification, detection, and segmentation.
We propose an elegant curriculum based scheme that smoothes the feature embedding of a CNN using anti-aliasing or low-pass filters.
As the amount of information in the feature maps increases during training, the network is able to progressively learn better representations of the data.
arXiv Detail & Related papers (2020-03-03T07:27:44Z) - Improving the Resolution of CNN Feature Maps Efficiently with
Multisampling [8.655380837944188]
One version of our method, which we call subsampling, significantly improves the accuracy of state-of-the-art architectures such as DenseNet and ResNet without any additional parameters.
We glean possible insight into the nature of data augmentations and demonstrate experimentally that coarse feature maps are bottlenecking the performance of neural networks in image classification.
arXiv Detail & Related papers (2018-05-28T04:29:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.