A Large-scale Evaluation of Pretraining Paradigms for the Detection of
Defects in Electroluminescence Solar Cell Images
- URL: http://arxiv.org/abs/2402.17611v1
- Date: Tue, 27 Feb 2024 15:37:15 GMT
- Title: A Large-scale Evaluation of Pretraining Paradigms for the Detection of
Defects in Electroluminescence Solar Cell Images
- Authors: David Torpey and Lawrence Pratt and Richard Klein
- Abstract summary: This work is a large-scale evaluation and benchmarking of various pretraining methods for Solar Cell Defect Detection.
We cover supervised training with semantic segmentation, semi-supervised learning, and two self-supervised techniques.
We achieve a new state-of-the-art for SCDD and demonstrate that certain pretraining schemes result in superior performance on underrepresented classes.
- Score: 3.729242965449096
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Pretraining has been shown to improve performance in many domains, including
semantic segmentation, especially in domains with limited labelled data. In
this work, we perform a large-scale evaluation and benchmarking of various
pretraining methods for Solar Cell Defect Detection (SCDD) in
electroluminescence images, a field with limited labelled datasets. We cover
supervised training with semantic segmentation, semi-supervised learning, and
two self-supervised techniques. We also experiment with both in-distribution
and out-of-distribution (OOD) pretraining and observe how this affects
downstream performance. The results suggest that supervised training on a large
OOD dataset (COCO), self-supervised pretraining on a large OOD dataset
(ImageNet), and semi-supervised pretraining (CCT) all yield statistically
equivalent performance for mean Intersection over Union (mIoU). We achieve a
new state-of-the-art for SCDD and demonstrate that certain pretraining schemes
result in superior performance on underrepresented classes. Additionally, we
provide a large-scale unlabelled EL image dataset of $22000$ images, and a
$642$-image labelled semantic segmentation EL dataset, for further research in
developing self- and semi-supervised training techniques in this domain.
Related papers
- Rethinking Transformers Pre-training for Multi-Spectral Satellite
Imagery [78.43828998065071]
Recent advances in unsupervised learning have demonstrated the ability of large vision models to achieve promising results on downstream tasks.
Such pre-training techniques have also been explored recently in the remote sensing domain due to the availability of large amount of unlabelled data.
In this paper, we re-visit transformers pre-training and leverage multi-scale information that is effectively utilized with multiple modalities.
arXiv Detail & Related papers (2024-03-08T16:18:04Z) - GenCo: An Auxiliary Generator from Contrastive Learning for Enhanced
Few-Shot Learning in Remote Sensing [9.504503675097137]
We introduce a generator-based contrastive learning framework (GenCo) that pre-trains backbones and simultaneously explores variants of feature samples.
In fine-tuning, the auxiliary generator can be used to enrich limited labeled data samples in feature space.
We demonstrate the effectiveness of our method in improving few-shot learning performance on two key remote sensing datasets.
arXiv Detail & Related papers (2023-07-27T03:59:19Z) - In-Domain Self-Supervised Learning Improves Remote Sensing Image Scene
Classification [5.323049242720532]
Self-supervised learning has emerged as a promising approach for remote sensing image classification.
We present a study of different self-supervised pre-training strategies and evaluate their effect across 14 downstream datasets.
arXiv Detail & Related papers (2023-07-04T10:57:52Z) - Self-Supervised Pretraining for 2D Medical Image Segmentation [0.0]
Self-supervised learning offers a way to lower the need for manually annotated data by pretraining models for a specific domain on unlabelled data.
We find that self-supervised pretraining on natural images and target-domain-specific images leads to the fastest and most stable downstream convergence.
In low-data scenarios, supervised ImageNet pretraining achieves the best accuracy, requiring less than 100 annotated samples to realise close to minimal error.
arXiv Detail & Related papers (2022-09-01T09:25:22Z) - On Efficient Transformer and Image Pre-training for Low-level Vision [74.22436001426517]
Pre-training has marked numerous state of the arts in high-level computer vision.
We present an in-depth study of image pre-training.
We find pre-training plays strikingly different roles in low-level tasks.
arXiv Detail & Related papers (2021-12-19T15:50:48Z) - Self-Supervised Pre-Training for Transformer-Based Person
Re-Identification [54.55281692768765]
Transformer-based supervised pre-training achieves great performance in person re-identification (ReID)
Due to the domain gap between ImageNet and ReID datasets, it usually needs a larger pre-training dataset to boost the performance.
This work aims to mitigate the gap between the pre-training and ReID datasets from the perspective of data and model structure.
arXiv Detail & Related papers (2021-11-23T18:59:08Z) - Efficient Conditional Pre-training for Transfer Learning [71.01129334495553]
We propose efficient filtering methods to select relevant subsets from the pre-training dataset.
We validate our techniques by pre-training on ImageNet in both the unsupervised and supervised settings.
We improve standard ImageNet pre-training by 1-3% by tuning available models on our subsets and pre-training on a dataset filtered from a larger scale dataset.
arXiv Detail & Related papers (2020-11-20T06:16:15Z) - Neural Semi-supervised Learning for Text Classification Under
Large-Scale Pretraining [51.19885385587916]
We conduct studies on semi-supervised learning in the task of text classification under the context of large-scale LM pretraining.
Our work marks an initial step in understanding the behavior of semi-supervised learning models under the context of large-scale pretraining.
arXiv Detail & Related papers (2020-11-17T13:39:05Z) - Tackling the Problem of Limited Data and Annotations in Semantic
Segmentation [1.0152838128195467]
To tackle the problem of limited data annotations in image segmentation, different pre-trained models and CRF based methods are applied.
To this end, RotNet, DeeperCluster, and Semi&Weakly Supervised Learning (SWSL) pre-trained models are transferred and finetuned in a DeepLab-v2 baseline.
The results of my study show that, on this small dataset, using a pre-trained ResNet50 SWSL model gives results that are 7.4% better than applying an ImageNet pre-trained model.
arXiv Detail & Related papers (2020-07-14T21:11:11Z) - Cheaper Pre-training Lunch: An Efficient Paradigm for Object Detection [86.0580214485104]
We propose a general and efficient pre-training paradigm, Montage pre-training, for object detection.
Montage pre-training needs only the target detection dataset while taking only 1/4 computational resources compared to the widely adopted ImageNet pre-training.
The efficiency and effectiveness of Montage pre-training are validated by extensive experiments on the MS-COCO dataset.
arXiv Detail & Related papers (2020-04-25T16:09:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.