Training CNN Classifiers for Semantic Segmentation using Partially
Annotated Images: with Application on Human Thigh and Calf MRI
- URL: http://arxiv.org/abs/2008.07030v1
- Date: Sun, 16 Aug 2020 23:38:02 GMT
- Title: Training CNN Classifiers for Semantic Segmentation using Partially
Annotated Images: with Application on Human Thigh and Calf MRI
- Authors: Chun Kit Wong, Stephanie Marchesseau, Maria Kalimeri, Tiang Siew Yap,
Serena S. H. Teo, Lingaraj Krishna, Alfredo Franco-Obreg\'on, Stacey K. H.
Tay, Chin Meng Khoo, Philip T. H. Lee, Melvin K. S. Leow, John J. Totman,
Mary C. Stephenson
- Abstract summary: We propose a set of strategies to train one single classifier in segmenting all label classes that are heterogeneously annotated across multiple datasets.
We show that presence masking is capable of significantly improving both training and inference efficiency across imaging modalities and anatomical regions.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Objective: Medical image datasets with pixel-level labels tend to have a
limited number of organ or tissue label classes annotated, even when the images
have wide anatomical coverage. With supervised learning, multiple classifiers
are usually needed given these partially annotated datasets. In this work, we
propose a set of strategies to train one single classifier in segmenting all
label classes that are heterogeneously annotated across multiple datasets
without moving into semi-supervised learning. Methods: Masks were first created
from each label image through a process we termed presence masking. Three
presence masking modes were evaluated, differing mainly in weightage assigned
to the annotated and unannotated classes. These masks were then applied to the
loss function during training to remove the influence of unannotated classes.
Results: Evaluation against publicly available CT datasets shows that presence
masking is a viable method for training class-generic classifiers. Our
class-generic classifier can perform as well as multiple class-specific
classifiers combined, while the training duration is similar to that required
for one class-specific classifier. Furthermore, the class-generic classifier
can outperform the class-specific classifiers when trained on smaller datasets.
Finally, consistent results are observed from evaluations against human thigh
and calf MRI datasets collected in-house. Conclusion: The evaluation outcomes
show that presence masking is capable of significantly improving both training
and inference efficiency across imaging modalities and anatomical regions.
Improved performance may even be observed on small datasets. Significance:
Presence masking strategies can reduce the computational resources and costs
involved in manual medical image annotations. All codes are publicly available
at https://github.com/wong-ck/DeepSegment.
Related papers
- Spatial Action Unit Cues for Interpretable Deep Facial Expression Recognition [55.97779732051921]
State-of-the-art classifiers for facial expression recognition (FER) lack interpretability, an important feature for end-users.
A new learning strategy is proposed to explicitly incorporate AU cues into classifier training, allowing to train deep interpretable models.
Our new strategy is generic, and can be applied to any deep CNN- or transformer-based classifier without requiring any architectural change or significant additional training time.
arXiv Detail & Related papers (2024-10-01T10:42:55Z) - Accurate Explanation Model for Image Classifiers using Class Association Embedding [5.378105759529487]
We propose a generative explanation model that combines the advantages of global and local knowledge.
Class association embedding (CAE) encodes each sample into a pair of separated class-associated and individual codes.
Building-block coherency feature extraction algorithm is proposed that efficiently separates class-associated features from individual ones.
arXiv Detail & Related papers (2024-06-12T07:41:00Z) - Efficient Subclass Segmentation in Medical Images [3.383033695275859]
One feasible way to reduce the cost is to annotate with coarse-grained superclass labels while using limited fine-grained annotations as a complement.
There is a lack of research on efficient learning of fine-grained subclasses in semantic segmentation tasks.
Our approach achieves comparable accuracy to a model trained with full subclass annotations, with limited subclass annotations and sufficient superclass annotations.
arXiv Detail & Related papers (2023-07-01T07:39:08Z) - Triple-View Feature Learning for Medical Image Segmentation [9.992387025633805]
TriSegNet is a semi-supervised semantic segmentation framework.
It uses triple-view feature learning on a limited amount of labelled data and a large amount of unlabeled data.
arXiv Detail & Related papers (2022-08-12T14:41:40Z) - Semantic Segmentation In-the-Wild Without Seeing Any Segmentation
Examples [34.97652735163338]
We propose a novel approach for creating semantic segmentation masks for every object.
Our method takes as input the image-level labels of the class categories present in the image.
The output of this stage provides pixel-level pseudo-labels, instead of the manual pixel-level labels required by supervised methods.
arXiv Detail & Related papers (2021-12-06T17:32:38Z) - Self-Supervised Learning for Fine-Grained Image Classification [0.0]
Fine-grained datasets usually provide bounding box annotations along with class labels to aid the process of classification.
On the other hand, self-supervised learning exploits the freely available data to generate supervisory signals which act as labels.
Our idea is to leverage self-supervision such that the model learns useful representations of fine-grained image classes.
arXiv Detail & Related papers (2021-07-29T14:01:31Z) - Per-Pixel Classification is Not All You Need for Semantic Segmentation [184.2905747595058]
Mask classification is sufficiently general to solve both semantic- and instance-level segmentation tasks.
We propose MaskFormer, a simple mask classification model which predicts a set of binary masks.
Our method outperforms both current state-of-the-art semantic (55.6 mIoU on ADE20K) and panoptic segmentation (52.7 PQ on COCO) models.
arXiv Detail & Related papers (2021-07-13T17:59:50Z) - A Closer Look at Self-training for Zero-Label Semantic Segmentation [53.4488444382874]
Being able to segment unseen classes not observed during training is an important technical challenge in deep learning.
Prior zero-label semantic segmentation works approach this task by learning visual-semantic embeddings or generative models.
We propose a consistency regularizer to filter out noisy pseudo-labels by taking the intersections of the pseudo-labels generated from different augmentations of the same image.
arXiv Detail & Related papers (2021-04-21T14:34:33Z) - Generative Multi-Label Zero-Shot Learning [136.17594611722285]
Multi-label zero-shot learning strives to classify images into multiple unseen categories for which no data is available during training.
Our work is the first to tackle the problem of multi-label feature in the (generalized) zero-shot setting.
Our cross-level fusion-based generative approach outperforms the state-of-the-art on all three datasets.
arXiv Detail & Related papers (2021-01-27T18:56:46Z) - Attribute Propagation Network for Graph Zero-shot Learning [57.68486382473194]
We introduce the attribute propagation network (APNet), which is composed of 1) a graph propagation model generating attribute vector for each class and 2) a parameterized nearest neighbor (NN) classifier.
APNet achieves either compelling performance or new state-of-the-art results in experiments with two zero-shot learning settings and five benchmark datasets.
arXiv Detail & Related papers (2020-09-24T16:53:40Z) - A Systematic Evaluation: Fine-Grained CNN vs. Traditional CNN
Classifiers [54.996358399108566]
We investigate the performance of the landmark general CNN classifiers, which presented top-notch results on large scale classification datasets.
We compare it against state-of-the-art fine-grained classifiers.
We show an extensive evaluation on six datasets to determine whether the fine-grained classifier is able to elevate the baseline in their experiments.
arXiv Detail & Related papers (2020-03-24T23:49:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.