Prototypical Cross-domain Knowledge Transfer for Cervical Dysplasia
Visual Inspection
- URL: http://arxiv.org/abs/2308.09983v1
- Date: Sat, 19 Aug 2023 11:25:09 GMT
- Title: Prototypical Cross-domain Knowledge Transfer for Cervical Dysplasia
Visual Inspection
- Authors: Yichen Zhang, Yifang Yin, Ying Zhang, Zhenguang Liu, Zheng Wang, Roger
Zimmermann
- Abstract summary: We propose to leverage cross-domain cervical images that were collected in different but related clinical studies to improve the model's performance.
To robustly learn the transferable information across datasets, we propose a novel prototype-based knowledge filtering method.
Our proposed method outperforms the state-of-the-art cervical dysplasia visual inspection by an absolute improvement of 4.7% in top-1 accuracy.
- Score: 37.66559952448587
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Early detection of dysplasia of the cervix is critical for cervical cancer
treatment. However, automatic cervical dysplasia diagnosis via visual
inspection, which is more appropriate in low-resource settings, remains a
challenging problem. Though promising results have been obtained by recent deep
learning models, their performance is significantly hindered by the limited
scale of the available cervix datasets. Distinct from previous methods that
learn from a single dataset, we propose to leverage cross-domain cervical
images that were collected in different but related clinical studies to improve
the model's performance on the targeted cervix dataset. To robustly learn the
transferable information across datasets, we propose a novel prototype-based
knowledge filtering method to estimate the transferability of cross-domain
samples. We further optimize the shared feature space by aligning the
cross-domain image representations simultaneously on domain level with early
alignment and class level with supervised contrastive learning, which endows
model training and knowledge transfer with stronger robustness. The empirical
results on three real-world benchmark cervical image datasets show that our
proposed method outperforms the state-of-the-art cervical dysplasia visual
inspection by an absolute improvement of 4.7% in top-1 accuracy, 7.0% in
precision, 1.4% in recall, 4.6% in F1 score, and 0.05 in ROC-AUC.
Related papers
- Cervical Cancer Detection Using Multi-Branch Deep Learning Model [0.6249768559720121]
This research proposes an innovative and novel approach to automate cervical cancer image classification using Multi-Head Self-Attention (MHSA) and convolutional neural networks (CNNs)
Our model achieved a remarkable accuracy of 98.522%, which holds promise for its applicability in other medical image recognition tasks.
arXiv Detail & Related papers (2024-08-20T02:44:48Z) - Deep Learning Descriptor Hybridization with Feature Reduction for Accurate Cervical Cancer Colposcopy Image Classification [0.9374652839580183]
We propose a novel Computer Aided Diagnosis (CAD) system that combines the strengths of various deep-learning descriptors with appropriate feature normalization.
Our approach achieves exceptional performance in the range of 97%-100% for both the normal-abnormal and the type classification.
arXiv Detail & Related papers (2024-05-01T06:05:13Z) - MM-SurvNet: Deep Learning-Based Survival Risk Stratification in Breast
Cancer Through Multimodal Data Fusion [18.395418853966266]
We propose a novel deep learning approach for breast cancer survival risk stratification.
We employ vision transformers, specifically the MaxViT model, for image feature extraction, and self-attention to capture intricate image relationships at the patient level.
A dual cross-attention mechanism fuses these features with genetic data, while clinical data is incorporated at the final layer to enhance predictive accuracy.
arXiv Detail & Related papers (2024-02-19T02:31:36Z) - DACov: A Deeper Analysis of Data Augmentation on the Computed Tomography
Segmentation Problem [0.0]
We present a deeper analysis of how data augmentation techniques improve segmentation performance on medical images.
We propose a novel data augmentation technique based on Generative Adversarial Networks (GANs) to create new healthy and unhealthy lung CT images.
Our findings show that GAN-based techniques and spatial-level transformations are the most promising for improving the learning of deep models on this problem.
arXiv Detail & Related papers (2023-03-10T13:41:20Z) - Significantly improving zero-shot X-ray pathology classification via fine-tuning pre-trained image-text encoders [50.689585476660554]
We propose a new fine-tuning strategy that includes positive-pair loss relaxation and random sentence sampling.
Our approach consistently improves overall zero-shot pathology classification across four chest X-ray datasets and three pre-trained models.
arXiv Detail & Related papers (2022-12-14T06:04:18Z) - Dissecting Self-Supervised Learning Methods for Surgical Computer Vision [51.370873913181605]
Self-Supervised Learning (SSL) methods have begun to gain traction in the general computer vision community.
The effectiveness of SSL methods in more complex and impactful domains, such as medicine and surgery, remains limited and unexplored.
We present an extensive analysis of the performance of these methods on the Cholec80 dataset for two fundamental and popular tasks in surgical context understanding, phase recognition and tool presence detection.
arXiv Detail & Related papers (2022-07-01T14:17:11Z) - Open-Set Recognition of Breast Cancer Treatments [91.3247063132127]
Open-set recognition generalizes a classification task by classifying test samples as one of the known classes from training or "unknown"
We apply a recent existing Gaussian mixture variational autoencoder model, which achieves state-of-the-art results for image datasets, to breast cancer patient data.
Not only do we obtain more accurate and robust classification results, with a 24.5% average F1 increase compared to a recent method, but we also reexamine open-set recognition in terms of deployability to a clinical setting.
arXiv Detail & Related papers (2022-01-09T04:35:55Z) - Cross-Site Severity Assessment of COVID-19 from CT Images via Domain
Adaptation [64.59521853145368]
Early and accurate severity assessment of Coronavirus disease 2019 (COVID-19) based on computed tomography (CT) images offers a great help to the estimation of intensive care unit event.
To augment the labeled data and improve the generalization ability of the classification model, it is necessary to aggregate data from multiple sites.
This task faces several challenges including class imbalance between mild and severe infections, domain distribution discrepancy between sites, and presence of heterogeneous features.
arXiv Detail & Related papers (2021-09-08T07:56:51Z) - Classification of COVID-19 in CT Scans using Multi-Source Transfer
Learning [91.3755431537592]
We propose the use of Multi-Source Transfer Learning to improve upon traditional Transfer Learning for the classification of COVID-19 from CT scans.
With our multi-source fine-tuning approach, our models outperformed baseline models fine-tuned with ImageNet.
Our best performing model was able to achieve an accuracy of 0.893 and a Recall score of 0.897, outperforming its baseline Recall score by 9.3%.
arXiv Detail & Related papers (2020-09-22T11:53:06Z) - Contrastive Cross-site Learning with Redesigned Net for COVID-19 CT
Classification [20.66003113364796]
The pandemic of coronavirus disease 2019 (COVID-19) has lead to a global public health crisis spreading hundreds of countries.
To assist the clinical diagnosis and reduce the tedious workload of image interpretation, developing automated tools for COVID-19 identification with CT image is highly desired.
This paper proposes a novel joint learning framework to perform accurate COVID-19 identification by effectively learning with heterogeneous datasets.
arXiv Detail & Related papers (2020-09-15T11:09:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.