Semi-Supervised Learning for Eye Image Segmentation
- URL: http://arxiv.org/abs/2103.09369v1
- Date: Wed, 17 Mar 2021 00:05:19 GMT
- Title: Semi-Supervised Learning for Eye Image Segmentation
- Authors: Aayush K. Chaudhary, Prashnna K. Gyawali, Linwei Wang, Jeff B. Pelz
- Abstract summary: Recent advances in appearance-based models have shown improved eye tracking performance in difficult scenarios.
The improved accuracy often comes at the cost of labeling an enormous dataset.
This work presents two semi-supervised learning frameworks to identify eye-parts by taking advantage of unlabeled images.
- Score: 7.084953573199144
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent advances in appearance-based models have shown improved eye tracking
performance in difficult scenarios like occlusion due to eyelashes, eyelids or
camera placement, and environmental reflections on the cornea and glasses. The
key reason for the improvement is the accurate and robust identification of eye
parts (pupil, iris, and sclera regions). The improved accuracy often comes at
the cost of labeling an enormous dataset, which is complex and time-consuming.
This work presents two semi-supervised learning frameworks to identify
eye-parts by taking advantage of unlabeled images where labeled datasets are
scarce. With these frameworks, leveraging the domain-specific augmentation and
novel spatially varying transformations for image segmentation, we show
improved performance on various test cases. For instance, for a model trained
on just 48 labeled images, these frameworks achieved an improvement of 0.38%
and 0.65% in segmentation performance over the baseline model, which is trained
only with the labeled dataset.
Related papers
- Synergy and Diversity in CLIP: Enhancing Performance Through Adaptive Backbone Ensembling [58.50618448027103]
Contrastive Language-Image Pretraining (CLIP) stands out as a prominent method for image representation learning.
This paper explores the differences across various CLIP-trained vision backbones.
Method achieves a remarkable increase in accuracy of up to 39.1% over the best single backbone.
arXiv Detail & Related papers (2024-05-27T12:59:35Z) - Visual Context-Aware Person Fall Detection [52.49277799455569]
We present a segmentation pipeline to semi-automatically separate individuals and objects in images.
Background objects such as beds, chairs, or wheelchairs can challenge fall detection systems, leading to false positive alarms.
We demonstrate that object-specific contextual transformations during training effectively mitigate this challenge.
arXiv Detail & Related papers (2024-04-11T19:06:36Z) - Deep Domain Adaptation: A Sim2Real Neural Approach for Improving Eye-Tracking Systems [80.62854148838359]
Eye image segmentation is a critical step in eye tracking that has great influence over the final gaze estimate.
We use dimensionality-reduction techniques to measure the overlap between the target eye images and synthetic training data.
Our methods result in robust, improved performance when tackling the discrepancy between simulation and real-world data samples.
arXiv Detail & Related papers (2024-03-23T22:32:06Z) - Few-shot Image Classification based on Gradual Machine Learning [6.935034849731568]
Few-shot image classification aims to accurately classify unlabeled images using only a few labeled samples.
We propose a novel approach based on the non-i.i.d paradigm of gradual machine learning (GML)
We show that the proposed approach can improve the SOTA performance by 1-5% in terms of accuracy.
arXiv Detail & Related papers (2023-07-28T12:30:41Z) - Weakly supervised segmentation with point annotations for histopathology
images via contrast-based variational model [7.021021047695508]
We propose a contrast-based variational model to generate segmentation results for histopathology images.
The proposed method considers the common characteristics of target regions in histopathology images and can be trained in an end-to-end manner.
It can generate more regionally consistent and smoother boundary segmentation, and is more robust to unlabeled novel' regions.
arXiv Detail & Related papers (2023-04-07T10:12:21Z) - Learning Customized Visual Models with Retrieval-Augmented Knowledge [104.05456849611895]
We propose REACT, a framework to acquire the relevant web knowledge to build customized visual models for target domains.
We retrieve the most relevant image-text pairs from the web-scale database as external knowledge, and propose to customize the model by only training new modualized blocks while freezing all the original weights.
The effectiveness of REACT is demonstrated via extensive experiments on classification, retrieval, detection and segmentation tasks, including zero, few, and full-shot settings.
arXiv Detail & Related papers (2023-01-17T18:59:06Z) - Towards Automated Polyp Segmentation Using Weakly- and Semi-Supervised
Learning and Deformable Transformers [8.01814397869811]
Polyp segmentation is a crucial step towards computer-aided diagnosis of colorectal cancer.
Most of the polyp segmentation methods require pixel-wise annotated datasets.
We propose a novel framework that can be trained using only weakly annotated images along with exploiting unlabeled images.
arXiv Detail & Related papers (2022-11-21T20:44:12Z) - EllSeg-Gen, towards Domain Generalization for head-mounted eyetracking [19.913297057204357]
We show that convolutional networks excel at extracting gaze features despite the presence of such artifacts.
We compare the performance of a single model trained with multiple datasets against a pool of models trained on individual datasets.
Results indicate that models tested on datasets in which eye images exhibit higher appearance variability benefit from multiset training.
arXiv Detail & Related papers (2022-05-04T08:35:52Z) - Unsupervised Domain Adaptation with Contrastive Learning for OCT
Segmentation [49.59567529191423]
We propose a novel semi-supervised learning framework for segmentation of volumetric images from new unlabeled domains.
We jointly use supervised and contrastive learning, also introducing a contrastive pairing scheme that leverages similarity between nearby slices in 3D.
arXiv Detail & Related papers (2022-03-07T19:02:26Z) - Towards End-to-end Video-based Eye-Tracking [50.0630362419371]
Estimating eye-gaze from images alone is a challenging task due to un-observable person-specific factors.
We propose a novel dataset and accompanying method which aims to explicitly learn these semantic and temporal relationships.
We demonstrate that the fusion of information from visual stimuli as well as eye images can lead towards achieving performance similar to literature-reported figures.
arXiv Detail & Related papers (2020-07-26T12:39:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.