Involution Fused ConvNet for Classifying Eye-Tracking Patterns of
Children with Autism Spectrum Disorder
- URL: http://arxiv.org/abs/2401.03575v1
- Date: Sun, 7 Jan 2024 20:08:17 GMT
- Title: Involution Fused ConvNet for Classifying Eye-Tracking Patterns of
Children with Autism Spectrum Disorder
- Authors: Md. Farhadul Islam and Meem Arafat Manab and Joyanta Jyoti Mondal and
Sarah Zabeen and Fardin Bin Rahman and Md. Zahidul Hasan and Farig Sadeque
and Jannatun Noor
- Abstract summary: Autism Spectrum Disorder (ASD) is a complicated neurological condition which is challenging to diagnose. Numerous studies demonstrate that children diagnosed with ASD struggle with maintaining attention spans and have less focused vision.
Eye-tracking technology has drawn special attention in the context of ASD since anomalies in gaze have long been acknowledged as a defining feature of autism in general.
- Score: 1.225920962851304
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Autism Spectrum Disorder (ASD) is a complicated neurological condition which
is challenging to diagnose. Numerous studies demonstrate that children
diagnosed with autism struggle with maintaining attention spans and have less
focused vision. The eye-tracking technology has drawn special attention in the
context of ASD since anomalies in gaze have long been acknowledged as a
defining feature of autism in general. Deep Learning (DL) approaches coupled
with eye-tracking sensors are exploiting additional capabilities to advance the
diagnostic and its applications. By learning intricate nonlinear input-output
relations, DL can accurately recognize the various gaze and eye-tracking
patterns and adjust to the data. Convolutions alone are insufficient to capture
the important spatial information in gaze patterns or eye tracking. The dynamic
kernel-based process known as involutions can improve the efficiency of
classifying gaze patterns or eye tracking data. In this paper, we utilise two
different image-processing operations to see how these processes learn
eye-tracking patterns. Since these patterns are primarily based on spatial
information, we use involution with convolution making it a hybrid, which adds
location-specific capability to a deep learning model. Our proposed model is
implemented in a simple yet effective approach, which makes it easier for
applying in real life. We investigate the reasons why our approach works well
for classifying eye-tracking patterns. For comparative analysis, we experiment
with two separate datasets as well as a combined version of both. The results
show that IC with three involution layers outperforms the previous approaches.
Related papers
- Exploring Gaze Pattern in Autistic Children: Clustering, Visualization, and Prediction [9.251838958621684]
We propose a novel method to automatically analyze gaze behaviors in ASD children with superior accuracy.
We first apply and optimize seven clustering algorithms to automatically group gaze points to compare ASD subjects with typically developing peers.
Lastly, using these features as prior knowledge, we train multiple predictive machine learning models to predict and diagnose ASD based on their gaze behaviors.
arXiv Detail & Related papers (2024-09-18T06:56:06Z) - Improving Deep Facial Phenotyping for Ultra-rare Disorder Verification
Using Model Ensembles [52.77024349608834]
We analyze the influence of replacing a DCNN with a state-of-the-art face recognition approach, iResNet with ArcFace.
Our proposed ensemble model achieves state-of-the-art performance on both seen and unseen disorders.
arXiv Detail & Related papers (2022-11-12T23:28:54Z) - Vision-Based Activity Recognition in Children with Autism-Related
Behaviors [15.915410623440874]
We demonstrate the effect of a region-based computer vision system to help clinicians and parents analyze a child's behavior.
The data is pre-processed by detecting the target child in the video to reduce the impact of background noise.
Motivated by the effectiveness of temporal convolutional models, we propose both light-weight and conventional models capable of extracting action features from video frames.
arXiv Detail & Related papers (2022-08-08T15:12:27Z) - Detection of ADHD based on Eye Movements during Natural Viewing [3.1890959219836574]
ADHD is a neurodevelopmental disorder that is highly prevalent and requires clinical specialists to diagnose.
We develop an end-to-end deep learning-based sequence model which we pre-train on a related task.
We find that the method is in fact able to detect ADHD and outperforms relevant baselines.
arXiv Detail & Related papers (2022-07-04T12:56:04Z) - Neurosymbolic hybrid approach to driver collision warning [64.02492460600905]
There are two main algorithmic approaches to autonomous driving systems.
Deep learning alone has achieved state-of-the-art results in many areas.
But sometimes it can be very difficult to debug if the deep learning model doesn't work.
arXiv Detail & Related papers (2022-03-28T20:29:50Z) - Assessing glaucoma in retinal fundus photographs using Deep Feature
Consistent Variational Autoencoders [63.391402501241195]
glaucoma is challenging to detect since it remains asymptomatic until the symptoms are severe.
Early identification of glaucoma is generally made based on functional, structural, and clinical assessments.
Deep learning methods have partially solved this dilemma by bypassing the marker identification stage and analyzing high-level information directly to classify the data.
arXiv Detail & Related papers (2021-10-04T16:06:49Z) - Understanding Character Recognition using Visual Explanations Derived
from the Human Visual System and Deep Networks [6.734853055176694]
We examine the congruence, or lack thereof, in the information-gathering strategies of deep neural networks.
The deep learning model considered similar regions in character, which humans have fixated in the case of correctly classified characters.
We propose to use the visual fixation maps obtained from the eye-tracking experiment as a supervisory input to align the model's focus on relevant character regions.
arXiv Detail & Related papers (2021-08-10T10:09:37Z) - Classifying Eye-Tracking Data Using Saliency Maps [8.524684315458245]
This paper proposes a visual saliency based novel feature extraction method for automatic and quantitative classification of eye-tracking data.
Comparing the saliency amplitudes, similarity and dissimilarity of saliency maps with the corresponding eye fixations maps gives an extra dimension of information which is effectively utilized to generate discriminative features to classify the eye-tracking data.
arXiv Detail & Related papers (2020-10-24T15:18:07Z) - Modeling Shared Responses in Neuroimaging Studies through MultiView ICA [94.31804763196116]
Group studies involving large cohorts of subjects are important to draw general conclusions about brain functional organization.
We propose a novel MultiView Independent Component Analysis model for group studies, where data from each subject are modeled as a linear combination of shared independent sources plus noise.
We demonstrate the usefulness of our approach first on fMRI data, where our model demonstrates improved sensitivity in identifying common sources among subjects.
arXiv Detail & Related papers (2020-06-11T17:29:53Z) - Towards Efficient Processing and Learning with Spikes: New Approaches
for Multi-Spike Learning [59.249322621035056]
We propose two new multi-spike learning rules which demonstrate better performance over other baselines on various tasks.
In the feature detection task, we re-examine the ability of unsupervised STDP with its limitations being presented.
Our proposed learning rules can reliably solve the task over a wide range of conditions without specific constraints being applied.
arXiv Detail & Related papers (2020-05-02T06:41:20Z) - Retinopathy of Prematurity Stage Diagnosis Using Object Segmentation and
Convolutional Neural Networks [68.96150598294072]
Retinopathy of Prematurity (ROP) is an eye disorder primarily affecting premature infants with lower weights.
It causes proliferation of vessels in the retina and could result in vision loss and, eventually, retinal detachment, leading to blindness.
In recent years, there has been a significant effort to automate the diagnosis using deep learning.
This paper builds upon the success of previous models and develops a novel architecture, which combines object segmentation and convolutional neural networks (CNN)
Our proposed system first trains an object segmentation model to identify the demarcation line at a pixel level and adds the resulting mask as an additional "color" channel in
arXiv Detail & Related papers (2020-04-03T14:07:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.