Real-Time Face and Landmark Localization for Eyeblink Detection
- URL: http://arxiv.org/abs/2006.00816v2
- Date: Wed, 15 Jul 2020 16:16:11 GMT
- Title: Real-Time Face and Landmark Localization for Eyeblink Detection
- Authors: Paul Bakker, Henk-Jan Boele, Zaid Al-Ars and Christos Strydis
- Abstract summary: A face- and landmark-detection algorithm have been carefully combined in order to provide fully automated eyelid tracking.
Our experimental evaluation reveals an overall runtime application of 0.533 ms per frame.
This is 1,101$times$ faster than the sequential implementation and well within the real-time requirements of eyeblink conditioning in humans.
- Score: 1.617280000789628
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Pavlovian eyeblink conditioning is a powerful experiment used in the field of
neuroscience to measure multiple aspects of how we learn in our daily life. To
track the movement of the eyelid during an experiment, researchers have
traditionally made use of potentiometers or electromyography. More recently,
the use of computer vision and image processing alleviated the need for these
techniques but currently employed methods require human intervention and are
not fast enough to enable real-time processing. In this work, a face- and
landmark-detection algorithm have been carefully combined in order to provide
fully automated eyelid tracking, and have further been accelerated to make the
first crucial step towards online, closed-loop experiments. Such experiments
have not been achieved so far and are expected to offer significant insights in
the workings of neurological and psychiatric disorders. Based on an extensive
literature search, various different algorithms for face detection and landmark
detection have been analyzed and evaluated. Two algorithms were identified as
most suitable for eyelid detection: the Histogram-of-Oriented-Gradients (HOG)
algorithm for face detection and the Ensemble-of-Regression-Trees (ERT)
algorithm for landmark detection. These two algorithms have been accelerated on
GPU and CPU, achieving speedups of 1,753$\times$ and 11$\times$, respectively.
To demonstrate the usefulness of our eyelid-detection algorithm, a research
hypothesis was formed and a well-established neuroscientific experiment was
employed: eyeblink detection. Our experimental evaluation reveals an overall
application runtime of 0.533 ms per frame, which is 1,101$\times$ faster than
the sequential implementation and well within the real-time requirements of
eyeblink conditioning in humans, i.e. faster than 500 frames per second.
Related papers
- A Framework for Pupil Tracking with Event Cameras [1.708806485130162]
Saccades are extremely rapid movements of both eyes that occur simultaneously.
The peak angular speed of the eye during a saccade can reach as high as 700deg/s in humans.
We present events as frames that can be readily utilized by standard deep learning algorithms.
arXiv Detail & Related papers (2024-07-23T17:32:02Z) - Spatiotemporal Pyramidal CNN with Depth-Wise Separable Convolution for
Eye Blinking Detection in the Wild [0.0]
Eye blinking detection plays an essential role in deception detection, driving fatigue detection, etc.
Two problems are addressed: how the eye blinking detection model can learn efficiently from different resolutions of eye pictures in diverse conditions; and how to reduce the size of the detection model for faster inference time.
arXiv Detail & Related papers (2023-06-20T04:59:09Z) - Measuring Human Perception to Improve Open Set Recognition [4.124573231232705]
Human ability to recognize when an object belongs or does not belong to a particular vision task outperforms all open set recognition algorithms.
measured reaction time from human subjects can offer insight as to whether a class sample is prone to be confused with a different class.
New psychophysical loss function enforces consistency with human behavior in deep networks which exhibit variable reaction time for different images.
arXiv Detail & Related papers (2022-09-08T01:19:36Z) - Human Activity Recognition Using Cascaded Dual Attention CNN and
Bi-Directional GRU Framework [3.3721926640077795]
Vision-based human activity recognition has emerged as one of the essential research areas in video analytics domain.
This paper presents a computationally efficient yet generic spatial-temporal cascaded framework that exploits the deep discriminative spatial and temporal features for human activity recognition.
The proposed framework attains an improvement in execution time up to 167 times in terms of frames per second as compared to most of the contemporary action recognition methods.
arXiv Detail & Related papers (2022-08-09T20:34:42Z) - A Deep Learning Approach for the Segmentation of Electroencephalography
Data in Eye Tracking Applications [56.458448869572294]
We introduce DETRtime, a novel framework for time-series segmentation of EEG data.
Our end-to-end deep learning-based framework brings advances in Computer Vision to the forefront.
Our model generalizes well in the task of EEG sleep stage segmentation.
arXiv Detail & Related papers (2022-06-17T10:17:24Z) - PosePipe: Open-Source Human Pose Estimation Pipeline for Clinical
Research [0.0]
We develop a human pose estimation pipeline that facilitates running state-of-the-art algorithms on data acquired in clinical context.
Our goal in this work is not to train new algorithms, but to advance the use of cutting-edge human pose estimation algorithms for clinical and translation research.
arXiv Detail & Related papers (2022-03-16T17:54:37Z) - MTCD: Cataract Detection via Near Infrared Eye Images [69.62768493464053]
cataract is a common eye disease and one of the leading causes of blindness and vision impairment.
We present a novel algorithm for cataract detection using near-infrared eye images.
Deep learning-based eye segmentation and multitask network classification networks are presented.
arXiv Detail & Related papers (2021-10-06T08:10:28Z) - Optimal Sequential Detection of Signals with Unknown Appearance and
Disappearance Points in Time [64.26593350748401]
The paper addresses a sequential changepoint detection problem, assuming that the duration of change may be finite and unknown.
We focus on a reliable maximin change detection criterion of maximizing the minimal probability of detection in a given time (or space) window.
The FMA algorithm is applied to detecting faint streaks of satellites in optical images.
arXiv Detail & Related papers (2021-02-02T04:58:57Z) - Searching for Efficient Architecture for Instrument Segmentation in
Robotic Surgery [58.63306322525082]
Most applications rely on accurate real-time segmentation of high-resolution surgical images.
We design a light-weight and highly-efficient deep residual architecture which is tuned to perform real-time inference of high-resolution images.
arXiv Detail & Related papers (2020-07-08T21:38:29Z) - Detecting Parkinsonian Tremor from IMU Data Collected In-The-Wild using
Deep Multiple-Instance Learning [59.74684475991192]
Parkinson's Disease (PD) is a slowly evolving neuro-logical disease that affects about 1% of the population above 60 years old.
PD symptoms include tremor, rigidity and braykinesia.
We present a method for automatically identifying tremorous episodes related to PD, based on IMU signals captured via a smartphone device.
arXiv Detail & Related papers (2020-05-06T09:02:30Z) - Robust Medical Instrument Segmentation Challenge 2019 [56.148440125599905]
Intraoperative tracking of laparoscopic instruments is often a prerequisite for computer and robotic-assisted interventions.
Our challenge was based on a surgical data set comprising 10,040 annotated images acquired from a total of 30 surgical procedures.
The results confirm the initial hypothesis, namely that algorithm performance degrades with an increasing domain gap.
arXiv Detail & Related papers (2020-03-23T14:35:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.