Towards Real-time Drowsiness Detection for Elderly Care
- URL: http://arxiv.org/abs/2010.10771v1
- Date: Wed, 21 Oct 2020 05:48:59 GMT
- Title: Towards Real-time Drowsiness Detection for Elderly Care
- Authors: Boris Ba\v{c}i\'c and Jason Zhang
- Abstract summary: This paper produces a proof of concept extracting drowsiness information from videos to help elderly living on their own.
To yawning, eyelid and head movement over time, we extracted 3000 images from videos for training and testing of deep learning models integrated with OpenCV.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The primary focus of this paper is to produce a proof of concept for
extracting drowsiness information from videos to help elderly living on their
own. To quantify yawning, eyelid and head movement over time, we extracted 3000
images from captured videos for training and testing of deep learning models
integrated with OpenCV library. The achieved classification accuracy for eyelid
and mouth open/close status were between 94.3%-97.2%. Visual inspection of head
movement from videos with generated 3D coordinate overlays, indicated clear
spatiotemporal patterns in collected data (yaw, roll and pitch). Extraction
methodology of the drowsiness information as timeseries is applicable to other
contexts including support for prior work in privacy-preserving augmented
coaching, sport rehabilitation, and integration with big data platform in
healthcare.
Related papers
- Blind Dates: Examining the Expression of Temporality in Historical
Photographs [57.07335632641355]
We investigate the dating of images using OpenCLIP, an open-source implementation of CLIP, a multi-modal language and vision model.
We use the textitDe Boer Scene Detection dataset, containing 39,866 gray-scale historical press photographs from 1950 to 1999.
Our analysis reveals that images featuring buses, cars, cats, dogs, and people are more accurately dated, suggesting the presence of temporal markers.
arXiv Detail & Related papers (2023-10-10T13:51:24Z) - Time Does Tell: Self-Supervised Time-Tuning of Dense Image
Representations [79.87044240860466]
We propose a novel approach that incorporates temporal consistency in dense self-supervised learning.
Our approach, which we call time-tuning, starts from image-pretrained models and fine-tunes them with a novel self-supervised temporal-alignment clustering loss on unlabeled videos.
Time-tuning improves the state-of-the-art by 8-10% for unsupervised semantic segmentation on videos and matches it for images.
arXiv Detail & Related papers (2023-08-22T21:28:58Z) - Pose2Gait: Extracting Gait Features from Monocular Video of Individuals
with Dementia [3.2739089842471136]
Video-based ambient monitoring of gait for older adults with dementia has the potential to detect negative changes in health.
Computer vision-based pose tracking models can process video data automatically and extract joint locations.
These models are not optimized for gait analysis on older adults or clinical populations.
arXiv Detail & Related papers (2023-08-22T14:59:17Z) - Segmentation of kidney stones in endoscopic video feeds [2.572404739180802]
We describe how we built a dataset from the raw videos and how we developed a pipeline to automate as much of the process as possible.
To show clinical potential for real-time use, we also confirmed that our best trained model can accurately annotate new videos at 30 frames per second.
arXiv Detail & Related papers (2022-04-29T16:00:52Z) - Using Human Gaze For Surgical Activity Recognition [0.40611352512781856]
We propose to use human gaze with a spatial temporal attention mechanism for activity recognition in surgical videos.
Our model consists of an I3D-based architecture, learns temporal features using 3D convolutions, as well as learning an attention map using human gaze.
arXiv Detail & Related papers (2022-03-09T14:28:00Z) - Video Salient Object Detection via Contrastive Features and Attention
Modules [106.33219760012048]
We propose a network with attention modules to learn contrastive features for video salient object detection.
A co-attention formulation is utilized to combine the low-level and high-level features.
We show that the proposed method requires less computation, and performs favorably against the state-of-the-art approaches.
arXiv Detail & Related papers (2021-11-03T17:40:32Z) - Self-supervised Video Representation Learning by Uncovering
Spatio-temporal Statistics [74.6968179473212]
This paper proposes a novel pretext task to address the self-supervised learning problem.
We compute a series of partitioning-temporal statistical summaries, such as the spatial location and dominant direction of the largest motion.
A neural network is built and trained to yield the statistical summaries given the video frames as inputs.
arXiv Detail & Related papers (2020-08-31T08:31:56Z) - A Plug-and-play Scheme to Adapt Image Saliency Deep Model for Video Data [54.198279280967185]
This paper proposes a novel plug-and-play scheme to weakly retrain a pretrained image saliency deep model for video data.
Our method is simple yet effective for adapting any off-the-shelf pre-trained image saliency deep model to obtain high-quality video saliency detection.
arXiv Detail & Related papers (2020-08-02T13:23:14Z) - Automatic Operating Room Surgical Activity Recognition for
Robot-Assisted Surgery [1.1033115844630357]
We investigate automatic surgical activity recognition in robot-assisted operations.
We collect the first large-scale dataset including 400 full-length multi-perspective videos.
We densely annotate the videos with 10 most recognized and clinically relevant classes of activities.
arXiv Detail & Related papers (2020-06-29T16:30:31Z) - Self-Supervised Human Depth Estimation from Monocular Videos [99.39414134919117]
Previous methods on estimating detailed human depth often require supervised training with ground truth' depth data.
This paper presents a self-supervised method that can be trained on YouTube videos without known depth.
Experiments demonstrate that our method enjoys better generalization and performs much better on data in the wild.
arXiv Detail & Related papers (2020-05-07T09:45:11Z) - Deep Reinforcement Learning for Active Human Pose Estimation [35.229529080763925]
We introduce Pose-DRL, a fully trainable deep reinforcement learning-based active pose estimation architecture.
We show that our model learns to select viewpoints that yield significantly more accurate pose estimates compared to strong multi-view baselines.
arXiv Detail & Related papers (2020-01-07T13:35:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.