CDN-MEDAL: Two-stage Density and Difference Approximation Framework for
Motion Analysis
- URL: http://arxiv.org/abs/2106.03776v1
- Date: Mon, 7 Jun 2021 16:39:42 GMT
- Title: CDN-MEDAL: Two-stage Density and Difference Approximation Framework for
Motion Analysis
- Authors: Synh Viet-Uyen Ha, Cuong Tien Nguyen, Hung Ngoc Phan, Nhat Minh Chung,
Phuong Hoai Ha
- Abstract summary: We propose a novel, two-stage method of change detection with two convolutional neural networks.
Our two-stage framework contains approximately 3.5K parameters in total but still maintains rapid convergence to intricate motion patterns.
- Score: 3.337126420148156
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Background modeling is a promising research area in video analysis with a
variety of video surveillance applications. Recent years have witnessed the
proliferation of deep neural networks via effective learning-based approaches
in motion analysis. However, these techniques only provide a limited
description of the observed scenes' insufficient properties where a
single-valued mapping is learned to approximate the temporal conditional
averages of the target background. On the other hand, statistical learning in
imagery domains has become one of the most prevalent approaches with high
adaptation to dynamic context transformation, notably Gaussian Mixture Models,
combined with a foreground extraction step. In this work, we propose a novel,
two-stage method of change detection with two convolutional neural networks.
The first architecture is grounded on the unsupervised Gaussian mixtures
statistical learning to describe the scenes' salient features. The second one
implements a light-weight pipeline of foreground detection. Our two-stage
framework contains approximately 3.5K parameters in total but still maintains
rapid convergence to intricate motion patterns. Our experiments on publicly
available datasets show that our proposed networks are not only capable of
generalizing regions of moving objects in unseen cases with promising results
but also are competitive in performance efficiency and effectiveness regarding
foreground segmentation.
Related papers
- Open Set Action Recognition via Multi-Label Evidential Learning [25.15753429188536]
We propose a new method for open set action recognition and novelty detection via MUlti-Label Evidential learning (MULE)
Our Beta Evidential Neural Network estimates multi-action uncertainty with Beta densities based on actor-context-object relation representations.
Our proposed approach achieves promising performance in single/multi-actor, single/multi-action settings.
arXiv Detail & Related papers (2023-02-27T18:34:18Z) - Dyna-DepthFormer: Multi-frame Transformer for Self-Supervised Depth
Estimation in Dynamic Scenes [19.810725397641406]
We propose a novel Dyna-Depthformer framework, which predicts scene depth and 3D motion field jointly.
Our contributions are two-fold. First, we leverage multi-view correlation through a series of self- and cross-attention layers in order to obtain enhanced depth feature representation.
Second, we propose a warping-based Motion Network to estimate the motion field of dynamic objects without using semantic prior.
arXiv Detail & Related papers (2023-01-14T09:43:23Z) - Video Segmentation Learning Using Cascade Residual Convolutional Neural
Network [0.0]
We propose a novel deep learning video segmentation approach that incorporates residual information into the foreground detection learning process.
Experiments conducted on the Change Detection 2014 and on the private dataset PetrobrasROUTES from Petrobras support the effectiveness of the proposed approach.
arXiv Detail & Related papers (2022-12-20T16:56:54Z) - Learning Multi-Granular Spatio-Temporal Graph Network for Skeleton-based
Action Recognition [49.163326827954656]
We propose a novel multi-granular-temporal graph network for skeleton-based action classification.
We develop a dual-head graph network consisting of two inter-leaved branches, which enables us to extract at least two-temporal resolutions.
We conduct extensive experiments on three large-scale datasets.
arXiv Detail & Related papers (2021-08-10T09:25:07Z) - EAN: Event Adaptive Network for Enhanced Action Recognition [66.81780707955852]
We propose a unified action recognition framework to investigate the dynamic nature of video content.
First, when extracting local cues, we generate the spatial-temporal kernels of dynamic-scale to adaptively fit the diverse events.
Second, to accurately aggregate these cues into a global video representation, we propose to mine the interactions only among a few selected foreground objects by a Transformer.
arXiv Detail & Related papers (2021-07-22T15:57:18Z) - Learning to Segment Rigid Motions from Two Frames [72.14906744113125]
We propose a modular network, motivated by a geometric analysis of what independent object motions can be recovered from an egomotion field.
It takes two consecutive frames as input and predicts segmentation masks for the background and multiple rigidly moving objects, which are then parameterized by 3D rigid transformations.
Our method achieves state-of-the-art performance for rigid motion segmentation on KITTI and Sintel.
arXiv Detail & Related papers (2021-01-11T04:20:30Z) - Adversarial Bipartite Graph Learning for Video Domain Adaptation [50.68420708387015]
Domain adaptation techniques, which focus on adapting models between distributionally different domains, are rarely explored in the video recognition area.
Recent works on visual domain adaptation which leverage adversarial learning to unify the source and target video representations are not highly effective on the videos.
This paper proposes an Adversarial Bipartite Graph (ABG) learning framework which directly models the source-target interactions.
arXiv Detail & Related papers (2020-07-31T03:48:41Z) - One-Shot Object Detection without Fine-Tuning [62.39210447209698]
We introduce a two-stage model consisting of a first stage Matching-FCOS network and a second stage Structure-Aware Relation Module.
We also propose novel training strategies that effectively improve detection performance.
Our method exceeds the state-of-the-art one-shot performance consistently on multiple datasets.
arXiv Detail & Related papers (2020-05-08T01:59:23Z) - Depthwise Non-local Module for Fast Salient Object Detection Using a
Single Thread [136.2224792151324]
We propose a new deep learning algorithm for fast salient object detection.
The proposed algorithm achieves competitive accuracy and high inference efficiency simultaneously with a single CPU thread.
arXiv Detail & Related papers (2020-01-22T15:23:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.