Revealing the Invisible with Model and Data Shrinking for
Composite-database Micro-expression Recognition
- URL: http://arxiv.org/abs/2006.09674v1
- Date: Wed, 17 Jun 2020 06:19:24 GMT
- Title: Revealing the Invisible with Model and Data Shrinking for
Composite-database Micro-expression Recognition
- Authors: Zhaoqiang Xia, Wei Peng, Huai-Qian Khor, Xiaoyi Feng, Guoying Zhao
- Abstract summary: We analyze the influence of learning complexity, including the input complexity and model complexity.
We propose a recurrent convolutional network (RCN) to explore the shallower-architecture and lower-resolution input data.
We develop three parameter-free modules to integrate with RCN without increasing any learnable parameters.
- Score: 49.463864096615254
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Composite-database micro-expression recognition is attracting increasing
attention as it is more practical to real-world applications. Though the
composite database provides more sample diversity for learning good
representation models, the important subtle dynamics are prone to disappearing
in the domain shift such that the models greatly degrade their performance,
especially for deep models. In this paper, we analyze the influence of learning
complexity, including the input complexity and model complexity, and discover
that the lower-resolution input data and shallower-architecture model are
helpful to ease the degradation of deep models in composite-database task.
Based on this, we propose a recurrent convolutional network (RCN) to explore
the shallower-architecture and lower-resolution input data, shrinking model and
input complexities simultaneously. Furthermore, we develop three parameter-free
modules (i.e., wide expansion, shortcut connection and attention unit) to
integrate with RCN without increasing any learnable parameters. These three
modules can enhance the representation ability in various perspectives while
preserving not-very-deep architecture for lower-resolution data. Besides, three
modules can further be combined by an automatic strategy (a neural architecture
search strategy) and the searched architecture becomes more robust. Extensive
experiments on MEGC2019 dataset (composited of existing SMIC, CASME II and SAMM
datasets) have verified the influence of learning complexity and shown that
RCNs with three modules and the searched combination outperform the
state-of-the-art approaches.
Related papers
- YOLO-RD: Introducing Relevant and Compact Explicit Knowledge to YOLO by Retriever-Dictionary [12.39040757106137]
We introduce an innovative em textbfRetriever-emtextbfDictionary (RD) module to address this issue.
This architecture enables YOLO-based models to efficiently retrieve features from a Dictionary that contains the insight of the dataset.
arXiv Detail & Related papers (2024-10-20T09:38:58Z) - Images in Discrete Choice Modeling: Addressing Data Isomorphism in
Multi-Modality Inputs [77.54052164713394]
This paper explores the intersection of Discrete Choice Modeling (DCM) and machine learning.
We investigate the consequences of embedding high-dimensional image data that shares isomorphic information with traditional tabular inputs within a DCM framework.
arXiv Detail & Related papers (2023-12-22T14:33:54Z) - Serving Deep Learning Model in Relational Databases [70.53282490832189]
Serving deep learning (DL) models on relational data has become a critical requirement across diverse commercial and scientific domains.
We highlight three pivotal paradigms: The state-of-the-art DL-centric architecture offloads DL computations to dedicated DL frameworks.
The potential UDF-centric architecture encapsulates one or more tensor computations into User Defined Functions (UDFs) within the relational database management system (RDBMS)
arXiv Detail & Related papers (2023-10-07T06:01:35Z) - Homological Convolutional Neural Networks [4.615338063719135]
We propose a novel deep learning architecture that exploits the data structural organization through topologically constrained network representations.
We test our model on 18 benchmark datasets against 5 classic machine learning and 3 deep learning models.
arXiv Detail & Related papers (2023-08-26T08:48:51Z) - General-Purpose Multimodal Transformer meets Remote Sensing Semantic
Segmentation [35.100738362291416]
Multimodal AI seeks to exploit complementary data sources, particularly for complex tasks like semantic segmentation.
Recent trends in general-purpose multimodal networks have shown great potential to achieve state-of-the-art performance.
We propose a UNet-inspired module that employs 3D convolution to encode vital local information and learn cross-modal features simultaneously.
arXiv Detail & Related papers (2023-07-07T04:58:34Z) - Neural Attentive Circuits [93.95502541529115]
We introduce a general purpose, yet modular neural architecture called Neural Attentive Circuits (NACs)
NACs learn the parameterization and a sparse connectivity of neural modules without using domain knowledge.
NACs achieve an 8x speedup at inference time while losing less than 3% performance.
arXiv Detail & Related papers (2022-10-14T18:00:07Z) - Adversarial Audio Synthesis with Complex-valued Polynomial Networks [60.231877895663956]
Time-frequency (TF) representations in audio have been increasingly modeled real-valued networks.
We introduce complex-valued networks called APOLLO, that integrate such complex-valued representations in a natural way.
APOLLO results in $17.5%$ improvement over adversarial methods and $8.2%$ over the state-of-the-art diffusion models on SC09 in audio generation.
arXiv Detail & Related papers (2022-06-14T12:58:59Z) - Triple-level Model Inferred Collaborative Network Architecture for Video
Deraining [43.06607185181434]
We develop a model-guided triple-level optimization framework to deduce network architecture with cooperating optimization and auto-searching mechanism.
Our model shows significant improvements in fidelity and temporal consistency over the state-of-the-art works.
arXiv Detail & Related papers (2021-11-08T13:09:00Z) - Deep Submodular Networks for Extractive Data Summarization [0.46898263272139784]
We propose an end-to-end learning framework for summarization problems.
The Deep Submodular Networks (DSN) framework can be used to learn features appropriate for summarization from scratch.
In particular, we show that DSNs outperform simple mixture models using off the shelf features.
arXiv Detail & Related papers (2020-10-16T19:06:15Z) - When Residual Learning Meets Dense Aggregation: Rethinking the
Aggregation of Deep Neural Networks [57.0502745301132]
We propose Micro-Dense Nets, a novel architecture with global residual learning and local micro-dense aggregations.
Our micro-dense block can be integrated with neural architecture search based models to boost their performance.
arXiv Detail & Related papers (2020-04-19T08:34:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.