VMSMO: Learning to Generate Multimodal Summary for Video-based News
Articles
- URL: http://arxiv.org/abs/2010.05406v1
- Date: Mon, 12 Oct 2020 02:19:16 GMT
- Title: VMSMO: Learning to Generate Multimodal Summary for Video-based News
Articles
- Authors: Mingzhe Li, Xiuying Chen, Shen Gao, Zhangming Chan, Dongyan Zhao and
Rui Yan
- Abstract summary: We propose the task of Video-based Multimodal Summarization with Multimodal Output (VMSMO)
The main challenge in this task is to jointly model the temporal dependency of video with semantic meaning of article.
We propose a Dual-Interaction-based Multimodal Summarizer (DIMS), consisting of a dual interaction module and multimodal generator.
- Score: 63.32111010686954
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A popular multimedia news format nowadays is providing users with a lively
video and a corresponding news article, which is employed by influential news
media including CNN, BBC, and social media including Twitter and Weibo. In such
a case, automatically choosing a proper cover frame of the video and generating
an appropriate textual summary of the article can help editors save time, and
readers make the decision more effectively. Hence, in this paper, we propose
the task of Video-based Multimodal Summarization with Multimodal Output (VMSMO)
to tackle such a problem. The main challenge in this task is to jointly model
the temporal dependency of video with semantic meaning of article. To this end,
we propose a Dual-Interaction-based Multimodal Summarizer (DIMS), consisting of
a dual interaction module and multimodal generator. In the dual interaction
module, we propose a conditional self-attention mechanism that captures local
semantic information within video and a global-attention mechanism that handles
the semantic relationship between news text and video from a high level.
Extensive experiments conducted on a large-scale real-world VMSMO dataset show
that DIMS achieves the state-of-the-art performance in terms of both automatic
metrics and human evaluations.
Related papers
- Ocean-omni: To Understand the World with Omni-modality [28.306965534325904]
Ocean-omni is the first open-source 7B Multimodal Large Language Model (MLLM)
We introduce Ocean-omni, the first open-source 7B Multimodal Large Language Model (MLLM)
arXiv Detail & Related papers (2024-10-11T06:44:31Z) - Tuning Large Multimodal Models for Videos using Reinforcement Learning from AI Feedback [38.708690624594794]
Video and text multimodal alignment remains challenging, primarily due to the deficient volume and quality of multimodal instruction-tune data.
We present a novel alignment strategy that employs multimodal AI system to oversee itself called Reinforcement Learning from AI Feedback (RLAIF)
In specific, we propose context-aware reward modeling by providing detailed video descriptions as context during the generation of preference feedback.
arXiv Detail & Related papers (2024-02-06T06:27:40Z) - MVBench: A Comprehensive Multi-modal Video Understanding Benchmark [63.14000659130736]
We introduce a comprehensive Multi-modal Video understanding Benchmark, namely MVBench.
We first introduce a novel static-to-dynamic method to define these temporal-related tasks.
Then, guided by the task definition, we automatically convert public video annotations into multiple-choice QA to evaluate each task.
arXiv Detail & Related papers (2023-11-28T17:59:04Z) - mPLUG-2: A Modularized Multi-modal Foundation Model Across Text, Image
and Video [89.19867891570945]
mPLUG-2 is a new unified paradigm with modularized design for multi-modal pretraining.
It shares common universal modules for modality collaboration and disentangling different modality modules to deal with modality entanglement.
It is flexible to select different modules for different understanding and generation tasks across all modalities including text, image, and video.
arXiv Detail & Related papers (2023-02-01T12:40:03Z) - TLDW: Extreme Multimodal Summarisation of News Videos [76.50305095899958]
We introduce eXtreme Multimodal Summarisation with Multimodal Output (XMSMO) for the scenario of TL;DW - Too Long; Didn't Watch, akin to TL;DR.
XMSMO aims to summarise a video-document pair into a summary with an extremely short length, which consists of one cover frame as the visual summary and one sentence as the textual summary.
Our method is trained, without using reference summaries, by optimising the visual and textual coverage from the perspectives of the distance between the semantic distributions under optimal transport plans.
arXiv Detail & Related papers (2022-10-16T08:19:59Z) - Semantics-Consistent Cross-domain Summarization via Optimal Transport
Alignment [80.18786847090522]
We propose a Semantics-Consistent Cross-domain Summarization model based on optimal transport alignment with visual and textual segmentation.
We evaluated our method on three recent multimodal datasets and demonstrated the effectiveness of our method in producing high-quality multimodal summaries.
arXiv Detail & Related papers (2022-10-10T14:27:10Z) - MHMS: Multimodal Hierarchical Multimedia Summarization [80.18786847090522]
We propose a multimodal hierarchical multimedia summarization (MHMS) framework by interacting visual and language domains.
Our method contains video and textual segmentation and summarization module, respectively.
It formulates a cross-domain alignment objective with optimal transport distance to generate the representative and textual summary.
arXiv Detail & Related papers (2022-04-07T21:00:40Z) - See, Hear, Read: Leveraging Multimodality with Guided Attention for
Abstractive Text Summarization [14.881597737762316]
We introduce the first large-scale dataset for abstractive text summarization with videos of diverse duration, compiled from presentations in well-known academic conferences like NDSS, ICML, NeurIPS, etc.
We then propose name, a factorized multi-modal Transformer based decoder-only language model, which inherently captures the intra-modal and inter-modal dynamics within various input modalities for the text summarization task.
arXiv Detail & Related papers (2021-05-20T08:56:33Z) - GPT2MVS: Generative Pre-trained Transformer-2 for Multi-modal Video
Summarization [18.543372365239673]
The proposed model consists of a contextualized video summary controller, multi-modal attention mechanisms, an interactive attention network, and a video summary generator.
Results show that the proposed model is effective with the increase of +5.88% in accuracy and +4.06% increase of F1-score, compared with the state-of-the-art method.
arXiv Detail & Related papers (2021-04-26T10:50:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.