Logically at Factify 2023: A Multi-Modal Fact Checking System Based on
Evidence Retrieval techniques and Transformer Encoder Architecture
- URL: http://arxiv.org/abs/2301.03127v1
- Date: Mon, 9 Jan 2023 00:19:11 GMT
- Title: Logically at Factify 2023: A Multi-Modal Fact Checking System Based on
Evidence Retrieval techniques and Transformer Encoder Architecture
- Authors: Pim Jordi Verschuuren, Jie Gao, Adelize van Eeden, Stylianos Oikonomou
and Anil Bandhakavi
- Abstract summary: We present the Logically submissions to De-Factify 2 challenge (DE-FACTIFY 2023) on the task 1 of Multi-Modal Fact Checking.
We describe our submissions to this challenge including explored evidence retrieval and selection techniques, pre-trained cross-modal and unimodal models, and a cross-modal veracity model based on the well established Transformer (TE) architecture.
The final system, a standard two-stage evidence based veracity detection system, yields weighted avg. 0.79 on both val set and final blind test set on the task 1, which achieves 3rd place with a small margin to the top
- Score: 3.7529756903595963
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In this paper, we present the Logically submissions to De-Factify 2 challenge
(DE-FACTIFY 2023) on the task 1 of Multi-Modal Fact Checking. We describes our
submissions to this challenge including explored evidence retrieval and
selection techniques, pre-trained cross-modal and unimodal models, and a
cross-modal veracity model based on the well established Transformer Encoder
(TE) architecture which is heavily relies on the concept of self-attention.
Exploratory analysis is also conducted on this Factify 2 data set that uncovers
the salient multi-modal patterns and hypothesis motivating the architecture
proposed in this work. A series of preliminary experiments were done to
investigate and benchmarking different pre-trained embedding models, evidence
retrieval settings and thresholds. The final system, a standard two-stage
evidence based veracity detection system, yields weighted avg. 0.79 on both val
set and final blind test set on the task 1, which achieves 3rd place with a
small margin to the top performing system on the leaderboard among 9
participants.
Related papers
- SmurfCat at SemEval-2024 Task 6: Leveraging Synthetic Data for Hallucination Detection [51.99159169107426]
We present our novel systems developed for the SemEval-2024 hallucination detection task.
Our investigation spans a range of strategies to compare model predictions with reference standards.
We introduce three distinct methods that exhibit strong performance metrics.
arXiv Detail & Related papers (2024-04-09T09:03:44Z) - MTP: Advancing Remote Sensing Foundation Model via Multi-Task Pretraining [73.81862342673894]
Foundation models have reshaped the landscape of Remote Sensing (RS) by enhancing various image interpretation tasks.
transferring the pretrained models to downstream tasks may encounter task discrepancy due to their formulation of pretraining as image classification or object discrimination tasks.
We conduct multi-task supervised pretraining on the SAMRS dataset, encompassing semantic segmentation, instance segmentation, and rotated object detection.
Our models are finetuned on various RS downstream tasks, such as scene classification, horizontal and rotated object detection, semantic segmentation, and change detection.
arXiv Detail & Related papers (2024-03-20T09:17:22Z) - Team Trifecta at Factify5WQA: Setting the Standard in Fact Verification with Fine-Tuning [8.292728619484636]
We present Pre-CoFactv3, a comprehensive framework comprised of Question Answering and Text Classification components for fact verification.
Our team secured first place in the AAAI-24 Factify 3.0 Workshop, surpassing the baseline accuracy by 103% and maintaining a 70% lead over the second competitor.
arXiv Detail & Related papers (2024-03-15T13:24:28Z) - X-InstructBLIP: A Framework for aligning X-Modal instruction-aware representations to LLMs and Emergent Cross-modal Reasoning [109.9413329636322]
This paper introduces an efficient framework that integrates multiple modalities (images, 3D, audio and video) to a frozen Large Language Models (LLMs)
Our approach explores two distinct projection mechanisms: Q-Formers and Linear Projections (LPs)
arXiv Detail & Related papers (2023-11-30T18:43:51Z) - An Empirical Study of Multimodal Model Merging [148.48412442848795]
Model merging is a technique that fuses multiple models trained on different tasks to generate a multi-task solution.
We conduct our study for a novel goal where we can merge vision, language, and cross-modal transformers of a modality-specific architecture.
We propose two metrics that assess the distance between weights to be merged and can serve as an indicator of the merging outcomes.
arXiv Detail & Related papers (2023-04-28T15:43:21Z) - Automatically Summarizing Evidence from Clinical Trials: A Prototype
Highlighting Current Challenges [20.74608114488094]
TrialsSummarizer aims to automatically summarize evidence presented in the set of randomized controlled trials most relevant to a given query.
System retrieves trial publications matching a query specifying a combination of condition, intervention(s), and outcome(s)
Top-k such studies are passed through a neural multi-document summarization system, yielding a synopsis of these trials.
arXiv Detail & Related papers (2023-03-07T17:30:48Z) - Team Triple-Check at Factify 2: Parameter-Efficient Large Foundation
Models with Feature Representations for Multi-Modal Fact Verification [5.552606716659022]
Multi-modal fact verification has become an important but challenging issue on social media.
In this paper, we propose the Pre-CoFactv2 framework for modeling fine-grained text and input embeddings with lightening parameters.
We show that Pre-CoFactv2 outperforms Pre-CoFact by a large margin and achieved new state-of-the-art results at the Factify challenge at AAAI 2023.
arXiv Detail & Related papers (2023-02-12T18:08:54Z) - Logically at the Factify 2022: Multimodal Fact Verification [2.8914815569249823]
This paper describes our participant system for the multi-modal fact verification (Factify) challenge at AAAI 2022.
Two baseline approaches are proposed and explored including an ensemble model and a multi-modal attention network.
Our best model is ranked first in leaderboard which obtains a weighted average F-measure of 0.77 on both validation and test set.
arXiv Detail & Related papers (2021-12-16T23:34:07Z) - Uni-Perceiver: Pre-training Unified Architecture for Generic Perception
for Zero-shot and Few-shot Tasks [73.63892022944198]
We present a generic perception architecture named Uni-Perceiver.
It processes a variety of modalities and tasks with unified modeling and shared parameters.
Results show that our pre-trained model without any tuning can achieve reasonable performance even on novel tasks.
arXiv Detail & Related papers (2021-12-02T18:59:50Z) - A Multi-Level Attention Model for Evidence-Based Fact Checking [58.95413968110558]
We present a simple model that can be trained on sequence structures.
Results on a large-scale dataset for Fact Extraction and VERification show that our model outperforms the graph-based approaches.
arXiv Detail & Related papers (2021-06-02T05:40:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.