Automated Artefact Relevancy Determination from Artefact Metadata and
Associated Timeline Events
- URL: http://arxiv.org/abs/2012.01972v1
- Date: Wed, 2 Dec 2020 14:14:26 GMT
- Title: Automated Artefact Relevancy Determination from Artefact Metadata and
Associated Timeline Events
- Authors: Xiaoyu Du, Quan Le and Mark Scanlon
- Abstract summary: Case-hindering, multi-year digital forensic evidence backlogs have become commonplace in law enforcement agencies throughout the world.
This is due to an ever-growing number of cases requiring digital forensic investigation coupled with the growing volume of data to be processed per case.
Leveraging previously processed digital forensic cases and their component artefact relevancy classifications can facilitate an opportunity for training automated artificial intelligence based evidence processing systems.
- Score: 7.219077740523683
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Case-hindering, multi-year digital forensic evidence backlogs have become
commonplace in law enforcement agencies throughout the world. This is due to an
ever-growing number of cases requiring digital forensic investigation coupled
with the growing volume of data to be processed per case. Leveraging previously
processed digital forensic cases and their component artefact relevancy
classifications can facilitate an opportunity for training automated artificial
intelligence based evidence processing systems. These can significantly aid
investigators in the discovery and prioritisation of evidence. This paper
presents one approach for file artefact relevancy determination building on the
growing trend towards a centralised, Digital Forensics as a Service (DFaaS)
paradigm. This approach enables the use of previously encountered pertinent
files to classify newly discovered files in an investigation. Trained models
can aid in the detection of these files during the acquisition stage, i.e.,
during their upload to a DFaaS system. The technique generates a relevancy
score for file similarity using each artefact's filesystem metadata and
associated timeline events. The approach presented is validated against three
experimental usage scenarios.
Related papers
- Advancing Cyber Incident Timeline Analysis Through Rule Based AI and Large Language Models [0.0]
This paper introduces a novel framework, GenDFIR, which combines Rule-Based Artificial Intelligence (R-BAI) algorithms with Large Language Models (LLMs) to enhance and automate the Timeline Analysis process.
arXiv Detail & Related papers (2024-09-04T09:46:33Z) - Behind the (Digital Crime) Scenes: An MSC Model [0.0]
The establishment of digital forensics as a foundational discipline for extracting digital evidence further exacerbates the complex nature of criminal investigations.
We delineate the protocols that compose digital forensics within a criminal case, formalise them as message sequence charts (MSCs) and identify their functional requirements.
arXiv Detail & Related papers (2024-03-24T15:29:08Z) - Automatically Summarizing Evidence from Clinical Trials: A Prototype
Highlighting Current Challenges [20.74608114488094]
TrialsSummarizer aims to automatically summarize evidence presented in the set of randomized controlled trials most relevant to a given query.
System retrieves trial publications matching a query specifying a combination of condition, intervention(s), and outcome(s)
Top-k such studies are passed through a neural multi-document summarization system, yielding a synopsis of these trials.
arXiv Detail & Related papers (2023-03-07T17:30:48Z) - Natural Logic-guided Autoregressive Multi-hop Document Retrieval for
Fact Verification [21.04611844009438]
We propose a novel retrieve-and-rerank method for multi-hop retrieval.
It consists of a retriever that jointly scores documents in the knowledge source and sentences from previously retrieved documents.
It is guided by a proof system that dynamically terminates the retrieval process if the evidence is deemed sufficient.
arXiv Detail & Related papers (2022-12-10T11:32:38Z) - Questions Are All You Need to Train a Dense Passage Retriever [123.13872383489172]
ART is a new corpus-level autoencoding approach for training dense retrieval models that does not require any labeled training data.
It uses a new document-retrieval autoencoding scheme, where (1) an input question is used to retrieve a set of evidence documents, and (2) the documents are then used to compute the probability of reconstructing the original question.
arXiv Detail & Related papers (2022-06-21T18:16:31Z) - GERE: Generative Evidence Retrieval for Fact Verification [57.78768817972026]
We propose GERE, the first system that retrieves evidences in a generative fashion.
The experimental results on the FEVER dataset show that GERE achieves significant improvements over the state-of-the-art baselines.
arXiv Detail & Related papers (2022-04-12T03:49:35Z) - Human-in-the-Loop Disinformation Detection: Stance, Sentiment, or
Something Else? [93.91375268580806]
Both politics and pandemics have recently provided ample motivation for the development of machine learning-enabled disinformation (a.k.a. fake news) detection algorithms.
Existing literature has focused primarily on the fully-automated case, but the resulting techniques cannot reliably detect disinformation on the varied topics, sources, and time scales required for military applications.
By leveraging an already-available analyst as a human-in-the-loop, canonical machine learning techniques of sentiment analysis, aspect-based sentiment analysis, and stance detection become plausible methods to use for a partially-automated disinformation detection system.
arXiv Detail & Related papers (2021-11-09T13:30:34Z) - MD-CSDNetwork: Multi-Domain Cross Stitched Network for Deepfake
Detection [80.83725644958633]
Current deepfake generation methods leave discriminative artifacts in the frequency spectrum of fake images and videos.
We present a novel approach, termed as MD-CSDNetwork, for combining the features in the spatial and frequency domains to mine a shared discriminative representation.
arXiv Detail & Related papers (2021-09-15T14:11:53Z) - Combining Feature and Instance Attribution to Detect Artifacts [62.63504976810927]
We propose methods to facilitate identification of training data artifacts.
We show that this proposed training-feature attribution approach can be used to uncover artifacts in training data.
We execute a small user study to evaluate whether these methods are useful to NLP researchers in practice.
arXiv Detail & Related papers (2021-07-01T09:26:13Z) - Fast(er) Reconstruction of Shredded Text Documents via Self-Supervised
Deep Asymmetric Metric Learning [62.34197797857823]
A central problem in automatic reconstruction of shredded documents is the pairwise compatibility evaluation of the shreds.
This work proposes a scalable deep learning approach for measuring pairwise compatibility in which the number of inferences scales linearly.
Our method has accuracy comparable to the state-of-the-art with a speed-up of about 22 times for a test instance with 505 shreds.
arXiv Detail & Related papers (2020-03-23T03:22:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.