Framework for developing quantitative agent based models based on
qualitative expert knowledge: an organised crime use-case
- URL: http://arxiv.org/abs/2308.00505v1
- Date: Fri, 21 Jul 2023 11:26:54 GMT
- Title: Framework for developing quantitative agent based models based on
qualitative expert knowledge: an organised crime use-case
- Authors: Frederike Oetker, Vittorio Nespeca, Thijs Vis, Paul Duijn, Peter
Sloot, Rick Quax
- Abstract summary: In order to model criminal networks for law enforcement purposes, a limited supply of data needs to be translated into validated agent-based models.
What is missing in current criminological modelling is a systematic and transparent framework for modelers and domain experts.
We propose FREIDA ( Framework for Expert-Informed Data-driven Agent-based models) for this purpose.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In order to model criminal networks for law enforcement purposes, a limited
supply of data needs to be translated into validated agent-based models. What
is missing in current criminological modelling is a systematic and transparent
framework for modelers and domain experts that establishes a modelling
procedure for computational criminal modelling that includes translating
qualitative data into quantitative rules. For this, we propose FREIDA
(Framework for Expert-Informed Data-driven Agent-based models). Throughout the
paper, the criminal cocaine replacement model (CCRM) will be used as an example
case to demonstrate the FREIDA methodology. For the CCRM, a criminal cocaine
network in the Netherlands is being modelled where the kingpin node is being
removed, the goal being for the remaining agents to reorganize after the
disruption and return the network into a stable state. Qualitative data sources
such as case files, literature and interviews are translated into empirical
laws, and combined with the quantitative sources such as databases form the
three dimensions (environment, agents, behaviour) of a networked ABM. Four case
files are being modelled and scored both for training as well as for validation
scores to transition to the computational model and application phase
respectively. In the last phase, iterative sensitivity analysis, uncertainty
quantification and scenario testing eventually lead to a robust model that can
help law enforcement plan their intervention strategies. Results indicate the
need for flexible parameters as well as additional case file simulations to be
performed.
Related papers
- Efficient Multi-Agent System Training with Data Influence-Oriented Tree Search [59.75749613951193]
We propose Data Influence-oriented Tree Search (DITS) to guide both tree search and data selection.
By leveraging influence scores, we effectively identify the most impactful data for system improvement.
We derive influence score estimation methods tailored for non-differentiable metrics.
arXiv Detail & Related papers (2025-02-02T23:20:16Z) - Testing and Improving the Robustness of Amortized Bayesian Inference for Cognitive Models [0.5223954072121659]
Contaminant observations and outliers often cause problems when estimating the parameters of cognitive models.
In this study, we test and improve the robustness of parameter estimation using amortized Bayesian inference.
The proposed method is straightforward and practical to implement and has a broad applicability in fields where outlier detection or removal is challenging.
arXiv Detail & Related papers (2024-12-29T21:22:24Z) - A Conformal Approach to Feature-based Newsvendor under Model Misspecification [2.801095519296785]
We propose a model-free and distribution-free framework inspired by conformal prediction.
We validate our framework using both simulated data and a real-world dataset from the Capital Bikeshare program in Washington, D.C.
arXiv Detail & Related papers (2024-12-17T18:34:43Z) - Using Instruction-Tuned Large Language Models to Identify Indicators of Vulnerability in Police Incident Narratives [0.0]
We analyze publicly available text narratives of police-public interactions recorded by Boston Police Department.
We compare qualitative coding of instruction tuned large language models (IT-LLMs) against human coders.
arXiv Detail & Related papers (2024-12-16T15:27:37Z) - KLCBL: An Improved Police Incident Classification Model [0.0]
Police incident data is crucial for public security intelligence, yet grassroots agencies struggle with efficient classification due to manual inefficiency and automated system limitations.
This research proposes a multichannel neural network model, KLCBL, integrating Kolmogorov-Arnold Networks (KAN), a linguistically enhanced text preprocessing approach (LERT), Convolutional Neural Network (CNN), and Bidirectional Long Short-Term Memory (BiLSTM) for police incident classification.
The model addresses classification challenges, enhances police informatization, improves resource allocation, and offers broad applicability to other classification tasks.
arXiv Detail & Related papers (2024-11-11T07:02:23Z) - Steering Masked Discrete Diffusion Models via Discrete Denoising Posterior Prediction [88.65168366064061]
We introduce Discrete Denoising Posterior Prediction (DDPP), a novel framework that casts the task of steering pre-trained MDMs as a problem of probabilistic inference.
Our framework leads to a family of three novel objectives that are all simulation-free, and thus scalable.
We substantiate our designs via wet-lab validation, where we observe transient expression of reward-optimized protein sequences.
arXiv Detail & Related papers (2024-10-10T17:18:30Z) - Investigating a Benchmark for Training-set free Evaluation of Linguistic Capabilities in Machine Reading Comprehension [12.09297288867446]
We examine a framework for evaluating optimised models in training-set free setting on synthetically generated challenge sets.
We find that despite the simplicity of the generation method, the data can compete with crowd-sourced datasets with regard to naturalness and lexical diversity.
We conduct further experiments and show that state-of-the-art language model-based MRC systems can learn to succeed on the challenge set correctly.
arXiv Detail & Related papers (2024-08-09T12:23:36Z) - Trustless Audits without Revealing Data or Models [49.23322187919369]
We show that it is possible to allow model providers to keep their model weights (but not architecture) and data secret while allowing other parties to trustlessly audit model and data properties.
We do this by designing a protocol called ZkAudit in which model providers publish cryptographic commitments of datasets and model weights.
arXiv Detail & Related papers (2024-04-06T04:43:06Z) - How Much Data are Enough? Investigating Dataset Requirements for Patch-Based Brain MRI Segmentation Tasks [74.21484375019334]
Training deep neural networks reliably requires access to large-scale datasets.
To mitigate both the time and financial costs associated with model development, a clear understanding of the amount of data required to train a satisfactory model is crucial.
This paper proposes a strategic framework for estimating the amount of annotated data required to train patch-based segmentation networks.
arXiv Detail & Related papers (2024-04-04T13:55:06Z) - Explainable Fraud Detection with Deep Symbolic Classification [4.1205832766381985]
We present Deep Classification, an extension of the Deep Symbolic Regression framework to classification problems.
Because the functions are mathematical expressions that are in closed-form and concise, the model is inherently explainable both at the level of a single classification decision and the model's decision process.
An evaluation on the PaySim data set demonstrates competitive predictive performance with state-of-the-art models, while surpassing them in terms of explainability.
arXiv Detail & Related papers (2023-12-01T13:50:55Z) - QualEval: Qualitative Evaluation for Model Improvement [82.73561470966658]
We propose QualEval, which augments quantitative scalar metrics with automated qualitative evaluation as a vehicle for model improvement.
QualEval uses a powerful LLM reasoner and our novel flexible linear programming solver to generate human-readable insights.
We demonstrate that leveraging its insights, for example, improves the absolute performance of the Llama 2 model by up to 15% points relative.
arXiv Detail & Related papers (2023-11-06T00:21:44Z) - Assessing Privacy Risks in Language Models: A Case Study on
Summarization Tasks [65.21536453075275]
We focus on the summarization task and investigate the membership inference (MI) attack.
We exploit text similarity and the model's resistance to document modifications as potential MI signals.
We discuss several safeguards for training summarization models to protect against MI attacks and discuss the inherent trade-off between privacy and utility.
arXiv Detail & Related papers (2023-10-20T05:44:39Z) - Learning Evaluation Models from Large Language Models for Sequence Generation [61.8421748792555]
We propose a three-stage evaluation model training method that utilizes large language models to generate labeled data for model-based metric development.
Experimental results on the SummEval benchmark demonstrate that CSEM can effectively train an evaluation model without human-labeled data.
arXiv Detail & Related papers (2023-08-08T16:41:16Z) - SimSCOOD: Systematic Analysis of Out-of-Distribution Generalization in
Fine-tuned Source Code Models [58.78043959556283]
We study the behaviors of models under different fine-tuning methodologies, including full fine-tuning and Low-Rank Adaptation (LoRA) fine-tuning methods.
Our analysis uncovers that LoRA fine-tuning consistently exhibits significantly better OOD generalization performance than full fine-tuning across various scenarios.
arXiv Detail & Related papers (2022-10-10T16:07:24Z) - Efficient training of lightweight neural networks using Online
Self-Acquired Knowledge Distillation [51.66271681532262]
Online Self-Acquired Knowledge Distillation (OSAKD) is proposed, aiming to improve the performance of any deep neural model in an online manner.
We utilize k-nn non-parametric density estimation technique for estimating the unknown probability distributions of the data samples in the output feature space.
arXiv Detail & Related papers (2021-08-26T14:01:04Z) - Validation and Inference of Agent Based Models [0.0]
Agent Based Modelling (ABM) is a computational framework for simulating the behaviours and interactions of autonomous agents.
Recent research in ABC has yielded increasingly efficient algorithms for calculating the approximate likelihood.
These are investigated and compared using a pedestrian model in the Hamilton CBD.
arXiv Detail & Related papers (2021-07-08T05:53:37Z) - How Faithful is your Synthetic Data? Sample-level Metrics for Evaluating
and Auditing Generative Models [95.8037674226622]
We introduce a 3-dimensional evaluation metric that characterizes the fidelity, diversity and generalization performance of any generative model in a domain-agnostic fashion.
Our metric unifies statistical divergence measures with precision-recall analysis, enabling sample- and distribution-level diagnoses of model fidelity and diversity.
arXiv Detail & Related papers (2021-02-17T18:25:30Z) - AIRSENSE-TO-ACT: A Concept Paper for COVID-19 Countermeasures based on
Artificial Intelligence algorithms and multi-sources Data Processing [0.0]
This paper describes a new tool to support institutions in the implementation of targeted countermeasures, based on quantitative and multi-scale elements, for the fight and prevention of emergencies, such as the current COVID-19 pandemic.
The tool is a centralized system (web application), single multi-user platform, which relies on Artificial Intelligence (AI) algorithms for the processing of heterogeneous data, and which can produce an output level of risk.
The model includes a specific neural network which will be first trained to learn the correlation between selected inputs, related to the case of interest: environmental variables (chemical-physical, such as meteorological), human activity
arXiv Detail & Related papers (2020-11-07T17:50:14Z) - PermuteAttack: Counterfactual Explanation of Machine Learning Credit
Scorecards [0.0]
This paper is a note on new directions and methodologies for validation and explanation of Machine Learning (ML) models employed for retail credit scoring in finance.
Our proposed framework draws motivation from the field of Artificial Intelligence (AI) security and adversarial ML.
arXiv Detail & Related papers (2020-08-24T00:05:13Z) - Structural Causal Models Are (Solvable by) Credal Networks [70.45873402967297]
Causal inferences can be obtained by standard algorithms for the updating of credal nets.
This contribution should be regarded as a systematic approach to represent structural causal models by credal networks.
Experiments show that approximate algorithms for credal networks can immediately be used to do causal inference in real-size problems.
arXiv Detail & Related papers (2020-08-02T11:19:36Z) - Model Repair: Robust Recovery of Over-Parameterized Statistical Models [24.319310729283636]
A new type of robust estimation problem is introduced where the goal is to recover a statistical model that has been corrupted after it has been estimated from data.
Methods are proposed for "repairing" the model using only the design and not the response values used to fit the model in a supervised learning setting.
arXiv Detail & Related papers (2020-05-20T08:41:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.