Simulating and classifying behavior in adversarial environments based on
action-state traces: an application to money laundering
- URL: http://arxiv.org/abs/2011.01826v1
- Date: Tue, 3 Nov 2020 16:30:53 GMT
- Title: Simulating and classifying behavior in adversarial environments based on
action-state traces: an application to money laundering
- Authors: Daniel Borrajo, Manuela Veloso, Sameena Shah
- Abstract summary: We present a novel way of approaching these types of applications, in particular in the context of Anti-Money Laundering.
We provide a mechanism through which diverse, realistic and new unobserved behavior may be generated to discover potential unobserved adversarial actions.
- Score: 18.625578105241
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many business applications involve adversarial relationships in which both
sides adapt their strategies to optimize their opposing benefits. One of the
key characteristics of these applications is the wide range of strategies that
an adversary may choose as they adapt their strategy dynamically to sustain
benefits and evade authorities. In this paper, we present a novel way of
approaching these types of applications, in particular in the context of
Anti-Money Laundering. We provide a mechanism through which diverse, realistic
and new unobserved behavior may be generated to discover potential unobserved
adversarial actions to enable organizations to preemptively mitigate these
risks. In this regard, we make three main contributions. (a) Propose a novel
behavior-based model as opposed to individual transactions-based models
currently used by financial institutions. We introduce behavior traces as
enriched relational representation to represent observed human behavior. (b) A
modelling approach that observes these traces and is able to accurately infer
the goals of actors by classifying the behavior into money laundering or
standard behavior despite significant unobserved activity. And (c) a synthetic
behavior simulator that can generate new previously unseen traces. The
simulator incorporates a high level of flexibility in the behavioral parameters
so that we can challenge the detection algorithm. Finally, we provide
experimental results that show that the learning module (automated
investigator) that has only partial observability can still successfully infer
the type of behavior, and thus the simulated goals, followed by customers based
on traces - a key aspiration for many applications today.
Related papers
- Learning Utilities from Demonstrations in Markov Decision Processes [18.205765143671858]
We propose a novel model of behavior in Markov Decision Processes (MDPs) that explicitly represents the agent's risk attitude through a utility function.
We then define the Utility Learning problem as the task of inferring the observed agent's risk attitude, encoded via a utility function, from demonstrations in MDPs.
We devise two provably efficient algorithms for UL in a finite-data regime, and we analyze their sample complexity.
arXiv Detail & Related papers (2024-09-25T21:01:15Z) - Learning to Generate All Feasible Actions [4.333208181196761]
We introduce action mapping, a novel approach that divides the learning process into two steps: first learn feasibility and subsequently, the objective.
This paper focuses on the feasibility part by learning to generate all feasible actions through self-supervised querying of the feasibility model.
We demonstrate the agent's proficiency in generating actions across disconnected feasible action sets.
arXiv Detail & Related papers (2023-01-26T23:15:51Z) - Emergent Behaviors in Multi-Agent Target Acquisition [0.0]
We simulate a Multi-Agent System (MAS) using Reinforcement Learning (RL) in a pursuit-evasion game.
We create different adversarial scenarios by replacing RL-trained pursuers' policies with two distinct (non-RL) analytical strategies.
The novelty of our approach entails the creation of an influential feature set that reveals underlying data regularities.
arXiv Detail & Related papers (2022-12-15T15:20:58Z) - Inferring Versatile Behavior from Demonstrations by Matching Geometric
Descriptors [72.62423312645953]
Humans intuitively solve tasks in versatile ways, varying their behavior in terms of trajectory-based planning and for individual steps.
Current Imitation Learning algorithms often only consider unimodal expert demonstrations and act in a state-action-based setting.
Instead, we combine a mixture of movement primitives with a distribution matching objective to learn versatile behaviors that match the expert's behavior and versatility.
arXiv Detail & Related papers (2022-10-17T16:42:59Z) - Learning Self-Modulating Attention in Continuous Time Space with
Applications to Sequential Recommendation [102.24108167002252]
We propose a novel attention network, named self-modulating attention, that models the complex and non-linearly evolving dynamic user preferences.
We empirically demonstrate the effectiveness of our method on top-N sequential recommendation tasks, and the results on three large-scale real-world datasets show that our model can achieve state-of-the-art performance.
arXiv Detail & Related papers (2022-03-30T03:54:11Z) - Learning Complex Spatial Behaviours in ABM: An Experimental
Observational Study [0.0]
This paper explores how Reinforcement Learning can be applied to create emergent agent behaviours.
Running a series of simulations, we demonstrate that agents trained using the novel Proximal Policy optimisation algorithm behave in ways that exhibit properties of real-world intelligent adaptive behaviours.
arXiv Detail & Related papers (2022-01-04T11:56:11Z) - Deceptive Decision-Making Under Uncertainty [25.197098169762356]
We study the design of autonomous agents that are capable of deceiving outside observers about their intentions while carrying out tasks.
By modeling the agent's behavior as a Markov decision process, we consider a setting where the agent aims to reach one of multiple potential goals.
We propose a novel approach to model observer predictions based on the principle of maximum entropy and to efficiently generate deceptive strategies.
arXiv Detail & Related papers (2021-09-14T14:56:23Z) - Online reinforcement learning with sparse rewards through an active
inference capsule [62.997667081978825]
This paper introduces an active inference agent which minimizes the novel free energy of the expected future.
Our model is capable of solving sparse-reward problems with a very high sample efficiency.
We also introduce a novel method for approximating the prior model from the reward function, which simplifies the expression of complex objectives.
arXiv Detail & Related papers (2021-06-04T10:03:36Z) - Generative Adversarial Reward Learning for Generalized Behavior Tendency
Inference [71.11416263370823]
We propose a generative inverse reinforcement learning for user behavioral preference modelling.
Our model can automatically learn the rewards from user's actions based on discriminative actor-critic network and Wasserstein GAN.
arXiv Detail & Related papers (2021-05-03T13:14:25Z) - Social NCE: Contrastive Learning of Socially-aware Motion
Representations [87.82126838588279]
Experimental results show that the proposed method dramatically reduces the collision rates of recent trajectory forecasting, behavioral cloning and reinforcement learning algorithms.
Our method makes few assumptions about neural architecture designs, and hence can be used as a generic way to promote the robustness of neural motion models.
arXiv Detail & Related papers (2020-12-21T22:25:06Z) - Behavior Priors for Efficient Reinforcement Learning [97.81587970962232]
We consider how information and architectural constraints can be combined with ideas from the probabilistic modeling literature to learn behavior priors.
We discuss how such latent variable formulations connect to related work on hierarchical reinforcement learning (HRL) and mutual information and curiosity based objectives.
We demonstrate the effectiveness of our framework by applying it to a range of simulated continuous control domains.
arXiv Detail & Related papers (2020-10-27T13:17:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.