Adaptive ABAC Policy Learning: A Reinforcement Learning Approach
- URL: http://arxiv.org/abs/2105.08587v1
- Date: Tue, 18 May 2021 15:18:02 GMT
- Title: Adaptive ABAC Policy Learning: A Reinforcement Learning Approach
- Authors: Leila Karimi, Mai Abdelhakim, James Joshi
- Abstract summary: We propose an adaptive ABAC policy learning approach to automate the authorization management task.
In particular, we propose a contextual bandit system, in which an authorization engine adapts an ABAC model through a feedback control loop.
We focus on developing an adaptive ABAC policy learning model for a home IoT environment as a running example.
- Score: 2.5997274006052544
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With rapid advances in computing systems, there is an increasing demand for
more effective and efficient access control (AC) approaches. Recently,
Attribute Based Access Control (ABAC) approaches have been shown to be
promising in fulfilling the AC needs of such emerging complex computing
environments. An ABAC model grants access to a requester based on attributes of
entities in a system and an authorization policy; however, its generality and
flexibility come with a higher cost. Further, increasing complexities of
organizational systems and the need for federated accesses to their resources
make the task of AC enforcement and management much more challenging. In this
paper, we propose an adaptive ABAC policy learning approach to automate the
authorization management task. We model ABAC policy learning as a reinforcement
learning problem. In particular, we propose a contextual bandit system, in
which an authorization engine adapts an ABAC model through a feedback control
loop; it relies on interacting with users/administrators of the system to
receive their feedback that assists the model in making authorization
decisions. We propose four methods for initializing the learning model and a
planning approach based on attribute value hierarchy to accelerate the learning
process. We focus on developing an adaptive ABAC policy learning model for a
home IoT environment as a running example. We evaluate our proposed approach
over real and synthetic data. We consider both complete and sparse datasets in
our evaluations. Our experimental results show that the proposed approach
achieves performance that is comparable to ones based on supervised learning in
many scenarios and even outperforms them in several situations.
Related papers
- Fast Context Adaptation in Cost-Aware Continual Learning [10.515324071327903]
5G and Beyond networks require more complex learning agents and the learning process itself might end up competing with users for communication and computational resources.
This creates friction: on the one hand, the learning process needs resources to quickly convergence to an effective strategy; on the other hand, the learning process needs to be efficient, i.e. take as few resources as possible from the user's data plane, so as not to throttle users' resources.
In this paper, we propose a dynamic strategy to balance the resources assigned to the data plane and those reserved for learning.
arXiv Detail & Related papers (2023-06-06T17:46:48Z) - Sparsity-Aware Intelligent Massive Random Access Control in Open RAN: A
Reinforcement Learning Based Approach [61.74489383629319]
Massive random access of devices in the emerging Open Radio Access Network (O-RAN) brings great challenge to the access control and management.
reinforcement-learning (RL)-assisted scheme of closed-loop access control is proposed to preserve sparsity of access requests.
Deep-RL-assisted SAUD is proposed to resolve highly complex environments with continuous and high-dimensional state and action spaces.
arXiv Detail & Related papers (2023-03-05T12:25:49Z) - Fully Decentralized Model-based Policy Optimization for Networked
Systems [23.46407780093797]
This work aims to improve data efficiency of multi-agent control by model-based learning.
We consider networked systems where agents are cooperative and communicate only locally with their neighbors.
In our method, each agent learns a dynamic model to predict future states and broadcast their predictions by communication, and then the policies are trained under the model rollouts.
arXiv Detail & Related papers (2022-07-13T23:52:14Z) - Toward Deep Learning Based Access Control [3.2511618464944547]
This paper proposes Deep Learning Based Access Control (DLBAC) by leveraging significant advances in deep learning technology.
DLBAC could complement and, in the long-term, has the potential to even replace, classical access control models with a neural network.
We demonstrate the feasibility of the proposed approach by addressing issues related to accuracy, generalization, and explainability.
arXiv Detail & Related papers (2022-03-28T22:05:11Z) - Unified Instance and Knowledge Alignment Pretraining for Aspect-based
Sentiment Analysis [96.53859361560505]
Aspect-based Sentiment Analysis (ABSA) aims to determine the sentiment polarity towards an aspect.
There always exists severe domain shift between the pretraining and downstream ABSA datasets.
We introduce a unified alignment pretraining framework into the vanilla pretrain-finetune pipeline.
arXiv Detail & Related papers (2021-10-26T04:03:45Z) - Evaluating model-based planning and planner amortization for continuous
control [79.49319308600228]
We take a hybrid approach, combining model predictive control (MPC) with a learned model and model-free policy learning.
We find that well-tuned model-free agents are strong baselines even for high DoF control problems.
We show that it is possible to distil a model-based planner into a policy that amortizes the planning without any loss of performance.
arXiv Detail & Related papers (2021-10-07T12:00:40Z) - DEALIO: Data-Efficient Adversarial Learning for Imitation from
Observation [57.358212277226315]
In imitation learning from observation IfO, a learning agent seeks to imitate a demonstrating agent using only observations of the demonstrated behavior without access to the control signals generated by the demonstrator.
Recent methods based on adversarial imitation learning have led to state-of-the-art performance on IfO problems, but they typically suffer from high sample complexity due to a reliance on data-inefficient, model-free reinforcement learning algorithms.
This issue makes them impractical to deploy in real-world settings, where gathering samples can incur high costs in terms of time, energy, and risk.
We propose a more data-efficient IfO algorithm
arXiv Detail & Related papers (2021-03-31T23:46:32Z) - Deep RL With Information Constrained Policies: Generalization in
Continuous Control [21.46148507577606]
We show that a natural constraint on information flow might confer onto artificial agents in continuous control tasks.
We implement a novel Capacity-Limited Actor-Critic (CLAC) algorithm.
Our experiments show that compared to alternative approaches, CLAC offers improvements in generalization between training and modified test environments.
arXiv Detail & Related papers (2020-10-09T15:42:21Z) - Data-driven Koopman Operators for Model-based Shared Control of
Human-Machine Systems [66.65503164312705]
We present a data-driven shared control algorithm that can be used to improve a human operator's control of complex machines.
Both the dynamics and information about the user's interaction are learned from observation through the use of a Koopman operator.
We find that model-based shared control significantly improves task and control metrics when compared to a natural learning, or user only, control paradigm.
arXiv Detail & Related papers (2020-06-12T14:14:07Z) - An Automatic Attribute Based Access Control Policy Extraction from
Access Logs [5.142415132534397]
An attribute-based access control (ABAC) model provides a more flexible approach for addressing the authorization needs of complex and dynamic systems.
We present a methodology for automatically learning ABAC policy rules from access logs of a system to simplify the policy development process.
arXiv Detail & Related papers (2020-03-16T15:08:54Z) - Information Theoretic Model Predictive Q-Learning [64.74041985237105]
We present a novel theoretical connection between information theoretic MPC and entropy regularized RL.
We develop a Q-learning algorithm that can leverage biased models.
arXiv Detail & Related papers (2019-12-31T00:29:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.