Managing caching strategies for stream reasoning with reinforcement
learning
- URL: http://arxiv.org/abs/2008.03212v1
- Date: Fri, 7 Aug 2020 15:01:41 GMT
- Title: Managing caching strategies for stream reasoning with reinforcement
learning
- Authors: Carmine Dodaro, Thomas Eiter, Paul Ogris, Konstantin Schekotihin
- Abstract summary: Stream reasoning allows efficient decision-making over continuously changing data.
We suggest a novel approach that uses the Conflict-Driven Constraint Learning (CDCL) to efficiently update legacy solutions.
In particular, we study the applicability of reinforcement learning to continuously assess the utility of learned constraints.
- Score: 18.998260813058305
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Efficient decision-making over continuously changing data is essential for
many application domains such as cyber-physical systems, industry
digitalization, etc. Modern stream reasoning frameworks allow one to model and
solve various real-world problems using incremental and continuous evaluation
of programs as new data arrives in the stream. Applied techniques use, e.g.,
Datalog-like materialization or truth maintenance algorithms to avoid costly
re-computations, thus ensuring low latency and high throughput of a stream
reasoner. However, the expressiveness of existing approaches is quite limited
and, e.g., they cannot be used to encode problems with constraints, which often
appear in practice. In this paper, we suggest a novel approach that uses the
Conflict-Driven Constraint Learning (CDCL) to efficiently update legacy
solutions by using intelligent management of learned constraints. In
particular, we study the applicability of reinforcement learning to
continuously assess the utility of learned constraints computed in previous
invocations of the solving algorithm for the current one. Evaluations conducted
on real-world reconfiguration problems show that providing a CDCL algorithm
with relevant learned constraints from previous iterations results in
significant performance improvements of the algorithm in stream reasoning
scenarios.
Under consideration for acceptance in TPLP.
Related papers
- Causal Context Adjustment Loss for Learned Image Compression [72.7300229848778]
In recent years, learned image compression (LIC) technologies have surpassed conventional methods notably in terms of rate-distortion (RD) performance.
Most present techniques are VAE-based with an autoregressive entropy model, which obviously promotes the RD performance by utilizing the decoded causal context.
In this paper, we make the first attempt in investigating the way to explicitly adjust the causal context with our proposed Causal Context Adjustment loss.
arXiv Detail & Related papers (2024-10-07T09:08:32Z) - Decentralized Learning Strategies for Estimation Error Minimization with Graph Neural Networks [94.2860766709971]
We address the challenge of sampling and remote estimation for autoregressive Markovian processes in a wireless network with statistically-identical agents.
Our goal is to minimize time-average estimation error and/or age of information with decentralized scalable sampling and transmission policies.
arXiv Detail & Related papers (2024-04-04T06:24:11Z) - Deep Neural Network for Constraint Acquisition through Tailored Loss
Function [0.0]
The significance of learning constraints from data is underscored by its potential applications in real-world problem-solving.
This work introduces a novel approach grounded in Deep Neural Network (DNN) based on Symbolic Regression.
arXiv Detail & Related papers (2024-03-04T13:47:33Z) - Large-Scale OD Matrix Estimation with A Deep Learning Method [70.78575952309023]
The proposed method integrates deep learning and numerical optimization algorithms to infer matrix structure and guide numerical optimization.
We conducted tests to demonstrate the good generalization performance of our method on a large-scale synthetic dataset.
arXiv Detail & Related papers (2023-10-09T14:30:06Z) - Interpretable Anomaly Detection via Discrete Optimization [1.7150329136228712]
We propose a framework for learning inherently interpretable anomaly detectors from sequential data.
We show that this problem is computationally hard and develop two learning algorithms based on constraint optimization.
Using a prototype implementation, we demonstrate that our approach shows promising results in terms of accuracy and F1 score.
arXiv Detail & Related papers (2023-03-24T16:19:15Z) - Learning to Optimize with Stochastic Dominance Constraints [103.26714928625582]
In this paper, we develop a simple yet efficient approach for the problem of comparing uncertain quantities.
We recast inner optimization in the Lagrangian as a learning problem for surrogate approximation, which bypasses apparent intractability.
The proposed light-SD demonstrates superior performance on several representative problems ranging from finance to supply chain management.
arXiv Detail & Related papers (2022-11-14T21:54:31Z) - Instance-Dependent Confidence and Early Stopping for Reinforcement
Learning [99.57168572237421]
Various algorithms for reinforcement learning (RL) exhibit dramatic variation in their convergence rates as a function of problem structure.
This research provides guarantees that explain textitex post the performance differences observed.
A natural next step is to convert these theoretical guarantees into guidelines that are useful in practice.
arXiv Detail & Related papers (2022-01-21T04:25:35Z) - Learning Algorithms for Regenerative Stopping Problems with Applications
to Shipping Consolidation in Logistics [8.111251824291244]
We study regenerative stopping problems in which the system starts anew whenever the controller decides to stop and the long-term average cost is to be minimized.
Traditional model-based solutions involve estimating the underlying process from data and computing strategies for the estimated model.
We compare such solutions to deep reinforcement learning and imitation learning which involve learning a neural network policy from simulations.
arXiv Detail & Related papers (2021-05-05T20:45:46Z) - Escaping from Zero Gradient: Revisiting Action-Constrained Reinforcement
Learning via Frank-Wolfe Policy Optimization [5.072893872296332]
Action-constrained reinforcement learning (RL) is a widely-used approach in various real-world applications.
We propose a learning algorithm that decouples the action constraints from the policy parameter update.
We show that the proposed algorithm significantly outperforms the benchmark methods on a variety of control tasks.
arXiv Detail & Related papers (2021-02-22T14:28:03Z) - Deep RL With Information Constrained Policies: Generalization in
Continuous Control [21.46148507577606]
We show that a natural constraint on information flow might confer onto artificial agents in continuous control tasks.
We implement a novel Capacity-Limited Actor-Critic (CLAC) algorithm.
Our experiments show that compared to alternative approaches, CLAC offers improvements in generalization between training and modified test environments.
arXiv Detail & Related papers (2020-10-09T15:42:21Z) - Combining Deep Learning and Optimization for Security-Constrained
Optimal Power Flow [94.24763814458686]
Security-constrained optimal power flow (SCOPF) is fundamental in power systems.
Modeling of APR within the SCOPF problem results in complex large-scale mixed-integer programs.
This paper proposes a novel approach that combines deep learning and robust optimization techniques.
arXiv Detail & Related papers (2020-07-14T12:38:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.