Conflict Generalisation in ASP: Learning Correct and Effective
Non-Ground Constraints
- URL: http://arxiv.org/abs/2008.03100v1
- Date: Fri, 7 Aug 2020 12:02:32 GMT
- Title: Conflict Generalisation in ASP: Learning Correct and Effective
Non-Ground Constraints
- Authors: Richard Taupe, Antonius Weinzierl, Gerhard Friedrich
- Abstract summary: Generalising and re-using knowledge learned while solving one problem instance has been neglected by state-of-the-art answer set solvers.
We suggest a new approach that generalises learned nogoods for re-use to speed-up the solving of future problem instances.
- Score: 3.8673630752805432
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generalising and re-using knowledge learned while solving one problem
instance has been neglected by state-of-the-art answer set solvers. We suggest
a new approach that generalises learned nogoods for re-use to speed-up the
solving of future problem instances. Our solution combines well-known ASP
solving techniques with deductive logic-based machine learning. Solving
performance can be improved by adding learned non-ground constraints to the
original program. We demonstrate the effects of our method by means of
realistic examples, showing that our approach requires low computational cost
to learn constraints that yield significant performance benefits in our test
cases. These benefits can be seen with ground-and-solve systems as well as
lazy-grounding systems. However, ground-and-solve systems suffer from
additional grounding overheads, induced by the additional constraints in some
cases. By means of conflict minimization, non-minimal learned constraints can
be reduced. This can result in significant reductions of grounding and solving
efforts, as our experiments show. (Under consideration for acceptance in TPLP.)
Related papers
- Efficient Imitation Learning with Conservative World Models [54.52140201148341]
We tackle the problem of policy learning from expert demonstrations without a reward function.
We re-frame imitation learning as a fine-tuning problem, rather than a pure reinforcement learning one.
arXiv Detail & Related papers (2024-05-21T20:53:18Z) - Near-Optimal Solutions of Constrained Learning Problems [85.48853063302764]
In machine learning systems, the need to curtail their behavior has become increasingly apparent.
This is evidenced by recent advancements towards developing models that satisfy dual robustness variables.
Our results show that rich parametrizations effectively mitigate non-dimensional, finite learning problems.
arXiv Detail & Related papers (2024-03-18T14:55:45Z) - Resilient Constrained Learning [94.27081585149836]
This paper presents a constrained learning approach that adapts the requirements while simultaneously solving the learning task.
We call this approach resilient constrained learning after the term used to describe ecological systems that adapt to disruptions by modifying their operation.
arXiv Detail & Related papers (2023-06-04T18:14:18Z) - Efficient lifting of symmetry breaking constraints for complex
combinatorial problems [9.156939957189502]
This work extends the learning framework and implementation of a model-based approach for Answer Set Programming.
In particular, we incorporate a new conflict analysis algorithm in the Inductive Logic Programming system ILASP.
arXiv Detail & Related papers (2022-05-14T20:42:13Z) - Exploring Viable Algorithmic Options for Learning from Demonstration
(LfD): A Parameterized Complexity Approach [0.0]
In this paper, we show how such a systematic exploration of algorithmic options can be done using parameterized complexity analysis.
We show that none of our problems can be solved efficiently either in general or relative to a number of (often simultaneous) restrictions on environments, demonstrations, and policies.
arXiv Detail & Related papers (2022-05-10T15:54:06Z) - Constrained Learning with Non-Convex Losses [119.8736858597118]
Though learning has become a core technology of modern information processing, there is now ample evidence that it can lead to biased, unsafe, and prejudiced solutions.
arXiv Detail & Related papers (2021-03-08T23:10:33Z) - Loss Bounds for Approximate Influence-Based Abstraction [81.13024471616417]
Influence-based abstraction aims to gain leverage by modeling local subproblems together with the 'influence' that the rest of the system exerts on them.
This paper investigates the performance of such approaches from a theoretical perspective.
We show that neural networks trained with cross entropy are well suited to learn approximate influence representations.
arXiv Detail & Related papers (2020-11-03T15:33:10Z) - Probably Approximately Correct Constrained Learning [135.48447120228658]
We develop a generalization theory based on the probably approximately correct (PAC) learning framework.
We show that imposing a learner does not make a learning problem harder in the sense that any PAC learnable class is also a constrained learner.
We analyze the properties of this solution and use it to illustrate how constrained learning can address problems in fair and robust classification.
arXiv Detail & Related papers (2020-06-09T19:59:29Z) - Teaching the Old Dog New Tricks: Supervised Learning with Constraints [18.88930622054883]
Adding constraint support in Machine Learning has the potential to address outstanding issues in data-driven AI systems.
Existing approaches typically apply constrained optimization techniques to ML training, enforce constraint satisfaction by adjusting the model design, or use constraints to correct the output.
Here, we investigate a different, complementary, strategy based on "teaching" constraint satisfaction to a supervised ML method via the direct use of a state-of-the-art constraint solver.
arXiv Detail & Related papers (2020-02-25T09:47:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.