Inductive Learning for Possibilistic Logic Programs Under Stable Models
- URL: http://arxiv.org/abs/2510.07069v1
- Date: Wed, 08 Oct 2025 14:32:10 GMT
- Title: Inductive Learning for Possibilistic Logic Programs Under Stable Models
- Authors: Hongbo Hu, Yisong Wang, Yi Huang, Kewen Wang,
- Abstract summary: Possibilistic logic programs (poss-programs) under stable models are a major variant of answer set programming (ASP)<n>This paper presents an approach to extracting poss-programs from a background program and examples (parts of intended possibilistic stable models)
- Score: 9.302291774874066
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Possibilistic logic programs (poss-programs) under stable models are a major variant of answer set programming (ASP). While its semantics (possibilistic stable models) and properties have been well investigated, the problem of inductive reasoning has not been investigated yet. This paper presents an approach to extracting poss-programs from a background program and examples (parts of intended possibilistic stable models). To this end, the notion of induction tasks is first formally defined, its properties are investigated and two algorithms ilpsm and ilpsmmin for computing induction solutions are presented. An implementation of ilpsmmin is also provided and experimental results show that when inputs are ordinary logic programs, the prototype outperforms a major inductive learning system for normal logic programs from stable models on the datasets that are randomly generated.
Related papers
- Schoenfeld's Anatomy of Mathematical Reasoning by Language Models [56.656180566692946]
We adopt Schoenfeld's Episode Theory as an inductive, intermediate-scale lens and introduce ThinkARM (Anatomy of Reasoning in Models)<n>ThinkARM explicitly abstracts reasoning traces into functional reasoning steps such as Analysis, Explore, Implement, verify, etc.<n>We show that episode-level representations make reasoning steps explicit, enabling systematic analysis of how reasoning is structured, stabilized, and altered in modern language models.
arXiv Detail & Related papers (2025-12-23T02:44:25Z) - REFINESTAT: Efficient Exploration for Probabilistic Program Synthesis [5.509795962249259]
RefineStat enforces semantic constraints ensuring synthesized programs contain valid and well-formed parameters.<n>It applies diagnostic-aware refinement by resampling prior or likelihood components whenever reliability checks fail.<n>It produces programs that are both syntactically sound and statistically reliable, often matching or surpassing those from closed-source large language models.
arXiv Detail & Related papers (2025-09-01T03:13:36Z) - Internal Causal Mechanisms Robustly Predict Language Model Out-of-Distribution Behaviors [61.92704516732144]
We show that the most robust features for correctness prediction are those that play a distinctive causal role in the model's behavior.<n>We propose two methods that leverage causal mechanisms to predict the correctness of model outputs.
arXiv Detail & Related papers (2025-05-17T00:31:39Z) - LLM-Guided Probabilistic Program Induction for POMDP Model Estimation [40.98644220584212]
Partially Observable Markov Decision Processes (POMDPs) model decision making under uncertainty.<n>We are interested in a subclass of POMDPs wherein the components of the model, including the observation function, reward function, transition function, and initial state distribution function, can be modeled as low-complexity probabilistic graphical models.
arXiv Detail & Related papers (2025-05-04T18:59:07Z) - Self-Steering Language Models [113.96916935955842]
DisCIPL is a method for "self-steering" language models (LMs)<n>DisCIPL generates a task-specific inference program that is executed by a population of Follower models.<n>Our work opens up a design space of highly-parallelized Monte Carlo inference strategies.
arXiv Detail & Related papers (2025-04-09T17:54:22Z) - Causality can systematically address the monsters under the bench(marks) [64.36592889550431]
Benchmarks are plagued by various biases, artifacts, or leakage.<n>Models may behave unreliably due to poorly explored failure modes.<n> causality offers an ideal framework to systematically address these challenges.
arXiv Detail & Related papers (2025-02-07T17:01:37Z) - JustLogic: A Comprehensive Benchmark for Evaluating Deductive Reasoning in Large Language Models [51.99046112135311]
We introduce JustLogic, a synthetically generated deductive reasoning benchmark for rigorous evaluation of Large Language Models (LLMs)<n>JustLogic is highly complex, capable of generating a diverse range of linguistic patterns, vocabulary, and argument structures.<n>Our experimental results reveal that (i) state-of-the-art (SOTA) reasoning LLMs perform on par or better than the human average but significantly worse than the human ceiling.
arXiv Detail & Related papers (2025-01-24T15:49:10Z) - Symbolic Parameter Learning in Probabilistic Answer Set Programming [0.16385815610837165]
We propose two algorithms to solve the formalism of Proabilistic Set Programming.
The first solves the task using an off-the-shelf constrained optimization solver.
The second is based on an implementation of the Expectation Maximization algorithm.
arXiv Detail & Related papers (2024-08-16T13:32:47Z) - Language Models can be Logical Solvers [99.40649402395725]
We introduce LoGiPT, a novel language model that directly emulates the reasoning processes of logical solvers.
LoGiPT is fine-tuned on a newly constructed instruction-tuning dataset derived from revealing and refining the invisible reasoning process of deductive solvers.
arXiv Detail & Related papers (2023-11-10T16:23:50Z) - Linear Temporal Logic Modulo Theories over Finite Traces (Extended
Version) [72.38188258853155]
This paper studies Linear Temporal Logic over Finite Traces (LTLf)
proposition letters are replaced with first-order formulas interpreted over arbitrary theories.
The resulting logic, called Satisfiability Modulo Theories (LTLfMT), is semi-decidable.
arXiv Detail & Related papers (2022-04-28T17:57:33Z) - Enforcing Consistency in Weakly Supervised Semantic Parsing [68.2211621631765]
We explore the use of consistency between the output programs for related inputs to reduce the impact of spurious programs.
We find that a more consistent formalism leads to improved model performance even without consistency-based training.
arXiv Detail & Related papers (2021-07-13T03:48:04Z) - Generating Random Logic Programs Using Constraint Programming [12.47276164048813]
We present a novel approach to generating random logic programs using constraint programming.
We show how the model scales with parameter values, and use the model to compare probabilistic inference algorithms across a range of synthetic problems.
arXiv Detail & Related papers (2020-06-02T19:12:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.