Standpoint Linear Temporal Logic
- URL: http://arxiv.org/abs/2304.14243v1
- Date: Thu, 27 Apr 2023 15:03:38 GMT
- Title: Standpoint Linear Temporal Logic
- Authors: Nicola Gigante, Lucia {Gomez Alvarez}, Tim S. Lyon
- Abstract summary: We present standpoint linear temporal logic (SLTL), a new logic that combines the temporal features of thepective with the multi-perspective modelling capacity of SL.
We define the logic SLTL, its syntax, and its semantics, establish its decidability and terminating complexity, and provide a tableau calculus to automate SLTL reasoning.
- Score: 2.552459629685159
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Many complex scenarios require the coordination of agents possessing unique
points of view and distinct semantic commitments. In response, standpoint logic
(SL) was introduced in the context of knowledge integration, allowing one to
reason with diverse and potentially conflicting viewpoints by means of indexed
modalities. Another multi-modal logic of import is linear temporal logic (LTL)
- a formalism used to express temporal properties of systems and processes,
having prominence in formal methods and fields related to artificial
intelligence. In this paper, we present standpoint linear temporal logic
(SLTL), a new logic that combines the temporal features of LTL with the
multi-perspective modelling capacity of SL. We define the logic SLTL, its
syntax, and its semantics, establish its decidability and complexity, and
provide a terminating tableau calculus to automate SLTL reasoning.
Conveniently, this offers a clear path to extend existing LTL reasoners with
practical reasoning support for temporal reasoning in multi-perspective
settings.
Related papers
- DeepLTL: Learning to Efficiently Satisfy Complex LTL Specifications [59.01527054553122]
Linear temporal logic (LTL) has recently been adopted as a powerful formalism for specifying complex, temporally extended tasks in reinforcement learning (RL)
Existing approaches suffer from several shortcomings: they are often only applicable to finite-horizon fragments, are restricted to suboptimal solutions, and do not adequately handle safety constraints.
In this work, we propose a novel learning approach to address these concerns.
Our method leverages the structure of B"uchia, which explicitly represent the semantics of automat- specifications, to learn policies conditioned on sequences of truth assignments that lead to satisfying the desired formulae.
arXiv Detail & Related papers (2024-10-06T21:30:38Z) - Temporal Many-valued Conditional Logics: a Preliminary Report [1.228936170812787]
We start from a many-valued logic with typicality, and extend it with the temporal operators of the Linear Time Temporal Logic.
We also consider an instantiation of the formalism for gradual argumentation.
arXiv Detail & Related papers (2024-09-06T16:23:31Z) - Inductive Learning of Logical Theories with LLMs: A Complexity-graded Analysis [9.865771016218549]
This work presents a novel systematic methodology to analyse the capabilities and limitations of Large Language Models (LLMs)
The analysis is complexity-graded w.r.t. rule dependency structure, allowing quantification of specific inference challenges on LLM performance.
arXiv Detail & Related papers (2024-08-15T16:41:00Z) - TLINet: Differentiable Neural Network Temporal Logic Inference [10.36033062385604]
This paper introduces TLINet, a neural-symbolic framework for learning STL formulas.
In contrast to existing approaches, we introduce approximation methods for max operator designed specifically for temporal logic-based gradient techniques.
Our framework not only learns the structure but also the parameters of STL formulas, allowing flexible combinations of operators and various logical structures.
arXiv Detail & Related papers (2024-05-03T16:38:14Z) - Logic Agent: Enhancing Validity with Logic Rule Invocation [24.815341366820753]
Chain-of-Thought prompting has emerged as a pivotal technique for augmenting the inferential capabilities of language models during reasoning tasks.
This paper introduces the Logic Agent (LA), an agent-based framework aimed at enhancing the validity of reasoning processes in Large Language Models.
arXiv Detail & Related papers (2024-04-28T10:02:28Z) - Towards LogiGLUE: A Brief Survey and A Benchmark for Analyzing Logical Reasoning Capabilities of Language Models [56.34029644009297]
Large language models (LLMs) have demonstrated the ability to overcome various limitations of formal Knowledge Representation (KR) systems.
LLMs excel most in abductive reasoning, followed by deductive reasoning, while they are least effective at inductive reasoning.
We study single-task training, multi-task training, and "chain-of-thought" knowledge distillation fine-tuning technique to assess the performance of model.
arXiv Detail & Related papers (2023-10-02T01:00:50Z) - Modeling Hierarchical Reasoning Chains by Linking Discourse Units and
Key Phrases for Reading Comprehension [80.99865844249106]
We propose a holistic graph network (HGN) which deals with context at both discourse level and word level, as the basis for logical reasoning.
Specifically, node-level and type-level relations, which can be interpreted as bridges in the reasoning process, are modeled by a hierarchical interaction mechanism.
arXiv Detail & Related papers (2023-06-21T07:34:27Z) - Unlocking Temporal Question Answering for Large Language Models with Tailor-Made Reasoning Logic [84.59255070520673]
Large language models (LLMs) face a challenge when engaging in temporal reasoning.
We propose TempLogic, a novel framework designed specifically for temporal question-answering tasks.
arXiv Detail & Related papers (2023-05-24T10:57:53Z) - The Transformation Logics [58.35574640378678]
We introduce a new family of temporal logics designed to balance the trade-off between expressivity and complexity.
Key feature is the possibility of defining operators of a new kind that we call transformation operators.
We show them to yield logics capable of creating hierarchies of increasing expressivity and complexity.
arXiv Detail & Related papers (2023-04-19T13:24:04Z) - Multi-Agent Reinforcement Learning with Temporal Logic Specifications [65.79056365594654]
We study the problem of learning to satisfy temporal logic specifications with a group of agents in an unknown environment.
We develop the first multi-agent reinforcement learning technique for temporal logic specifications.
We provide correctness and convergence guarantees for our main algorithm.
arXiv Detail & Related papers (2021-02-01T01:13:03Z) - Temporal Answer Set Programming [3.263632801414296]
We present an overview on Temporal Logic Programming under the perspective of its application for Knowledge Representation and declarative problem solving.
We focus on recent results of the non-monotonic formalism called Temporal Equilibrium Logic (TEL) that is defined for the full syntax.
In a second part, we focus on practical aspects, defining a syntactic fragment called temporal logic programs closer to ASP, and explain how this has been exploited in the construction of the solver TELINGO.
arXiv Detail & Related papers (2020-09-14T16:13:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.