Integrating problem structuring methods with formal design theory: collective water management policy design in Tunisia
- URL: http://arxiv.org/abs/2410.05303v1
- Date: Fri, 4 Oct 2024 13:55:43 GMT
- Title: Integrating problem structuring methods with formal design theory: collective water management policy design in Tunisia
- Authors: Berkay Tosunlu, Joseph H. A. Guillaume, Alexis Tsoukiàs, Emeline Hassenforder, Samia Chrii, Houssem Braiki, Irene Pluchinotta,
- Abstract summary: This paper proposes an innovative approach to policy design by merging Problem Structuring Methods (PSMs) and the Policy-Knowledge, Concepts, Proposals (P-KCP) methodology.
Utilizing cognitive maps and value trees, the study aims to generate new collective groundwater management practices.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Groundwater management, especially in regions like Tunisia, is challenging due to diverse stakeholder interests and the dry structure of climate, which is extremely challenging for the sustainability of water resources. This paper proposes an innovative approach to policy design by merging Problem Structuring Methods (PSMs) and the Policy-Knowledge, Concepts, Proposals (P-KCP) methodology. Utilizing cognitive maps and value trees, the study aims to generate new collective groundwater management practices. Bridging decision theory and design theory, the study addresses the gap in new alternative generation and highlights the P-KCP's role in innovative policy design. Integrating PSMs and C-K theory, the framework expands policy alternatives and advocates for participatory approaches. It emphasizes adaptability across contexts, provides replicable process descriptions, and encourages the creation of unconventional policy solutions. Ultimately, this comprehensive framework offers a practical guide for policy innovation and collaboration.
Related papers
- Theoretically Guaranteed Policy Improvement Distilled from Model-Based
Planning [64.10794426777493]
Model-based reinforcement learning (RL) has demonstrated remarkable successes on a range of continuous control tasks.
Recent practices tend to distill optimized action sequences into an RL policy during the training phase.
We develop an approach to distill from model-based planning to the policy.
arXiv Detail & Related papers (2023-07-24T16:52:31Z) - Representation-Driven Reinforcement Learning [57.44609759155611]
We present a representation-driven framework for reinforcement learning.
By representing policies as estimates of their expected values, we leverage techniques from contextual bandits to guide exploration and exploitation.
We demonstrate the effectiveness of this framework through its application to evolutionary and policy gradient-based approaches.
arXiv Detail & Related papers (2023-05-31T14:59:12Z) - A Strategy-Oriented Bayesian Soft Actor-Critic Model [1.52292571922932]
This paper proposes a novel hierarchical strategy decomposition approach based on the Bayesian chain rule.
We integrate this approach into the state-of-the-art DRL method -- soft actor-critic (SAC) and build the corresponding Bayesian soft actor-critic (BSAC) model.
We compare the proposed BSAC method with the SAC and other state-of-the-art approaches such as TD3, DDPG, and PPO on the standard continuous control benchmarks.
arXiv Detail & Related papers (2023-03-07T19:31:25Z) - Diverse Policy Optimization for Structured Action Space [59.361076277997704]
We propose Diverse Policy Optimization (DPO) to model the policies in structured action space as the energy-based models (EBM)
A novel and powerful generative model, GFlowNet, is introduced as the efficient, diverse EBM-based policy sampler.
Experiments on ATSC and Battle benchmarks demonstrate that DPO can efficiently discover surprisingly diverse policies.
arXiv Detail & Related papers (2023-02-23T10:48:09Z) - Generalised Policy Improvement with Geometric Policy Composition [18.80807234471197]
We introduce a method for policy improvement that interpolates between the greedy approach of value-based reinforcement learning (RL) and the full planning approach typical of model-based RL.
We show that we can evaluate any non-Markov policy that switches between a set of base Markov policies with fixed probability by a careful composition of the base policy GHMs.
We can then apply generalised policy improvement (GPI) to collections of such non-Markov policies to obtain a new Markov policy that will in general outperform its precursors.
arXiv Detail & Related papers (2022-06-17T12:52:13Z) - A Regularized Implicit Policy for Offline Reinforcement Learning [54.7427227775581]
offline reinforcement learning enables learning from a fixed dataset, without further interactions with the environment.
We propose a framework that supports learning a flexible yet well-regularized fully-implicit policy.
Experiments and ablation study on the D4RL dataset validate our framework and the effectiveness of our algorithmic designs.
arXiv Detail & Related papers (2022-02-19T20:22:04Z) - Beyond modeling: NLP Pipeline for efficient environmental policy
analysis [0.6597195879147557]
Policy analysis is necessary for policymakers to understand the actors and rules involved in forest restoration.
We propose a Knowledge Management Framework based on Natural Language Processing (NLP) techniques.
We describe the design of the NLP pipeline, review the state-of-the-art methods for each of its components, and discuss the challenges that rise when building a framework oriented towards policy analysis.
arXiv Detail & Related papers (2022-01-08T05:33:04Z) - An Entropy Regularization Free Mechanism for Policy-based Reinforcement
Learning [1.4566990078034239]
Policy-based reinforcement learning methods suffer from the policy collapse problem.
We propose an entropy regularization free mechanism that is designed for policy-based methods.
arXiv Detail & Related papers (2021-06-01T18:04:19Z) - Novel Policy Seeking with Constrained Optimization [131.67409598529287]
We propose to rethink the problem of generating novel policies in reinforcement learning tasks.
We first introduce a new metric to evaluate the difference between policies and then design two practical novel policy generation methods.
The two proposed methods, namely the Constrained Task Novel Bisector (CTNB) and the Interior Policy Differentiation (IPD), are derived from the feasible direction method and the interior point method commonly known in the constrained optimization literature.
arXiv Detail & Related papers (2020-05-21T14:39:14Z) - Stable Policy Optimization via Off-Policy Divergence Regularization [50.98542111236381]
Trust Region Policy Optimization (TRPO) and Proximal Policy Optimization (PPO) are among the most successful policy gradient approaches in deep reinforcement learning (RL)
We propose a new algorithm which stabilizes the policy improvement through a proximity term that constrains the discounted state-action visitation distribution induced by consecutive policies to be close to one another.
Our proposed method can have a beneficial effect on stability and improve final performance in benchmark high-dimensional control tasks.
arXiv Detail & Related papers (2020-03-09T13:05:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.