Specialization in Hierarchical Learning Systems
- URL: http://arxiv.org/abs/2011.01845v1
- Date: Tue, 3 Nov 2020 17:00:31 GMT
- Title: Specialization in Hierarchical Learning Systems
- Authors: Heinke Hihn and Daniel A. Braun
- Abstract summary: We investigate in how far information constraints in hierarchies of experts not only provide a principled method for regularization but also to enforce specialization.
We devise an information-theoretically motivated on-line learning rule that allows partitioning of the problem space into multiple sub-problems that can be solved by the individual experts.
We show the broad applicability of our approach on a range of problems including classification, regression, density estimation, and reinforcement learning problems.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Joining multiple decision-makers together is a powerful way to obtain more
sophisticated decision-making systems, but requires to address the questions of
division of labor and specialization. We investigate in how far information
constraints in hierarchies of experts not only provide a principled method for
regularization but also to enforce specialization. In particular, we devise an
information-theoretically motivated on-line learning rule that allows
partitioning of the problem space into multiple sub-problems that can be solved
by the individual experts. We demonstrate two different ways to apply our
method: (i) partitioning problems based on individual data samples and (ii)
based on sets of data samples representing tasks. Approach (i) equips the
system with the ability to solve complex decision-making problems by finding an
optimal combination of local expert decision-makers. Approach (ii) leads to
decision-makers specialized in solving families of tasks, which equips the
system with the ability to solve meta-learning problems. We show the broad
applicability of our approach on a range of problems including classification,
regression, density estimation, and reinforcement learning problems, both in
the standard machine learning setup and in a meta-learning setting.
Related papers
- A Unifying Post-Processing Framework for Multi-Objective Learn-to-Defer Problems [6.046591474843391]
Learn-to-Defer is a paradigm that enables learning algorithms to work not in isolation but as a team with human experts.
In this paper, we obtain the Bayes optimal solution for learn-to-defer systems under various constraints.
Our algorithm shows improvements in terms of constraint violation over a set of baselines.
arXiv Detail & Related papers (2024-07-17T16:32:30Z) - Eliciting Problem Specifications via Large Language Models [4.055489363682198]
Large language models (LLMs) can be utilized to map a problem class into a semi-formal specification.
A cognitive system can then use the problem-space specification to solve multiple instances of problems from the problem class.
arXiv Detail & Related papers (2024-05-20T16:19:02Z) - Explainable Data-Driven Optimization: From Context to Decision and Back
Again [76.84947521482631]
Data-driven optimization uses contextual information and machine learning algorithms to find solutions to decision problems with uncertain parameters.
We introduce a counterfactual explanation methodology tailored to explain solutions to data-driven problems.
We demonstrate our approach by explaining key problems in operations management such as inventory management and routing.
arXiv Detail & Related papers (2023-01-24T15:25:16Z) - On solving decision and risk management problems subject to uncertainty [91.3755431537592]
Uncertainty is a pervasive challenge in decision and risk management.
This paper develops a systematic understanding of such strategies, determine their range of application, and develop a framework to better employ them.
arXiv Detail & Related papers (2023-01-18T19:16:23Z) - Causal Fairness Analysis [68.12191782657437]
We introduce a framework for understanding, modeling, and possibly solving issues of fairness in decision-making settings.
The main insight of our approach will be to link the quantification of the disparities present on the observed data with the underlying, and often unobserved, collection of causal mechanisms.
Our effort culminates in the Fairness Map, which is the first systematic attempt to organize and explain the relationship between different criteria found in the literature.
arXiv Detail & Related papers (2022-07-23T01:06:34Z) - A Practical Guide to Multi-Objective Reinforcement Learning and Planning [24.81310809455139]
This paper serves as a guide to the application of multi-objective methods to difficult problems.
It identifies the factors that may influence the nature of the desired solution.
It illustrates by example how these influence the design of multi-objective decision-making systems.
arXiv Detail & Related papers (2021-03-17T11:07:28Z) - CoreDiag: Eliminating Redundancy in Constraint Sets [68.8204255655161]
We present a new algorithm which can be exploited for the determination of minimal cores (minimal non-redundant constraint sets)
The algorithm is especially useful for distributed knowledge engineering scenarios where the degree of redundancy can become high.
In order to show the applicability of our approach, we present an empirical study conducted with commercial configuration knowledge bases.
arXiv Detail & Related papers (2021-02-24T09:16:10Z) - Decision Rule Elicitation for Domain Adaptation [93.02675868486932]
Human-in-the-loop machine learning is widely used in artificial intelligence (AI) to elicit labels from experts.
In this work, we allow experts to additionally produce decision rules describing their decision-making.
We show that decision rule elicitation improves domain adaptation of the algorithm and helps to propagate expert's knowledge to the AI model.
arXiv Detail & Related papers (2021-02-23T08:07:22Z) - Leveraging Expert Consistency to Improve Algorithmic Decision Support [62.61153549123407]
We explore the use of historical expert decisions as a rich source of information that can be combined with observed outcomes to narrow the construct gap.
We propose an influence function-based methodology to estimate expert consistency indirectly when each case in the data is assessed by a single expert.
Our empirical evaluation, using simulations in a clinical setting and real-world data from the child welfare domain, indicates that the proposed approach successfully narrows the construct gap.
arXiv Detail & Related papers (2021-01-24T05:40:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.