A Quantitative Perspective on Values of Domain Knowledge for Machine
Learning
- URL: http://arxiv.org/abs/2011.08450v2
- Date: Tue, 9 Feb 2021 09:14:56 GMT
- Title: A Quantitative Perspective on Values of Domain Knowledge for Machine
Learning
- Authors: Jianyi Yang, Shaolei Ren
- Abstract summary: Domain knowledge in various forms has been playing a crucial role in improving the learning performance.
We study the problem of quantifying the values of domain knowledge in terms of its contribution to the learning performance.
- Score: 27.84415856657607
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the exploding popularity of machine learning, domain knowledge in
various forms has been playing a crucial role in improving the learning
performance, especially when training data is limited. Nonetheless, there is
little understanding of to what extent domain knowledge can affect a machine
learning task from a quantitative perspective. To increase the transparency and
rigorously explain the role of domain knowledge in machine learning, we study
the problem of quantifying the values of domain knowledge in terms of its
contribution to the learning performance in the context of informed machine
learning. We propose a quantification method based on Shapley value that fairly
attributes the overall learning performance improvement to different domain
knowledge. We also present Monte-Carlo sampling to approximate the fair value
of domain knowledge with a polynomial time complexity. We run experiments of
injecting symbolic domain knowledge into semi-supervised learning tasks on both
MNIST and CIFAR10 datasets, providing quantitative values of different symbolic
knowledge and rigorously explaining how it affects the machine learning
performance in terms of test accuracy.
Related papers
- Large Language Models are Limited in Out-of-Context Knowledge Reasoning [65.72847298578071]
Large Language Models (LLMs) possess extensive knowledge and strong capabilities in performing in-context reasoning.
This paper focuses on a significant aspect of out-of-context reasoning: Out-of-Context Knowledge Reasoning (OCKR), which is to combine multiple knowledge to infer new knowledge.
arXiv Detail & Related papers (2024-06-11T15:58:59Z) - Learning Beyond Pattern Matching? Assaying Mathematical Understanding in LLMs [58.09253149867228]
This paper assesses the domain knowledge of LLMs through its understanding of different mathematical skills required to solve problems.
Motivated by the use of LLMs as a general scientific assistant, we propose textitNTKEval to assess changes in LLM's probability distribution.
Our systematic analysis finds evidence of domain understanding during in-context learning.
Certain instruction-tuning leads to similar performance changes irrespective of training on different data, suggesting a lack of domain understanding across different skills.
arXiv Detail & Related papers (2024-05-24T12:04:54Z) - A Comprehensive Study of Knowledge Editing for Large Language Models [82.65729336401027]
Large Language Models (LLMs) have shown extraordinary capabilities in understanding and generating text that closely mirrors human communication.
This paper defines the knowledge editing problem and provides a comprehensive review of cutting-edge approaches.
We introduce a new benchmark, KnowEdit, for a comprehensive empirical evaluation of representative knowledge editing approaches.
arXiv Detail & Related papers (2024-01-02T16:54:58Z) - Beyond Factuality: A Comprehensive Evaluation of Large Language Models
as Knowledge Generators [78.63553017938911]
Large language models (LLMs) outperform information retrieval techniques for downstream knowledge-intensive tasks.
However, community concerns abound regarding the factuality and potential implications of using this uncensored knowledge.
We introduce CONNER, designed to evaluate generated knowledge from six important perspectives.
arXiv Detail & Related papers (2023-10-11T08:22:37Z) - Worth of knowledge in deep learning [3.132595571344153]
We present a framework inspired by interpretable machine learning to evaluate the worth of knowledge.
Our findings elucidate the complex relationship between data and knowledge, including dependence, synergistic, and substitution effects.
Our model-agnostic framework can be applied to a variety of common network architectures, providing a comprehensive understanding of the role of prior knowledge in deep learning models.
arXiv Detail & Related papers (2023-07-03T02:25:19Z) - Recognizing Unseen Objects via Multimodal Intensive Knowledge Graph
Propagation [68.13453771001522]
We propose a multimodal intensive ZSL framework that matches regions of images with corresponding semantic embeddings.
We conduct extensive experiments and evaluate our model on large-scale real-world data.
arXiv Detail & Related papers (2023-06-14T13:07:48Z) - Graph Enabled Cross-Domain Knowledge Transfer [1.52292571922932]
Cross-Domain Knowledge Transfer is an approach to mitigate the gap between good representation learning and the scarce knowledge in the domain of interest.
From the machine learning perspective, the paradigm of semi-supervised learning takes advantage of large amount of data without ground truth and achieves impressive learning performance improvement.
arXiv Detail & Related papers (2023-04-07T03:02:10Z) - Informed Learning by Wide Neural Networks: Convergence, Generalization
and Sampling Complexity [27.84415856657607]
We study how and why domain knowledge benefits the performance of informed learning.
We propose a generalized informed training objective to better exploit the benefits of knowledge and balance the label and knowledge imperfectness.
arXiv Detail & Related papers (2022-07-02T06:28:25Z) - Knowledge Modelling and Active Learning in Manufacturing [0.6299766708197884]
Ontologies and Knowledge Graphs provide means to model and relate a wide range of concepts, problems, and configurations.
Both can be used to generate new knowledge through deductive inference and identify missing knowledge.
Active learning can be used to identify the most informative data instances for which to obtain users' feedback, reduce friction, and maximize knowledge acquisition.
arXiv Detail & Related papers (2021-07-05T22:07:21Z) - Knowledge as Invariance -- History and Perspectives of
Knowledge-augmented Machine Learning [69.99522650448213]
Research in machine learning is at a turning point.
Research interests are shifting away from increasing the performance of highly parameterized models to exceedingly specific tasks.
This white paper provides an introduction and discussion of this emerging field in machine learning research.
arXiv Detail & Related papers (2020-12-21T15:07:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.