CogGPT: Unleashing the Power of Cognitive Dynamics on Large Language
Models
- URL: http://arxiv.org/abs/2401.08438v1
- Date: Sat, 6 Jan 2024 03:59:59 GMT
- Title: CogGPT: Unleashing the Power of Cognitive Dynamics on Large Language
Models
- Authors: Yaojia Lv, Haojie Pan, Ruiji Fu, Ming Liu, Zhongyuan Wang, Bing Qin
- Abstract summary: We propose the concept of the cognitive dynamics of large language models (LLMs) and present a corresponding task with the inspiration of longitudinal studies.
Towards the task, we develop CogBench, a novel benchmark to assess the cognitive dynamics of LLMs and validate it through participant surveys.
We introduce CogGPT for the task, which features an innovative iterative cognitive mechanism aimed at enhancing lifelong cognitive dynamics.
- Score: 27.81862535460598
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Cognitive dynamics are pivotal to advance human understanding of the world.
Recent advancements in large language models (LLMs) reveal their potential for
cognitive simulation. However, these LLM-based cognitive studies primarily
focus on static modeling, overlooking the dynamic nature of cognition. To
bridge this gap, we propose the concept of the cognitive dynamics of LLMs and
present a corresponding task with the inspiration of longitudinal studies.
Towards the task, we develop CogBench, a novel benchmark to assess the
cognitive dynamics of LLMs and validate it through participant surveys. We also
design two evaluation metrics for CogBench, including Authenticity and
Rationality. Recognizing the inherent static nature of LLMs, we introduce
CogGPT for the task, which features an innovative iterative cognitive mechanism
aimed at enhancing lifelong cognitive dynamics. Empirical results demonstrate
the superiority of CogGPT over existing methods, particularly in its ability to
facilitate role-specific cognitive dynamics under continuous information flows.
Related papers
- Exploring the LLM Journey from Cognition to Expression with Linear Representations [10.92882688742428]
This paper presents an in-depth examination of the evolution and interplay of cognitive and expressive capabilities in large language models (LLMs)
We define and explore the model's cognitive and expressive capabilities through linear representations across three critical phases: Pretraining, Supervised Fine-Tuning (SFT), and Reinforcement Learning from Human Feedback (RLHF)
Our findings unveil a sequential development pattern, where cognitive abilities are largely established during Pretraining, whereas expressive abilities predominantly advance during SFT and RLHF.
arXiv Detail & Related papers (2024-05-27T08:57:04Z) - Metacognitive Retrieval-Augmented Large Language Models [43.57020180706832]
This paper introduces MetaRAG, an approach that combines the retrieval-augmented generation process with metacognition.
By integrating this, MetaRAG enables the model to monitor, evaluate, and plan its response strategies.
Empirical evaluations show that MetaRAG significantly outperforms existing methods.
arXiv Detail & Related papers (2024-02-18T15:41:31Z) - K-Level Reasoning with Large Language Models [80.13817747270029]
We explore the dynamic reasoning capabilities of Large Language Models (LLMs) for decision-making in rapidly evolving environments.
We introduce two game theory-based pilot challenges that mirror the complexities of real-world dynamic decision-making.
These challenges are well-defined, enabling clear, controllable, and precise evaluation of LLMs' dynamic reasoning abilities.
arXiv Detail & Related papers (2024-02-02T16:07:05Z) - Interpreting Pretrained Language Models via Concept Bottlenecks [55.47515772358389]
Pretrained language models (PLMs) have made significant strides in various natural language processing tasks.
The lack of interpretability due to their black-box'' nature poses challenges for responsible implementation.
We propose a novel approach to interpreting PLMs by employing high-level, meaningful concepts that are easily understandable for humans.
arXiv Detail & Related papers (2023-11-08T20:41:18Z) - Exploring the Cognitive Knowledge Structure of Large Language Models: An
Educational Diagnostic Assessment Approach [50.125704610228254]
Large Language Models (LLMs) have not only exhibited exceptional performance across various tasks, but also demonstrated sparks of intelligence.
Recent studies have focused on assessing their capabilities on human exams and revealed their impressive competence in different domains.
We conduct an evaluation using MoocRadar, a meticulously annotated human test dataset based on Bloom taxonomy.
arXiv Detail & Related papers (2023-10-12T09:55:45Z) - A Novel Neural-symbolic System under Statistical Relational Learning [50.747658038910565]
We propose a general bi-level probabilistic graphical reasoning framework called GBPGR.
In GBPGR, the results of symbolic reasoning are utilized to refine and correct the predictions made by the deep learning models.
Our approach achieves high performance and exhibits effective generalization in both transductive and inductive tasks.
arXiv Detail & Related papers (2023-09-16T09:15:37Z) - Synergistic Integration of Large Language Models and Cognitive
Architectures for Robust AI: An Exploratory Analysis [12.9222727028798]
This paper explores the integration of two AI subdisciplines employed in the development of artificial agents that exhibit intelligent behavior: Large Language Models (LLMs) and Cognitive Architectures (CAs)
We present three integration approaches, each grounded in theoretical models and supported by preliminary empirical evidence.
These approaches aim to harness the strengths of both LLMs and CAs, while mitigating their weaknesses, thereby advancing the development of more robust AI systems.
arXiv Detail & Related papers (2023-08-18T21:42:47Z) - Modelling human logical reasoning process in dynamic environmental
stress with cognitive agents [13.171768256928509]
We propose a cognitive agent integrating drift-diffusion with deep reinforcement learning to simulate granular stress effects on logical reasoning process.
Leveraging a large dataset of 21,157 logical responses, we investigate performance impacts of dynamic stress.
Quantitatively, the framework improves cognition modelling by capturing both subject-specific and stimuli-specific behavioural differences.
Overall, this work demonstrates a powerful, data-driven methodology to simulate and understand the vagaries of human logical reasoning process in dynamic contexts.
arXiv Detail & Related papers (2023-01-15T23:46:37Z) - CogNGen: Constructing the Kernel of a Hyperdimensional Predictive
Processing Cognitive Architecture [79.07468367923619]
We present a new cognitive architecture that combines two neurobiologically plausible, computational models.
We aim to develop a cognitive architecture that has the power of modern machine learning techniques.
arXiv Detail & Related papers (2022-03-31T04:44:28Z) - Computational Metacognition [2.0552049801885746]
Computational metacognition represents a cognitive systems perspective on high-order reasoning in integrated artificial systems.
We show how computational metacognition improves performance by changing cognition through meta-level goal operations and learning.
arXiv Detail & Related papers (2022-01-30T17:34:53Z) - Backprop-Free Reinforcement Learning with Active Neural Generative
Coding [84.11376568625353]
We propose a computational framework for learning action-driven generative models without backpropagation of errors (backprop) in dynamic environments.
We develop an intelligent agent that operates even with sparse rewards, drawing inspiration from the cognitive theory of planning as inference.
The robust performance of our agent offers promising evidence that a backprop-free approach for neural inference and learning can drive goal-directed behavior.
arXiv Detail & Related papers (2021-07-10T19:02:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.