OmniThink: Expanding Knowledge Boundaries in Machine Writing through Thinking
- URL: http://arxiv.org/abs/2501.09751v4
- Date: Wed, 01 Oct 2025 08:33:49 GMT
- Title: OmniThink: Expanding Knowledge Boundaries in Machine Writing through Thinking
- Authors: Zekun Xi, Wenbiao Yin, Jizhan Fang, Jialong Wu, Runnan Fang, Yong Jiang, Pengjun Xie, Fei Huang, Huajun Chen, Ningyu Zhang,
- Abstract summary: Machine writing with large language models often relies on retrieval-augmented generation.<n> vanilla-retrieved information tends to lack depth, novelty, and suffers from redundancy.<n>We propose OmniThink, a slow-thinking machine writing framework that emulates the human-like process of iterative expansion and reflection.
- Score: 95.01555654262
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Machine writing with large language models often relies on retrieval-augmented generation. However, these approaches remain confined within the boundaries of the model's predefined scope, limiting the generation of content with rich information. Specifically, vanilla-retrieved information tends to lack depth, novelty, and suffers from redundancy, which negatively impacts the quality of generated articles, leading to shallow, unoriginal, and repetitive outputs. To address these issues, we propose OmniThink, a slow-thinking machine writing framework that emulates the human-like process of iterative expansion and reflection. The core idea behind OmniThink is to simulate the cognitive behavior of learners as they slowly deepen their knowledge of the topics. Experimental results demonstrate that OmniThink improves the knowledge density of generated articles without compromising metrics such as coherence and depth. Human evaluations and expert feedback further highlight the potential of OmniThink to address real-world challenges in the generation of long-form articles. Code is available at https://github.com/zjunlp/OmniThink.
Related papers
- SmartSwitch: Advancing LLM Reasoning by Overcoming Underthinking via Promoting Deeper Thought Exploration [49.290631188365786]
Long chain-of-thought (LongCoT) is central to the recent breakthroughs achieved by large language models in complex reasoning tasks.<n>We propose a simple yet effective reasoning strategy: the SmartSwitch inference framework.<n>This framework can be easily integrated into any large language model as a plug-and-play solution.
arXiv Detail & Related papers (2025-10-22T16:56:01Z) - CoCoNUTS: Concentrating on Content while Neglecting Uninformative Textual Styles for AI-Generated Peer Review Detection [60.52240468810558]
We introduce CoCoNUTS, a content-oriented benchmark built upon a fine-grained dataset of AI-generated peer reviews.<n>We also develop CoCoDet, an AI review detector via a multi-task learning framework, to achieve more accurate and robust detection of AI involvement in review content.
arXiv Detail & Related papers (2025-08-28T06:03:11Z) - Hop, Skip, and Overthink: Diagnosing Why Reasoning Models Fumble during Multi-Hop Analysis [3.711555701154055]
Reasoning models and their integration into practical AI chat bots have led to breakthroughs in solving advanced math, deep search, and extractive question answering problems.<n>Yet, a complete understanding of why these models hallucinate more than general purpose language models is missing.<n>In this study, we systematicallyexplore reasoning failures of contemporary language models on multi-hop question answering tasks.
arXiv Detail & Related papers (2025-08-06T17:58:36Z) - DeepWriter: A Fact-Grounded Multimodal Writing Assistant Based On Offline Knowledge Base [10.572896494190669]
We introduce DeepWriter, a customizable, multimodal, long-form writing assistant that operates on a curated, offline knowledge base.<n>By deeply mining information from a structured corpus, DeepWriter generates coherent, factually grounded, and professional-grade documents.<n>Our experiments on financial report generation demonstrate that DeepWriter produces high-quality, verifiable articles.
arXiv Detail & Related papers (2025-07-14T02:13:22Z) - Machine Learning: a Lecture Note [51.31735291774885]
This lecture note is intended to prepare early-year master's and PhD students in data science or a related discipline with foundational ideas in machine learning.<n>It starts with basic ideas in modern machine learning with classification as a main target task.<n>Based on these basic ideas, the lecture note explores in depth the probablistic approach to unsupervised learning.
arXiv Detail & Related papers (2025-05-06T16:03:41Z) - Information Retrieval in the Age of Generative AI: The RGB Model [77.96475639967431]
This paper presents a novel quantitative approach to shed light on the complex information dynamics arising from the growing use of generative AI tools.
We propose a model to characterize the generation, indexing, and dissemination of information in response to new topics.
Our findings suggest that the rapid pace of generative AI adoption, combined with increasing user reliance, can outpace human verification, escalating the risk of inaccurate information proliferation.
arXiv Detail & Related papers (2025-04-29T10:21:40Z) - Multi-Novelty: Improve the Diversity and Novelty of Contents Generated by Large Language Models via inference-time Multi-Views Brainstorming [3.591342811819669]
Large Language Models (LLMs) demonstrate remarkable proficiency in generating accurate and fluent text.
They often struggle with diversity and novelty, leading to repetitive or overly deterministic responses.
We introduce inference-time multi-view brainstorming method, a novel approach that enriches input prompts with diverse perspectives.
arXiv Detail & Related papers (2025-02-18T10:04:20Z) - Think More, Hallucinate Less: Mitigating Hallucinations via Dual Process of Fast and Slow Thinking [124.69672273754144]
HaluSearch is a novel framework that incorporates tree search-based algorithms.
It frames text generation as a step-by-step reasoning process.
We introduce a hierarchical thinking system switch mechanism inspired by the dual process theory in cognitive science.
arXiv Detail & Related papers (2025-01-02T15:36:50Z) - The Superalignment of Superhuman Intelligence with Large Language Models [63.96120398355404]
We discuss the concept of superalignment from the learning perspective to answer this question.<n>We highlight some key research problems in superalignment, namely, weak-to-strong generalization, scalable oversight, and evaluation.<n>We present a conceptual framework for superalignment, which consists of three modules: an attacker which generates adversary queries trying to expose the weaknesses of a learner model; a learner which will refine itself by learning from scalable feedbacks generated by a critic model along with minimal human experts; and a critic which generates critics or explanations for a given query-response pair, with a target of improving the learner by criticizing.
arXiv Detail & Related papers (2024-12-15T10:34:06Z) - Can Large Language Models Unlock Novel Scientific Research Ideas? [31.88070174767799]
This study examines the ability of Large Language Models (LLMs) to generate future research ideas from scientific papers.<n>Human evaluation in this setting is extremely challenging ie: it requires substantial domain expertise, contextual understanding of the paper, and awareness of the current research landscape.<n>We propose two automated evaluation metrics: Idea Alignment Score (IAScore) and Idea Distinctness Index.
arXiv Detail & Related papers (2024-09-10T03:26:42Z) - Good Idea or Not, Representation of LLM Could Tell [86.36317971482755]
We focus on idea assessment, which aims to leverage the knowledge of large language models to assess the merit of scientific ideas.
We release a benchmark dataset from nearly four thousand manuscript papers with full texts, meticulously designed to train and evaluate the performance of different approaches to this task.
Our findings suggest that the representations of large language models hold more potential in quantifying the value of ideas than their generative outputs.
arXiv Detail & Related papers (2024-09-07T02:07:22Z) - Analysis of Plan-based Retrieval for Grounded Text Generation [78.89478272104739]
hallucinations occur when a language model is given a generation task outside its parametric knowledge.
A common strategy to address this limitation is to infuse the language models with retrieval mechanisms.
We analyze how planning can be used to guide retrieval to further reduce the frequency of hallucinations.
arXiv Detail & Related papers (2024-08-20T02:19:35Z) - Machine Unlearning in Generative AI: A Survey [19.698620794387338]
Generative AI technologies have been deployed in many places, such as (multimodal) large language models and vision generative models.
New machine unlearning (MU) techniques are being developed to reduce or eliminate undesirable knowledge.
arXiv Detail & Related papers (2024-07-30T03:26:09Z) - Bridging the Intent Gap: Knowledge-Enhanced Visual Generation [27.568260631117365]
For visual content generation, discrepancies between user intentions and the generated content have been a longstanding problem.
We propose a knowledge-enhanced iterative refinement framework for visual content generation.
We demonstrate the efficacy of the proposed framework through preliminary results.
arXiv Detail & Related papers (2024-05-21T07:07:44Z) - Collapse of Self-trained Language Models [0.0]
We explore the potential of self-training models on their own outputs, akin to how humans learn and build on their previous thoughts and actions.
We find that extended self-training of the GPT-2 model leads to a significant degradation in performance, resulting in repetitive and collapsed token output.
arXiv Detail & Related papers (2024-04-02T21:03:37Z) - Awakening Augmented Generation: Learning to Awaken Internal Knowledge of Large Language Models for Question Answering [30.409828862670764]
A novel knowledge-augmented framework, $textbfAwakening-Augmented-Generation$ (AAG), is proposed.<n>Explicit awakening fine-tunes a context generator to create a synthetic, compressed document that functions as symbolic context.<n> Implicit awakening utilizes a hypernetwork to generate adapters based on the question and synthetic document, which are inserted into Large Language Models.
arXiv Detail & Related papers (2024-03-22T15:06:45Z) - ThinkNote: Enhancing Knowledge Integration and Utilization of Large Language Models via Constructivist Cognition Modeling [55.21641515545307]
Large Language Models (LLMs) have demonstrated strong performance across a wide range of NLP tasks.<n>They often exhibit suboptimal behaviors and inconsistencies when exposed to unfamiliar external information.<n>We propose ThinkNote, a novel framework that enhances the external knowledge utilization of LLMs.
arXiv Detail & Related papers (2024-02-21T06:04:53Z) - Beyond Factuality: A Comprehensive Evaluation of Large Language Models
as Knowledge Generators [78.63553017938911]
Large language models (LLMs) outperform information retrieval techniques for downstream knowledge-intensive tasks.
However, community concerns abound regarding the factuality and potential implications of using this uncensored knowledge.
We introduce CONNER, designed to evaluate generated knowledge from six important perspectives.
arXiv Detail & Related papers (2023-10-11T08:22:37Z) - AI for the Generation and Testing of Ideas Towards an AI Supported
Knowledge Development Environment [2.0305676256390934]
We discuss how generative AI can boost idea generation by eliminating human bias.
We also describe how search can verify facts, logic, and context.
This paper introduces a system for knowledge workers, Generate And Search Test, enabling individuals to efficiently create solutions.
arXiv Detail & Related papers (2023-07-17T22:17:40Z) - The Creative Frontier of Generative AI: Managing the Novelty-Usefulness
Tradeoff [0.4873362301533825]
We explore the optimal balance between novelty and usefulness in generative Artificial Intelligence (AI) systems.
Overemphasizing either aspect can lead to limitations such as hallucinations and memorization.
arXiv Detail & Related papers (2023-06-06T11:44:57Z) - Factual and Informative Review Generation for Explainable Recommendation [41.403493319602816]
Previous models' generated content often contain factual hallucinations.
Inspired by recent success in using retrieved content in addition to parametric knowledge for generation, we propose to augment the generator with a personalized retriever.
Experiments on Yelp, TripAdvisor, and Amazon Movie Reviews dataset show our model could generate explanations that more reliably entail existing reviews, are more diverse, and are rated more informative by human evaluators.
arXiv Detail & Related papers (2022-09-12T16:46:47Z) - Generating Diversified Comments via Reader-Aware Topic Modeling and
Saliency Detection [25.16392119801612]
We propose a reader-aware topic modeling and saliency information detection framework to enhance the quality of generated comments.
For reader-aware topic modeling, we design a variational generative clustering algorithm for latent semantic learning and topic mining from reader comments.
For saliency information detection, we introduce Bernoulli distribution estimating on news content to select saliency information.
arXiv Detail & Related papers (2021-02-13T03:50:31Z) - Can We Automate Scientific Reviewing? [89.50052670307434]
We discuss the possibility of using state-of-the-art natural language processing (NLP) models to generate first-pass peer reviews for scientific papers.
We collect a dataset of papers in the machine learning domain, annotate them with different aspects of content covered in each review, and train targeted summarization models that take in papers to generate reviews.
Comprehensive experimental results show that system-generated reviews tend to touch upon more aspects of the paper than human-written reviews.
arXiv Detail & Related papers (2021-01-30T07:16:53Z) - Knowledge as Invariance -- History and Perspectives of
Knowledge-augmented Machine Learning [69.99522650448213]
Research in machine learning is at a turning point.
Research interests are shifting away from increasing the performance of highly parameterized models to exceedingly specific tasks.
This white paper provides an introduction and discussion of this emerging field in machine learning research.
arXiv Detail & Related papers (2020-12-21T15:07:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.