SikuGPT: A Generative Pre-trained Model for Intelligent Information
Processing of Ancient Texts from the Perspective of Digital Humanities
- URL: http://arxiv.org/abs/2304.07778v1
- Date: Sun, 16 Apr 2023 13:25:24 GMT
- Title: SikuGPT: A Generative Pre-trained Model for Intelligent Information
Processing of Ancient Texts from the Perspective of Digital Humanities
- Authors: Liu Chang, Wang Dongbo, Zhao Zhixiao, Hu Die, Wu Mengcheng, Lin Litao,
Shen Si, Li Bin, Liu Jiangfeng, Zhang Hai, Zhao Lianzheng
- Abstract summary: We propose a GPT model called SikuGPT based on the corpus of Siku Quanshu.
The model's performance in tasks such as intralingual translation and text classification exceeds that of other GPT-type models.
SikuGPT's ability to process traditional Chinese ancient texts can help promote the organization of ancient information and knowledge services.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The rapid advance in artificial intelligence technology has facilitated the
prosperity of digital humanities research. Against such backdrop, research
methods need to be transformed in the intelligent processing of ancient texts,
which is a crucial component of digital humanities research, so as to adapt to
new development trends in the wave of AIGC. In this study, we propose a GPT
model called SikuGPT based on the corpus of Siku Quanshu. The model's
performance in tasks such as intralingual translation and text classification
exceeds that of other GPT-type models aimed at processing ancient texts.
SikuGPT's ability to process traditional Chinese ancient texts can help promote
the organization of ancient information and knowledge services, as well as the
international dissemination of Chinese ancient culture.
Related papers
- From ChatGPT, DALL-E 3 to Sora: How has Generative AI Changed Digital Humanities Research and Services? [5.3743115255502545]
This article profoundly explores the application of large-scale language models in digital humanities research.
The article first outlines the importance of ancient book resources and the necessity of digital preservation.
Through specific cases, the article demonstrates how AI can assist in the organization, classification, and content generation of ancient books.
arXiv Detail & Related papers (2024-04-29T09:03:19Z) - RenAIssance: A Survey into AI Text-to-Image Generation in the Era of
Large Model [93.8067369210696]
Text-to-image generation (TTI) refers to the usage of models that could process text input and generate high fidelity images based on text descriptions.
Diffusion models are one prominent type of generative model used for the generation of images through the systematic introduction of noises with repeating steps.
In the era of large models, scaling up model size and the integration with large language models have further improved the performance of TTI models.
arXiv Detail & Related papers (2023-09-02T03:27:20Z) - GujiBERT and GujiGPT: Construction of Intelligent Information Processing
Foundation Language Models for Ancient Texts [11.289265479095956]
GujiBERT and GujiGPT language models are foundational models specifically designed for intelligent information processing of ancient texts.
These models have been trained on an extensive dataset that encompasses both simplified and traditional Chinese characters.
These models have exhibited exceptional performance across a range of validation tasks using publicly available datasets.
arXiv Detail & Related papers (2023-07-11T15:44:01Z) - SciMON: Scientific Inspiration Machines Optimized for Novelty [68.46036589035539]
We explore and enhance the ability of neural language models to generate novel scientific directions grounded in literature.
We take a dramatic departure with a novel setting in which models use as input background contexts.
We present SciMON, a modeling framework that uses retrieval of "inspirations" from past scientific papers.
arXiv Detail & Related papers (2023-05-23T17:12:08Z) - Semantic Similarity Measure of Natural Language Text through Machine
Learning and a Keyword-Aware Cross-Encoder-Ranking Summarizer -- A Case Study
Using UCGIS GIS&T Body of Knowledge [2.4909170697740968]
GIS&T Body of Knowledge (BoK) is a community-driven endeavor to define, develop, and document geospatial topics.
This research evaluates the effectiveness of multiple natural language processing (NLP) techniques in extracting semantics from text.
It also offers a new perspective on the use of machine learning techniques for analyzing scientific publications.
arXiv Detail & Related papers (2023-05-17T01:17:57Z) - ChatGPT and a New Academic Reality: Artificial Intelligence-Written
Research Papers and the Ethics of the Large Language Models in Scholarly
Publishing [6.109522330180625]
ChatGPT is a generative pre-trained transformer, which uses natural language processing to fulfill text-based user requests.
Potential ethical issues that could arise with the emergence of large language models like GPT-3 are discussed.
arXiv Detail & Related papers (2023-03-21T14:35:07Z) - A Survey of Text Representation Methods and Their Genealogy [0.0]
In recent years, with the advent of highly scalable artificial-neural-network-based text representation methods the field of natural language processing has seen unprecedented growth and sophistication.
We provide a survey of current approaches, by arranging them in a genealogy, and by conceptualizing a taxonomy of text representation methods to examine and explain the state-of-the-art.
arXiv Detail & Related papers (2022-11-26T15:22:01Z) - A Survey on Retrieval-Augmented Text Generation [53.04991859796971]
Retrieval-augmented text generation has remarkable advantages and has achieved state-of-the-art performance in many NLP tasks.
It firstly highlights the generic paradigm of retrieval-augmented generation, and then it reviews notable approaches according to different tasks.
arXiv Detail & Related papers (2022-02-02T16:18:41Z) - A Survey of Knowledge-Enhanced Text Generation [81.24633231919137]
The goal of text generation is to make machines express in human language.
Various neural encoder-decoder models have been proposed to achieve the goal by learning to map input text to output text.
To address this issue, researchers have considered incorporating various forms of knowledge beyond the input text into the generation models.
arXiv Detail & Related papers (2020-10-09T06:46:46Z) - Generating Major Types of Chinese Classical Poetry in a Uniformed
Framework [88.57587722069239]
We propose a GPT-2 based framework for generating major types of Chinese classical poems.
Preliminary results show this enhanced model can generate Chinese classical poems of major types with high quality in both form and content.
arXiv Detail & Related papers (2020-03-13T14:16:25Z) - Exploring the Limits of Transfer Learning with a Unified Text-to-Text
Transformer [64.22926988297685]
Transfer learning, where a model is first pre-trained on a data-rich task before being fine-tuned on a downstream task, has emerged as a powerful technique in natural language processing (NLP)
In this paper, we explore the landscape of introducing transfer learning techniques for NLP by a unified framework that converts all text-based language problems into a text-to-text format.
arXiv Detail & Related papers (2019-10-23T17:37:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.