Zero- and Few-Shots Knowledge Graph Triplet Extraction with Large
Language Models
- URL: http://arxiv.org/abs/2312.01954v1
- Date: Mon, 4 Dec 2023 15:12:04 GMT
- Title: Zero- and Few-Shots Knowledge Graph Triplet Extraction with Large
Language Models
- Authors: Andrea Papaluca, Daniel Krefl, Sergio Mendez Rodriguez, Artem Lensky,
Hanna Suominen
- Abstract summary: In this work, we tested the Triplet Extraction capabilities of a variety of Large Language Models (LLMs) of different sizes in the Zero- and Few-Shots settings.
We proposed a pipeline that dynamically gathers contextual information from a Knowledge Base (KB), both in the form of context triplets and of (sentence, triplets) pairs as examples.
- Score: 7.919349589245355
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this work, we tested the Triplet Extraction (TE) capabilities of a variety
of Large Language Models (LLMs) of different sizes in the Zero- and Few-Shots
settings. In detail, we proposed a pipeline that dynamically gathers contextual
information from a Knowledge Base (KB), both in the form of context triplets
and of (sentence, triplets) pairs as examples, and provides it to the LLM
through a prompt. The additional context allowed the LLMs to be competitive
with all the older fully trained baselines based on the Bidirectional Long
Short-Term Memory (BiLSTM) Network architecture. We further conducted a
detailed analysis of the quality of the gathered KB context, finding it to be
strongly correlated with the final TE performance of the model. In contrast,
the size of the model appeared to only logarithmically improve the TE
capabilities of the LLMs.
Related papers
- TituLLMs: A Family of Bangla LLMs with Comprehensive Benchmarking [6.070192392563392]
We present TituLLMs, the first large pretrained Bangla LLMs, available in 1b and 3b parameter sizes.
To train TituLLMs, we collected a pretraining dataset of approximately 37 billion tokens.
We extended the Llama-3.2 tokenizer to incorporate language- and culture-specific knowledge.
arXiv Detail & Related papers (2025-02-16T16:22:23Z) - BaichuanSEED: Sharing the Potential of ExtensivE Data Collection and Deduplication by Introducing a Competitive Large Language Model Baseline [34.518474035662905]
General capabilities of Large Language Models (LLM) highly rely on extensive pretraining datasets, treated as commercial secrets by several institutions.
We open-source the details of a universally applicable data processing pipeline to validate its effectiveness and potential.
BaichuanSEED demonstrates consistency and predictability throughout training and achieves comparable performance on comprehensive benchmarks.
arXiv Detail & Related papers (2024-08-27T14:08:23Z) - TriSum: Learning Summarization Ability from Large Language Models with Structured Rationale [66.01943465390548]
We introduce TriSum, a framework for distilling large language models' text summarization abilities into a compact, local model.
Our method enhances local model performance on various benchmarks.
It also improves interpretability by providing insights into the summarization rationale.
arXiv Detail & Related papers (2024-03-15T14:36:38Z) - Contextualization Distillation from Large Language Model for Knowledge
Graph Completion [51.126166442122546]
We introduce the Contextualization Distillation strategy, a plug-in-and-play approach compatible with both discriminative and generative KGC frameworks.
Our method begins by instructing large language models to transform compact, structural triplets into context-rich segments.
Comprehensive evaluations across diverse datasets and KGC techniques highlight the efficacy and adaptability of our approach.
arXiv Detail & Related papers (2024-01-28T08:56:49Z) - LLM Augmented LLMs: Expanding Capabilities through Composition [56.40953749310957]
CALM -- Composition to Augment Language Models -- introduces cross-attention between models to compose their representations and enable new capabilities.
We illustrate that augmenting PaLM2-S with a smaller model trained on low-resource languages results in an absolute improvement of up to 13% on tasks like translation into English.
When PaLM2-S is augmented with a code-specific model, we see a relative improvement of 40% over the base model for code generation and explanation tasks.
arXiv Detail & Related papers (2024-01-04T18:53:01Z) - BLESS: Benchmarking Large Language Models on Sentence Simplification [55.461555829492866]
We present BLESS, a performance benchmark of the most recent state-of-the-art large language models (LLMs) on the task of text simplification (TS)
We assess a total of 44 models, differing in size, architecture, pre-training methods, and accessibility, on three test sets from different domains (Wikipedia, news, and medical) under a few-shot setting.
Our evaluation indicates that the best LLMs, despite not being trained on TS, perform comparably with state-of-the-art TS baselines.
arXiv Detail & Related papers (2023-10-24T12:18:17Z) - Improving Open Information Extraction with Large Language Models: A
Study on Demonstration Uncertainty [52.72790059506241]
Open Information Extraction (OIE) task aims at extracting structured facts from unstructured text.
Despite the potential of large language models (LLMs) like ChatGPT as a general task solver, they lag behind state-of-the-art (supervised) methods in OIE tasks.
arXiv Detail & Related papers (2023-09-07T01:35:24Z) - LLM-Pruner: On the Structural Pruning of Large Language Models [65.02607075556742]
Large language models (LLMs) have shown remarkable capabilities in language understanding and generation.
We tackle the compression of LLMs within the bound of two constraints: being task-agnostic and minimizing the reliance on the original training dataset.
Our method, named LLM-Pruner, adopts structural pruning that selectively removes non-critical coupled structures.
arXiv Detail & Related papers (2023-05-19T12:10:53Z) - LP-BERT: Multi-task Pre-training Knowledge Graph BERT for Link
Prediction [3.5382535469099436]
LP-BERT contains two training stages: multi-task pre-training and knowledge graph fine-tuning.
We achieve state-of-the-art results on WN18RR and UMLS datasets, especially the Hits@10 indicator improved by 5%.
arXiv Detail & Related papers (2022-01-13T09:18:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.