Experiments in Extractive Summarization: Integer Linear Programming,
Term/Sentence Scoring, and Title-driven Models
- URL: http://arxiv.org/abs/2008.00140v1
- Date: Sat, 1 Aug 2020 01:05:55 GMT
- Title: Experiments in Extractive Summarization: Integer Linear Programming,
Term/Sentence Scoring, and Title-driven Models
- Authors: Daniel Lee and Rakesh Verma and Avisha Das and Arjun Mukherjee
- Abstract summary: We describe a new framework, NewsSumm, that includes many existing and new approaches for summarization including ILP and title-driven approaches.
We show that the new title-driven reduction idea leads to improvement in performance for both unsupervised and supervised approaches considered.
- Score: 1.3286165491120467
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we revisit the challenging problem of unsupervised
single-document summarization and study the following aspects: Integer linear
programming (ILP) based algorithms, Parameterized normalization of term and
sentence scores, and Title-driven approaches for summarization. We describe a
new framework, NewsSumm, that includes many existing and new approaches for
summarization including ILP and title-driven approaches. NewsSumm's flexibility
allows to combine different algorithms and sentence scoring schemes seamlessly.
Our results combining sentence scoring with ILP and normalization are in
contrast to previous work on this topic, showing the importance of a broader
search for optimal parameters. We also show that the new title-driven reduction
idea leads to improvement in performance for both unsupervised and supervised
approaches considered.
Related papers
- Self-Calibrated Listwise Reranking with Large Language Models [137.6557607279876]
Large language models (LLMs) have been employed in reranking tasks through a sequence-to-sequence approach.
This reranking paradigm requires a sliding window strategy to iteratively handle larger candidate sets.
We propose a novel self-calibrated listwise reranking method, which aims to leverage LLMs to produce global relevance scores for ranking.
arXiv Detail & Related papers (2024-11-07T10:31:31Z) - Absolute Ranking: An Essential Normalization for Benchmarking Optimization Algorithms [0.0]
evaluating performance across optimization algorithms on many problems presents a complex challenge due to the diversity of numerical scales involved.
This paper extensively explores the problem, making a compelling case to underscore the issue and conducting a thorough analysis of its root causes.
Building on this research, this paper introduces a new mathematical model called "absolute ranking" and a sampling-based computational method.
arXiv Detail & Related papers (2024-09-06T00:55:03Z) - PARL: A Unified Framework for Policy Alignment in Reinforcement Learning from Human Feedback [106.63518036538163]
We present a novel unified bilevel optimization-based framework, textsfPARL, formulated to address the recently highlighted critical issue of policy alignment in reinforcement learning.
Our framework addressed these concerns by explicitly parameterizing the distribution of the upper alignment objective (reward design) by the lower optimal variable.
Our empirical results substantiate that the proposed textsfPARL can address the alignment concerns in RL by showing significant improvements.
arXiv Detail & Related papers (2023-08-03T18:03:44Z) - Enhancing Topic Extraction in Recommender Systems with Entropy
Regularization [2.7286395031146062]
This paper introduces a novel approach called entropy regularization to address the issue of low explainability of recommender systems.
Experiment results show a significant improvement in topic coherence, which is quantified by cosine similarity on word embeddings.
arXiv Detail & Related papers (2023-06-12T20:05:09Z) - Efficient Alternating Minimization Solvers for Wyner Multi-View
Unsupervised Learning [0.0]
We propose two novel formulations that enable the development of computational efficient solvers based the alternating principle.
The proposed solvers offer computational efficiency, theoretical convergence guarantees, local minima complexity with the number of views, and exceptional accuracy as compared with the state-of-the-art techniques.
arXiv Detail & Related papers (2023-03-28T10:17:51Z) - Linear Partial Monitoring for Sequential Decision-Making: Algorithms,
Regret Bounds and Applications [70.67112733968654]
Partial monitoring is an expressive framework for sequential decision-making.
We present a simple and unified analysis of partial monitoring, and further extend the model to the contextual and kernelized setting.
arXiv Detail & Related papers (2023-02-07T18:58:25Z) - A Character-Level Length-Control Algorithm for Non-Autoregressive
Sentence Summarization [23.495225374478295]
Sentence summarization aims at compressing a long sentence into a short one that keeps the main gist, and has extensive real-world applications such as headline generation.
In our work, we address a new problem of explicit character-level length control for summarization, and propose a dynamic programming algorithm based on the Connectionist Temporal Classification (CTC) model.
arXiv Detail & Related papers (2022-05-28T21:09:53Z) - Adaptive Affinity Loss and Erroneous Pseudo-Label Refinement for Weakly
Supervised Semantic Segmentation [48.294903659573585]
In this paper, we propose to embed affinity learning of multi-stage approaches in a single-stage model.
A deep neural network is used to deliver comprehensive semantic information in the training phase.
Experiments are conducted on the PASCAL VOC 2012 dataset to evaluate the effectiveness of our proposed approach.
arXiv Detail & Related papers (2021-08-03T07:48:33Z) - Multi-Fact Correction in Abstractive Text Summarization [98.27031108197944]
Span-Fact is a suite of two factual correction models that leverages knowledge learned from question answering models to make corrections in system-generated summaries via span selection.
Our models employ single or multi-masking strategies to either iteratively or auto-regressively replace entities in order to ensure semantic consistency w.r.t. the source text.
Experiments show that our models significantly boost the factual consistency of system-generated summaries without sacrificing summary quality in terms of both automatic metrics and human evaluation.
arXiv Detail & Related papers (2020-10-06T02:51:02Z) - Semi-Supervised Learning with Meta-Gradient [123.26748223837802]
We propose a simple yet effective meta-learning algorithm in semi-supervised learning.
We find that the proposed algorithm performs favorably against state-of-the-art methods.
arXiv Detail & Related papers (2020-07-08T08:48:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.