NILC-Metrix: assessing the complexity of written and spoken language in
Brazilian Portuguese
- URL: http://arxiv.org/abs/2201.03445v1
- Date: Fri, 17 Dec 2021 16:51:00 GMT
- Title: NILC-Metrix: assessing the complexity of written and spoken language in
Brazilian Portuguese
- Authors: Sidney Evaldo Leal and Magali Sanches Duran and Carolina Evaristo
Scarton and Nathan Siegle Hartmann and Sandra Maria Alu\'isio
- Abstract summary: This paper presents and makes publicly available the NILC-Metrix, a computational system comprising 200 metrics proposed in studies on discourse.
The metrics in NILC-Metrix were developed during the last 13 years, starting in 2008 with Coh-Metrix-Port, a tool developed within the scope of the PorSimples project.
- Score: 0.32622301272834514
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: This paper presents and makes publicly available the NILC-Metrix, a
computational system comprising 200 metrics proposed in studies on discourse,
psycholinguistics, cognitive and computational linguistics, to assess textual
complexity in Brazilian Portuguese (BP). These metrics are relevant for
descriptive analysis and the creation of computational models and can be used
to extract information from various linguistic levels of written and spoken
language. The metrics in NILC-Metrix were developed during the last 13 years,
starting in 2008 with Coh-Metrix-Port, a tool developed within the scope of the
PorSimples project. Coh-Metrix-Port adapted some metrics to BP from the
Coh-Metrix tool that computes metrics related to cohesion and coherence of
texts in English. After the end of PorSimples in 2010, new metrics were added
to the initial 48 metrics of Coh-Metrix-Port. Given the large number of
metrics, we present them following an organisation similar to the metrics of
Coh-Metrix v3.0 to facilitate comparisons made with metrics in Portuguese and
English. In this paper, we illustrate the potential of NILC-Metrix by
presenting three applications: (i) a descriptive analysis of the differences
between children's film subtitles and texts written for Elementary School I and
II (Final Years); (ii) a new predictor of textual complexity for the corpus of
original and simplified texts of the PorSimples project; (iii) a complexity
prediction model for school grades, using transcripts of children's story
narratives told by teenagers. For each application, we evaluate which groups of
metrics are more discriminative, showing their contribution for each task.
Related papers
- Rethinking Evaluation Metrics of Open-Vocabulary Segmentaion [78.76867266561537]
The evaluation process still heavily relies on closed-set metrics without considering the similarity between predicted and ground truth categories.
To tackle this issue, we first survey eleven similarity measurements between two categorical words.
We designed novel evaluation metrics, namely Open mIoU, Open AP, and Open PQ, tailored for three open-vocabulary segmentation tasks.
arXiv Detail & Related papers (2023-11-06T18:59:01Z) - StyloMetrix: An Open-Source Multilingual Tool for Representing
Stylometric Vectors [0.0]
This work aims to provide an overview of the open-source multilanguage tool called StyloMetrix.
It offers stylometric text representations that cover various aspects of grammar, syntax and lexicon.
StyloMetrix covers four languages: Polish as the primary language, English, Ukrainian and Russian.
arXiv Detail & Related papers (2023-09-22T11:53:47Z) - LyricSIM: A novel Dataset and Benchmark for Similarity Detection in
Spanish Song LyricS [52.77024349608834]
We present a new dataset and benchmark tailored to the task of semantic similarity in song lyrics.
Our dataset, originally consisting of 2775 pairs of Spanish songs, was annotated in a collective annotation experiment by 63 native annotators.
arXiv Detail & Related papers (2023-06-02T07:48:20Z) - Multi-Dimensional Evaluation of Text Summarization with In-Context
Learning [79.02280189976562]
In this paper, we study the efficacy of large language models as multi-dimensional evaluators using in-context learning.
Our experiments show that in-context learning-based evaluators are competitive with learned evaluation frameworks for the task of text summarization.
We then analyze the effects of factors such as the selection and number of in-context examples on performance.
arXiv Detail & Related papers (2023-06-01T23:27:49Z) - The Grammar and Syntax Based Corpus Analysis Tool For The Ukrainian
Language [0.0]
The StyloMetrix is a tool to analyze grammatical, stylistic, and syntactic patterns in English, Spanish, German, and others.
We describe the StyloMetrix pipeline and provide some experiments with this tool for the text classification task.
We also describe our package's main limitations and the metrics' evaluation procedure.
arXiv Detail & Related papers (2023-05-22T22:52:47Z) - Transformer-based Model for Word Level Language Identification in
Code-mixed Kannada-English Texts [55.41644538483948]
We propose the use of a Transformer based model for word-level language identification in code-mixed Kannada English texts.
The proposed model on the CoLI-Kenglish dataset achieves a weighted F1-score of 0.84 and a macro F1-score of 0.61.
arXiv Detail & Related papers (2022-11-26T02:39:19Z) - A Machine Learning Approach to Classifying Construction Cost Documents
into the International Construction Measurement Standard [0.0]
We introduce the first automated models for classifying natural language descriptions provided in cost documents called "Bills of Quantities"
We learn from a dataset of more than 50 thousand descriptions of items retrieved from 24 large infrastructure construction projects across the United Kingdom.
arXiv Detail & Related papers (2022-10-24T11:35:53Z) - InfoLM: A New Metric to Evaluate Summarization & Data2Text Generation [27.129551973093008]
InfoLM is a family of untrained metrics that can be viewed as a string-based metric.
This family of metrics also makes use of information measures allowing the adaptation of InfoLM to various evaluation criteria.
arXiv Detail & Related papers (2021-12-02T20:09:29Z) - BEAMetrics: A Benchmark for Language Generation Evaluation Evaluation [16.81712151903078]
Natural language processing (NLP) systems are increasingly trained to generate open-ended text.
Different metrics have different strengths and biases, and reflect human intuitions better on some tasks than others.
Here, we describe the Benchmark to Evaluate Automatic Metrics (BEAMetrics) to make research into new metrics itself easier to evaluate.
arXiv Detail & Related papers (2021-10-18T10:03:19Z) - Curious Case of Language Generation Evaluation Metrics: A Cautionary
Tale [52.663117551150954]
A few popular metrics remain as the de facto metrics to evaluate tasks such as image captioning and machine translation.
This is partly due to ease of use, and partly because researchers expect to see them and know how to interpret them.
In this paper, we urge the community for more careful consideration of how they automatically evaluate their models.
arXiv Detail & Related papers (2020-10-26T13:57:20Z) - GO FIGURE: A Meta Evaluation of Factuality in Summarization [131.1087461486504]
We introduce GO FIGURE, a meta-evaluation framework for evaluating factuality evaluation metrics.
Our benchmark analysis on ten factuality metrics reveals that our framework provides a robust and efficient evaluation.
It also reveals that while QA metrics generally improve over standard metrics that measure factuality across domains, performance is highly dependent on the way in which questions are generated.
arXiv Detail & Related papers (2020-10-24T08:30:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.