Implementing a Logical Inference System for Japanese Comparatives
- URL: http://arxiv.org/abs/2509.13734v1
- Date: Wed, 17 Sep 2025 06:37:10 GMT
- Title: Implementing a Logical Inference System for Japanese Comparatives
- Authors: Yosuke Mikami, Daiki Matsuoka, Hitomi Yanaka,
- Abstract summary: This study proposes ccg-jcomp, a logical inference system for Japanese comparatives based on compositional semantics.<n>We evaluate the proposed system on a Japanese NLI dataset containing comparative expressions.
- Score: 15.852779398905957
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Natural Language Inference (NLI) involving comparatives is challenging because it requires understanding quantities and comparative relations expressed by sentences. While some approaches leverage Large Language Models (LLMs), we focus on logic-based approaches grounded in compositional semantics, which are promising for robust handling of numerical and logical expressions. Previous studies along these lines have proposed logical inference systems for English comparatives. However, it has been pointed out that there are several morphological and semantic differences between Japanese and English comparatives. These differences make it difficult to apply such systems directly to Japanese comparatives. To address this gap, this study proposes ccg-jcomp, a logical inference system for Japanese comparatives based on compositional semantics. We evaluate the proposed system on a Japanese NLI dataset containing comparative expressions. We demonstrate the effectiveness of our system by comparing its accuracy with that of existing LLMs.
Related papers
- DivLogicEval: A Framework for Benchmarking Logical Reasoning Evaluation in Large Language Models [58.439517684779936]
This paper proposes a new classical logic benchmark DivLogicEval, consisting of natural sentences composed of diverse statements in a counterintuitive way.<n>To ensure a more reliable evaluation, we also introduce a new evaluation metric that mitigates the influence of bias and randomness inherent in Large Language Models.
arXiv Detail & Related papers (2025-09-19T04:40:46Z) - Can Large Language Models Robustly Perform Natural Language Inference for Japanese Comparatives? [15.852779398905957]
Large Language Models (LLMs) perform remarkably well in Natural Language Inference (NLI)<n>This paper focuses on comparatives and evaluate various LLMs in zero-shot and few-shot settings.<n>We observe that prompts containing logical semantic representations help the models predict the correct labels for inference problems that they struggle to solve even with few-shot examples.
arXiv Detail & Related papers (2025-09-17T04:56:51Z) - Disparities in LLM Reasoning Accuracy and Explanations: A Case Study on African American English [66.97110551643722]
We investigate dialectal disparities in Large Language Models (LLMs) reasoning tasks.<n>We find that LLMs produce less accurate responses and simpler reasoning chains and explanations for AAE inputs.<n>These findings highlight systematic differences in how LLMs process and reason about different language varieties.
arXiv Detail & Related papers (2025-03-06T05:15:34Z) - Predicting Text Preference Via Structured Comparative Reasoning [110.49560164568791]
We introduce SC, a prompting approach that predicts text preferences by generating structured intermediate comparisons.
We select consistent comparisons with a pairwise consistency comparator that ensures each aspect's comparisons clearly distinguish differences between texts.
Our comprehensive evaluations across various NLP tasks, including summarization, retrieval, and automatic rating, demonstrate that SC equips LLMs to achieve state-of-the-art performance in text preference prediction.
arXiv Detail & Related papers (2023-11-14T18:51:38Z) - LLM Comparative Assessment: Zero-shot NLG Evaluation through Pairwise
Comparisons using Large Language Models [55.60306377044225]
Large language models (LLMs) have enabled impressive zero-shot capabilities across various natural language tasks.
This paper explores two options for exploiting the emergent abilities of LLMs for zero-shot NLG assessment.
For moderate-sized open-source LLMs, such as FlanT5 and Llama2-chat, comparative assessment is superior to prompt scoring.
arXiv Detail & Related papers (2023-07-15T22:02:12Z) - RankCSE: Unsupervised Sentence Representations Learning via Learning to
Rank [54.854714257687334]
We propose a novel approach, RankCSE, for unsupervised sentence representation learning.
It incorporates ranking consistency and ranking distillation with contrastive learning into a unified framework.
An extensive set of experiments are conducted on both semantic textual similarity (STS) and transfer (TR) tasks.
arXiv Detail & Related papers (2023-05-26T08:27:07Z) - Pre-training Language Models for Comparative Reasoning [26.161185103553635]
We propose a novel framework to pre-train language models for enhancing their abilities of comparative reasoning over texts.
Our approach introduces a novel method of collecting scalable data for text-based entity comparison.
We present a framework of pre-training language models via three novel objectives on comparative reasoning.
arXiv Detail & Related papers (2023-05-23T18:28:42Z) - Combining Event Semantics and Degree Semantics for Natural Language
Inference [16.536018920603176]
We implement a logic-based NLI system that combines event semantics and degree semantics and their interaction with lexical knowledge.
We evaluate the system on various NLI datasets containing linguistically challenging problems.
arXiv Detail & Related papers (2020-11-02T13:27:21Z) - Modeling Voting for System Combination in Machine Translation [92.09572642019145]
We propose an approach to modeling voting for system combination in machine translation.
Our approach combines the advantages of statistical and neural methods since it can not only analyze the relations between hypotheses but also allow for end-to-end training.
arXiv Detail & Related papers (2020-07-14T09:59:38Z) - Logical Inferences with Comparatives and Generalized Quantifiers [18.58482811176484]
A logical inference system for comparatives has not been sufficiently developed for use in the Natural Language Inference task.
We present a compositional semantics that maps various comparative constructions in English to semantic representations via Category Grammar (CCG)
We show that the system outperforms previous logic-based systems as well as recent deep learning-based models.
arXiv Detail & Related papers (2020-05-16T11:11:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.