Towards Explainable Khmer Polarity Classification
- URL: http://arxiv.org/abs/2511.09313v1
- Date: Thu, 13 Nov 2025 01:46:00 GMT
- Title: Towards Explainable Khmer Polarity Classification
- Authors: Marry Kong, Rina Buoy, Sovisal Chenda, Nguonly Taing,
- Abstract summary: This paper proposes an explainable Khmer polarity by fine-tuning an instruction-based reasoning Qwen-3 model.<n> Experimental results show that the fine-tuned model not only predicts labels accurately but also provides reasoning by identifying polarity-related keywords.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Khmer polarity classification is a fundamental natural language processing task that assigns a positive, negative, or neutral label to a given Khmer text input. Existing Khmer models typically predict the label without explaining the rationale behind the prediction. This paper proposes an explainable Khmer polarity classifier by fine-tuning an instruction-based reasoning Qwen-3 model. The notion of explainability in this paper is limited to self-explanations, which the model uses to rationalize its predictions. Experimental results show that the fine-tuned model not only predicts labels accurately but also provides reasoning by identifying polarity-related keywords or phrases to support its predictions. In addition, we contribute a new Khmer polarity dataset consisting of short- to medium-length casual, romanized, and mixed-code Khmer expressions. This dataset was constructed using both heuristic rules and human curation and is publicly available through a gated Hugging Face repository (rinabuoy/khmerpolarity_nonreasoning). The fine-tuned Qwen-3 models are also made available in the same Hugging Face account.
Related papers
- Surprise! Uniform Information Density Isn't the Whole Story: Predicting Surprisal Contours in Long-form Discourse [54.08750245737734]
We propose that speakers modulate information rate based on location within a hierarchically-structured model of discourse.
We find that hierarchical predictors are significant predictors of a discourse's information contour and that deeply nested hierarchical predictors are more predictive than shallow ones.
arXiv Detail & Related papers (2024-10-21T14:42:37Z) - On the Proper Treatment of Tokenization in Psycholinguistics [53.960910019072436]
The paper argues that token-level language models should be marginalized into character-level language models before they are used in psycholinguistic studies.<n>We find various focal areas whose surprisal is a better psychometric predictor than the surprisal of the region of interest itself.
arXiv Detail & Related papers (2024-10-03T17:18:03Z) - LinkLogic: A New Method and Benchmark for Explainable Knowledge Graph Predictions [0.5999777817331317]
We present an in-depth exploration of a simple link prediction explanation method we call LinkLogic.
We construct the first-ever link prediction explanation benchmark, based on family structures present in the FB13 dataset.
arXiv Detail & Related papers (2024-06-02T20:22:22Z) - Hashmarks: Privacy-Preserving Benchmarks for High-Stakes AI Evaluation [0.0]
We propose hashmarking, a protocol for evaluating language models in the open without having to disclose the correct answers.
In its simplest form, a hashmark is a benchmark whose reference solutions have been cryptographically hashed prior to publication.
arXiv Detail & Related papers (2023-12-01T15:16:00Z) - Filtering Context Mitigates Scarcity and Selection Bias in Political
Ideology Prediction [42.31457743674423]
We propose a novel supervised learning approach for political ideology prediction (PIP)
We show that our model is capable of outputting predictions even when trained with as little as 5% biased data.
arXiv Detail & Related papers (2023-02-01T04:34:48Z) - Rationalizing Predictions by Adversarial Information Calibration [65.19407304154177]
We train two models jointly: one is a typical neural model that solves the task at hand in an accurate but black-box manner, and the other is a selector-predictor model that additionally produces a rationale for its prediction.
We use an adversarial technique to calibrate the information extracted by the two models such that the difference between them is an indicator of the missed or over-selected features.
arXiv Detail & Related papers (2023-01-15T03:13:09Z) - On the Usefulness of Embeddings, Clusters and Strings for Text Generator
Evaluation [86.19634542434711]
Mauve measures an information-theoretic divergence between two probability distributions over strings.
We show that Mauve was right for the wrong reasons, and that its newly proposed divergence is not necessary for its high performance.
We conclude that -- by encoding syntactic- and coherence-level features of text, while ignoring surface-level features -- such cluster-based substitutes to string distributions may simply be better for evaluating state-of-the-art language generators.
arXiv Detail & Related papers (2022-05-31T17:58:49Z) - Quark: Controllable Text Generation with Reinforced Unlearning [68.07749519374089]
Large-scale language models often learn behaviors that are misaligned with user expectations.
We introduce Quantized Reward Konditioning (Quark), an algorithm for optimizing a reward function that quantifies an (un)wanted property.
For unlearning toxicity, negative sentiment, and repetition, our experiments show that Quark outperforms both strong baselines and state-of-the-art reinforcement learning methods.
arXiv Detail & Related papers (2022-05-26T21:11:51Z) - Using Natural Sentences for Understanding Biases in Language Models [10.604991889372092]
We create a prompt dataset with respect to occupations collected from real-world natural sentences in Wikipedia.
We find bias evaluations are very sensitive to the design choices of template prompts.
We propose using natural sentence prompts for systematic evaluations to step away from design choices that could introduce bias in the observations.
arXiv Detail & Related papers (2022-05-12T18:36:33Z) - Locally Typical Sampling [84.62530743899025]
We show that today's probabilistic language generators fall short when it comes to producing coherent and fluent text.<n>We propose a simple and efficient procedure for enforcing this criterion when generating from probabilistic models.
arXiv Detail & Related papers (2022-02-01T18:58:45Z) - You Can Do Better! If You Elaborate the Reason When Making Prediction [13.658942796267015]
This paper proposes a novel neural predictive framework coupled with large pre-trained language models to make a prediction and generate its corresponding explanation simultaneously.
We conducted a preliminary empirical study on Chinese medical multiple-choice question answering, English natural language inference and commonsense question answering tasks.
The proposed method also achieves improved prediction accuracy on three datasets, which indicates that making predictions can benefit from generating the explanation in the decision process.
arXiv Detail & Related papers (2021-03-27T14:55:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.