Incorporating Behavioral Hypotheses for Query Generation
- URL: http://arxiv.org/abs/2010.02667v1
- Date: Tue, 6 Oct 2020 12:38:02 GMT
- Title: Incorporating Behavioral Hypotheses for Query Generation
- Authors: Ruey-Cheng Chen, Chia-Jung Lee
- Abstract summary: We propose a generic encoder-decoder Transformer framework to aggregate arbitrary hypotheses of choice for query generation.
Our experimental results show that the proposed approach leads to significant improvements on top-$k$ word error rate and Bert F1 Score.
- Score: 3.5873361386290643
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generative neural networks have been shown effective on query suggestion.
Commonly posed as a conditional generation problem, the task aims to leverage
earlier inputs from users in a search session to predict queries that they will
likely issue at a later time. User inputs come in various forms such as
querying and clicking, each of which can imply different semantic signals
channeled through the corresponding behavioral patterns. This paper induces
these behavioral biases as hypotheses for query generation, where a generic
encoder-decoder Transformer framework is presented to aggregate arbitrary
hypotheses of choice. Our experimental results show that the proposed approach
leads to significant improvements on top-$k$ word error rate and Bert F1 Score
compared to a recent BART model.
Related papers
- Towards a Path Dependent Account of Category Fluency [2.66269503676104]
We present evidence towards resolving the disagreement between each account of foraging by reformulating models as sequence generators.
We find category switch predictors do not necessarily produce human-like sequences, in fact the additional biases used by the Hills et al. (2012) model are required to improve generation quality.
arXiv Detail & Related papers (2024-05-09T16:36:56Z) - Source-Free Unsupervised Domain Adaptation with Hypothesis Consolidation
of Prediction Rationale [53.152460508207184]
Source-Free Unsupervised Domain Adaptation (SFUDA) is a challenging task where a model needs to be adapted to a new domain without access to target domain labels or source domain data.
This paper proposes a novel approach that considers multiple prediction hypotheses for each sample and investigates the rationale behind each hypothesis.
To achieve the optimal performance, we propose a three-step adaptation process: model pre-adaptation, hypothesis consolidation, and semi-supervised learning.
arXiv Detail & Related papers (2024-02-02T05:53:22Z) - Semantic Equivalence of e-Commerce Queries [6.232692545488813]
This paper introduces a framework to recognize and leverage query equivalence to enhance searcher and business outcomes.
The proposed approach addresses three key problems: mapping queries to vector representations of search intent, identifying nearest neighbor queries expressing equivalent or similar intent, and optimizing for user or business objectives.
arXiv Detail & Related papers (2023-08-07T18:40:13Z) - EEL: Efficiently Encoding Lattices for Reranking [44.77383151122229]
We use Transformers to efficiently encode lattices of generated outputs.
We combine this approach with a new class of token-factored rerankers (TFRs)
Our results show both substantial speedup compared to naive reranking and often better performance on downstream metrics than comparable approaches.
arXiv Detail & Related papers (2023-06-01T17:45:32Z) - Predictive Querying for Autoregressive Neural Sequence Models [23.85426261235507]
We introduce a general typology for predictive queries in neural autoregressive sequence models.
We show that such queries can be systematically represented by sets of elementary building blocks.
We leverage this typology to develop new query estimation methods.
arXiv Detail & Related papers (2022-10-12T17:59:36Z) - Paragraph-based Transformer Pre-training for Multi-Sentence Inference [99.59693674455582]
We show that popular pre-trained transformers perform poorly when used for fine-tuning on multi-candidate inference tasks.
We then propose a new pre-training objective that models the paragraph-level semantics across multiple input sentences.
arXiv Detail & Related papers (2022-05-02T21:41:14Z) - Generalization in the Face of Adaptivity: A Bayesian Perspective [3.0202264016476623]
Repeated use of a data sample via adaptively chosen queries can rapidly lead to overfitting.
It turns out that simple noise unbounded addition algorithms suffice to prevent this issue.
We show that the harm of adaptivity results from the covariance between the new query and a Bayes factor-based measure of how much information about the data sample was encoded in the responses given to past queries.
arXiv Detail & Related papers (2021-06-20T22:06:44Z) - Double Perturbation: On the Robustness of Robustness and Counterfactual
Bias Evaluation [109.06060143938052]
We propose a "double perturbation" framework to uncover model weaknesses beyond the test dataset.
We apply this framework to study two perturbation-based approaches that are used to analyze models' robustness and counterfactual bias in English.
arXiv Detail & Related papers (2021-04-12T06:57:36Z) - On the Reproducibility of Neural Network Predictions [52.47827424679645]
We study the problem of churn, identify factors that cause it, and propose two simple means of mitigating it.
We first demonstrate that churn is indeed an issue, even for standard image classification tasks.
We propose using emphminimum entropy regularizers to increase prediction confidences.
We present empirical results showing the effectiveness of both techniques in reducing churn while improving the accuracy of the underlying model.
arXiv Detail & Related papers (2021-02-05T18:51:01Z) - Toward Scalable and Unified Example-based Explanation and Outlier
Detection [128.23117182137418]
We argue for a broader adoption of prototype-based student networks capable of providing an example-based explanation for their prediction.
We show that our prototype-based networks beyond similarity kernels deliver meaningful explanations and promising outlier detection results without compromising classification accuracy.
arXiv Detail & Related papers (2020-11-11T05:58:17Z) - Pre-training Is (Almost) All You Need: An Application to Commonsense
Reasoning [61.32992639292889]
Fine-tuning of pre-trained transformer models has become the standard approach for solving common NLP tasks.
We introduce a new scoring method that casts a plausibility ranking task in a full-text format.
We show that our method provides a much more stable training phase across random restarts.
arXiv Detail & Related papers (2020-04-29T10:54:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.