Domain Divergences: a Survey and Empirical Analysis
- URL: http://arxiv.org/abs/2010.12198v2
- Date: Mon, 19 Apr 2021 06:47:30 GMT
- Title: Domain Divergences: a Survey and Empirical Analysis
- Authors: Abhinav Ramesh Kashyap, Devamanyu Hazarika, Min-Yen Kan, Roger
Zimmermann
- Abstract summary: We develop a taxonomy of divergence measures consisting of three classes -- Information-theoretic, Geometric, and Higher-order measures.
We perform correlation analysis spanning 130 domain adaptation scenarios, 3 varied NLP tasks and 12 divergence measures identified from our survey.
- Score: 47.535524183965464
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Domain divergence plays a significant role in estimating the performance of a
model in new domains. While there is a significant literature on divergence
measures, researchers find it hard to choose an appropriate divergence for a
given NLP application. We address this shortcoming by both surveying the
literature and through an empirical study. We develop a taxonomy of divergence
measures consisting of three classes -- Information-theoretic, Geometric, and
Higher-order measures and identify the relationships between them. Further, to
understand the common use-cases of these measures, we recognise three novel
applications -- 1) Data Selection, 2) Learning Representation, and 3) Decisions
in the Wild -- and use it to organise our literature. From this, we identify
that Information-theoretic measures are prevalent for 1) and 3), and
Higher-order measures are more common for 2). To further help researchers
choose appropriate measures to predict drop in performance -- an important
aspect of Decisions in the Wild, we perform correlation analysis spanning 130
domain adaptation scenarios, 3 varied NLP tasks and 12 divergence measures
identified from our survey. To calculate these divergences, we consider the
current contextual word representations (CWR) and contrast with the older
distributed representations. We find that traditional measures over word
distributions still serve as strong baselines, while higher-order measures with
CWR are effective.
Related papers
- Measuring publication relatedness using controlled vocabularies [0.0]
Controlled vocabularies provide a promising basis for measuring relatedness.
There exists no comprehensive and direct test of their accuracy and suitability for different types of research questions.
This paper reviews existing measures, develops a new measure, and benchmarks the measures using TREC Genomics data as a ground truth of topics.
arXiv Detail & Related papers (2024-08-27T12:41:37Z) - Chain-of-Thought Prompting for Demographic Inference with Large Multimodal Models [58.58594658683919]
Large multimodal models (LMMs) have shown transformative potential across various research tasks.
Our findings indicate LMMs possess advantages in zero-shot learning, interpretability, and handling uncurated 'in-the-wild' inputs.
We propose a Chain-of-Thought augmented prompting approach, which effectively mitigates the off-target prediction issue.
arXiv Detail & Related papers (2024-05-24T16:26:56Z) - Statistical Correlators and Tripartite Entanglement [0.07812210699650153]
We show that two measures can be empirically determined for the two important classes of tripartite entangled states.
Such a formulated scheme would provide for the first time the means to exactly quantify tripartite entanglement.
arXiv Detail & Related papers (2023-08-30T18:00:07Z) - Goodhart's Law Applies to NLP's Explanation Benchmarks [57.26445915212884]
We critically examine two sets of metrics: the ERASER metrics (comprehensiveness and sufficiency) and the EVAL-X metrics.
We show that we can inflate a model's comprehensiveness and sufficiency scores dramatically without altering its predictions or explanations on in-distribution test inputs.
Our results raise doubts about the ability of current metrics to guide explainability research, underscoring the need for a broader reassessment of what precisely these metrics are intended to capture.
arXiv Detail & Related papers (2023-08-28T03:03:03Z) - A Diachronic Analysis of Paradigm Shifts in NLP Research: When, How, and
Why? [84.46288849132634]
We propose a systematic framework for analyzing the evolution of research topics in a scientific field using causal discovery and inference techniques.
We define three variables to encompass diverse facets of the evolution of research topics within NLP.
We utilize a causal discovery algorithm to unveil the causal connections among these variables using observational data.
arXiv Detail & Related papers (2023-05-22T11:08:00Z) - Robust Saliency-Aware Distillation for Few-shot Fine-grained Visual
Recognition [57.08108545219043]
Recognizing novel sub-categories with scarce samples is an essential and challenging research topic in computer vision.
Existing literature addresses this challenge by employing local-based representation approaches.
This article proposes a novel model, Robust Saliency-aware Distillation (RSaD), for few-shot fine-grained visual recognition.
arXiv Detail & Related papers (2023-05-12T00:13:17Z) - Explaining Algorithmic Fairness Through Fairness-Aware Causal Path
Decomposition [37.823248189626014]
We propose to study the problem of identification of the source of model disparities.
Unlike existing interpretation methods which typically learn feature importance, we consider the causal relationships among feature variables.
Our framework is also model agnostic and applicable to a variety of quantitative disparity measures.
arXiv Detail & Related papers (2021-08-11T17:23:47Z) - Characterizing and comparing external measures for the assessment of
cluster analysis and community detection [1.5543116359698947]
Many external evaluation measures have been proposed in the literature to compare two partitions of the same set.
This makes the task of selecting the most appropriate measure for a given situation a challenge for the end user.
We propose a new empirical evaluation framework to solve this issue, and help the end user selecting an appropriate measure for their application.
arXiv Detail & Related papers (2021-02-01T09:10:25Z) - Metric-Learning-Assisted Domain Adaptation [18.62119154143642]
Many existing domain alignment methods assume that a low source risk, together with the alignment of distributions of source and target, means a low target risk.
We propose a novel metric-learning-assisted domain adaptation (MLA-DA) method, which employs a novel triplet loss for helping better feature alignment.
arXiv Detail & Related papers (2020-04-23T04:20:02Z) - A Survey on Causal Inference [64.45536158710014]
Causal inference is a critical research topic across many domains, such as statistics, computer science, education, public policy and economics.
Various causal effect estimation methods for observational data have sprung up.
arXiv Detail & Related papers (2020-02-05T21:35:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.