Error Diversity Matters: An Error-Resistant Ensemble Method for Unsupervised Dependency Parsing
- URL: http://arxiv.org/abs/2412.11543v2
- Date: Thu, 06 Feb 2025 21:00:42 GMT
- Title: Error Diversity Matters: An Error-Resistant Ensemble Method for Unsupervised Dependency Parsing
- Authors: Behzad Shayegh, Hobie H. -B. Lee, Xiaodan Zhu, Jackie Chi Kit Cheung, Lili Mou,
- Abstract summary: We propose an efficient ensemble-selection approach that considers error diversity and avoids error accumulation.
Results demonstrate that our approach outperforms each individual model as well as previous ensemble techniques.
- Score: 49.60415088899208
- License:
- Abstract: We address unsupervised dependency parsing by building an ensemble of diverse existing models through post hoc aggregation of their output dependency parse structures. We observe that these ensembles often suffer from low robustness against weak ensemble components due to error accumulation. To tackle this problem, we propose an efficient ensemble-selection approach that considers error diversity and avoids error accumulation. Results demonstrate that our approach outperforms each individual model as well as previous ensemble techniques. Additionally, our experiments show that the proposed ensemble-selection method significantly enhances the performance and robustness of our ensemble, surpassing previously proposed strategies, which have not accounted for error diversity.
Related papers
- A Unified Comparative Study with Generalized Conformity Scores for Multi-Output Conformal Regression [1.747623282473278]
We present a unified comparative study of nine conformal methods with different multi-output base models.
We also introduce two novel classes of conformity scores for multi-output regression.
One class is compatible with any generative model, while the other is computationally efficient, leveraging the properties of invertible generative models.
arXiv Detail & Related papers (2025-01-17T20:13:24Z) - Bridging Multicalibration and Out-of-distribution Generalization Beyond Covariate Shift [44.708914058803224]
We establish a new model-agnostic optimization framework for out-of-distribution generalization via multicalibration.
We propose MC-Pseudolabel, a post-processing algorithm to achieve both extended multicalibration and out-of-distribution generalization.
arXiv Detail & Related papers (2024-06-02T08:11:35Z) - Task Groupings Regularization: Data-Free Meta-Learning with Heterogeneous Pre-trained Models [83.02797560769285]
Data-Free Meta-Learning (DFML) aims to derive knowledge from a collection of pre-trained models without accessing their original data.
Current methods often overlook the heterogeneity among pre-trained models, which leads to performance degradation due to task conflicts.
arXiv Detail & Related papers (2024-05-26T13:11:55Z) - Exploring Model Learning Heterogeneity for Boosting Ensemble Robustness [17.127312781074245]
Deep neural network ensembles hold the potential of improving generalization performance for complex learning tasks.
This paper presents formal analysis and empirical evaluation of heterogeneous deep ensembles with high ensemble diversity.
arXiv Detail & Related papers (2023-10-03T17:47:25Z) - Ensemble Distillation for Unsupervised Constituency Parsing [40.96887945888518]
We investigate the unsupervised constituency parsing task, which organizes words and phrases of a sentence into a hierarchical structure without using linguistically annotated data.
We propose a notion of "tree averaging," based on which we further propose a novel ensemble method for unsupervised parsing.
arXiv Detail & Related papers (2023-10-03T01:02:44Z) - Joint Training of Deep Ensembles Fails Due to Learner Collusion [61.557412796012535]
Ensembles of machine learning models have been well established as a powerful method of improving performance over a single model.
Traditionally, ensembling algorithms train their base learners independently or sequentially with the goal of optimizing their joint performance.
We show that directly minimizing the loss of the ensemble appears to rarely be applied in practice.
arXiv Detail & Related papers (2023-01-26T18:58:07Z) - Rethinking Clustering-Based Pseudo-Labeling for Unsupervised
Meta-Learning [146.11600461034746]
Method for unsupervised meta-learning, CACTUs, is a clustering-based approach with pseudo-labeling.
This approach is model-agnostic and can be combined with supervised algorithms to learn from unlabeled data.
We prove that the core reason for this is lack of a clustering-friendly property in the embedding space.
arXiv Detail & Related papers (2022-09-27T19:04:36Z) - Examining and Combating Spurious Features under Distribution Shift [94.31956965507085]
We define and analyze robust and spurious representations using the information-theoretic concept of minimal sufficient statistics.
We prove that even when there is only bias of the input distribution, models can still pick up spurious features from their training data.
Inspired by our analysis, we demonstrate that group DRO can fail when groups do not directly account for various spurious correlations.
arXiv Detail & Related papers (2021-06-14T05:39:09Z) - Accounting for Unobserved Confounding in Domain Generalization [107.0464488046289]
This paper investigates the problem of learning robust, generalizable prediction models from a combination of datasets.
Part of the challenge of learning robust models lies in the influence of unobserved confounders.
We demonstrate the empirical performance of our approach on healthcare data from different modalities.
arXiv Detail & Related papers (2020-07-21T08:18:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.