Synthetic Model Combination: An Instance-wise Approach to Unsupervised
Ensemble Learning
- URL: http://arxiv.org/abs/2210.05320v1
- Date: Tue, 11 Oct 2022 10:20:31 GMT
- Title: Synthetic Model Combination: An Instance-wise Approach to Unsupervised
Ensemble Learning
- Authors: Alex J. Chan and Mihaela van der Schaar
- Abstract summary: Consider making a prediction over new test data without any opportunity to learn from a training set of labelled data.
Give access to a set of expert models and their predictions alongside some limited information about the dataset used to train them.
- Score: 92.89846887298852
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Consider making a prediction over new test data without any opportunity to
learn from a training set of labelled data - instead given access to a set of
expert models and their predictions alongside some limited information about
the dataset used to train them. In scenarios from finance to the medical
sciences, and even consumer practice, stakeholders have developed models on
private data they either cannot, or do not want to, share. Given the value and
legislation surrounding personal information, it is not surprising that only
the models, and not the data, will be released - the pertinent question
becoming: how best to use these models? Previous work has focused on global
model selection or ensembling, with the result of a single final model across
the feature space. Machine learning models perform notoriously poorly on data
outside their training domain however, and so we argue that when ensembling
models the weightings for individual instances must reflect their respective
domains - in other words models that are more likely to have seen information
on that instance should have more attention paid to them. We introduce a method
for such an instance-wise ensembling of models, including a novel
representation learning step for handling sparse high-dimensional domains.
Finally, we demonstrate the need and generalisability of our method on
classical machine learning tasks as well as highlighting a real world use case
in the pharmacological setting of vancomycin precision dosing.
Related papers
- Addressing Bias Through Ensemble Learning and Regularized Fine-Tuning [0.2812395851874055]
This paper proposes a comprehensive approach using multiple methods to remove bias in AI models.
We train multiple models with the counter-bias of the pre-trained model through data splitting, local training, and regularized fine-tuning.
We conclude our solution with knowledge distillation that results in a single unbiased neural network.
arXiv Detail & Related papers (2024-02-01T09:24:36Z) - Learning Defect Prediction from Unrealistic Data [57.53586547895278]
Pretrained models of code have become popular choices for code understanding and generation tasks.
Such models tend to be large and require commensurate volumes of training data.
It has become popular to train models with far larger but less realistic datasets, such as functions with artificially injected bugs.
Models trained on such data tend to only perform well on similar data, while underperforming on real world programs.
arXiv Detail & Related papers (2023-11-02T01:51:43Z) - Fantastic Gains and Where to Find Them: On the Existence and Prospect of
General Knowledge Transfer between Any Pretrained Model [74.62272538148245]
We show that for arbitrary pairings of pretrained models, one model extracts significant data context unavailable in the other.
We investigate if it is possible to transfer such "complementary" knowledge from one model to another without performance degradation.
arXiv Detail & Related papers (2023-10-26T17:59:46Z) - On Inductive Biases for Machine Learning in Data Constrained Settings [0.0]
This thesis explores a different answer to the problem of learning expressive models in data constrained settings.
Instead of relying on big datasets to learn neural networks, we will replace some modules by known functions reflecting the structure of the data.
Our approach falls under the hood of "inductive biases", which can be defined as hypothesis on the data at hand restricting the space of models to explore.
arXiv Detail & Related papers (2023-02-21T14:22:01Z) - Dataless Knowledge Fusion by Merging Weights of Language Models [51.8162883997512]
Fine-tuning pre-trained language models has become the prevalent paradigm for building downstream NLP models.
This creates a barrier to fusing knowledge across individual models to yield a better single model.
We propose a dataless knowledge fusion method that merges models in their parameter space.
arXiv Detail & Related papers (2022-12-19T20:46:43Z) - Revealing Secrets From Pre-trained Models [2.0249686991196123]
Transfer-learning has been widely adopted in many emerging deep learning algorithms.
We show that pre-trained models and fine-tuned models have significantly high similarities in weight values.
We propose a new model extraction attack that reveals the model architecture and the pre-trained model used by the black-box victim model.
arXiv Detail & Related papers (2022-07-19T20:19:03Z) - Forecasting COVID-19 spreading trough an ensemble of classical and
machine learning models: Spain's case study [0.0]
We evaluate the applicability of an ensemble of population models and machine learning models to predict the near future evolution of the COVID-19 pandemic.
We rely solely in open and public datasets, fusing incidence, vaccination, human mobility and weather data.
arXiv Detail & Related papers (2022-07-12T08:16:44Z) - On the Efficacy of Adversarial Data Collection for Question Answering:
Results from a Large-Scale Randomized Study [65.17429512679695]
In adversarial data collection (ADC), a human workforce interacts with a model in real time, attempting to produce examples that elicit incorrect predictions.
Despite ADC's intuitive appeal, it remains unclear when training on adversarial datasets produces more robust models.
arXiv Detail & Related papers (2021-06-02T00:48:33Z) - Learning from others' mistakes: Avoiding dataset biases without modeling
them [111.17078939377313]
State-of-the-art natural language processing (NLP) models often learn to model dataset biases and surface form correlations instead of features that target the intended task.
Previous work has demonstrated effective methods to circumvent these issues when knowledge of the bias is available.
We show a method for training models that learn to ignore these problematic correlations.
arXiv Detail & Related papers (2020-12-02T16:10:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.