Enabling Reproducibility and Meta-learning Through a Lifelong Database
of Experiments (LDE)
- URL: http://arxiv.org/abs/2202.10979v2
- Date: Wed, 23 Feb 2022 18:26:42 GMT
- Title: Enabling Reproducibility and Meta-learning Through a Lifelong Database
of Experiments (LDE)
- Authors: Jason Tsay, Andrea Bartezzaghi, Aleke Nolte, Cristiano Malossi
- Abstract summary: We present the Lifelong Database of Experiments (LDE) that automatically extracts and stores linked metadata from experiment artifacts.
We store context from multiple stages of the AI development lifecycle including datasets, pipelines, how each is configured, and training runs with information about their runtime environment.
We perform two experiments on this metadata: 1) examining the variability of the performance metrics and 2) implementing a number of meta-learning algorithms on top of the data.
- Score: 0.43012765978447565
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Artificial Intelligence (AI) development is inherently iterative and
experimental. Over the course of normal development, especially with the advent
of automated AI, hundreds or thousands of experiments are generated and are
often lost or never examined again. There is a lost opportunity to document
these experiments and learn from them at scale, but the complexity of tracking
and reproducing these experiments is often prohibitive to data scientists. We
present the Lifelong Database of Experiments (LDE) that automatically extracts
and stores linked metadata from experiment artifacts and provides features to
reproduce these artifacts and perform meta-learning across them. We store
context from multiple stages of the AI development lifecycle including
datasets, pipelines, how each is configured, and training runs with information
about their runtime environment. The standardized nature of the stored metadata
allows for querying and aggregation, especially in terms of ranking artifacts
by performance metrics. We exhibit the capabilities of the LDE by reproducing
an existing meta-learning study and storing the reproduced metadata in our
system. Then, we perform two experiments on this metadata: 1) examining the
reproducibility and variability of the performance metrics and 2) implementing
a number of meta-learning algorithms on top of the data and examining how
variability in experimental results impacts recommendation performance. The
experimental results suggest significant variation in performance, especially
depending on dataset configurations; this variation carries over when
meta-learning is built on top of the results, with performance improving when
using aggregated results. This suggests that a system that automatically
collects and aggregates results such as the LDE not only assists in
implementing meta-learning but may also improve its performance.
Related papers
- A Dataset For Computational Reproducibility [2.147712260420443]
This article introduces a dataset of computational experiments covering a broad spectrum of scientific fields.
It incorporates details about software dependencies, execution steps, and configurations necessary for accurate reproduction.
It provides a universal benchmark by establishing a standardized dataset for objectively evaluating and comparing the effectiveness of tools.
arXiv Detail & Related papers (2025-04-11T16:45:10Z) - Optimizing Instruction Synthesis: Effective Exploration of Evolutionary Space with Tree Search [25.108044778194536]
We introduce IDEA-MCTS (Instruction Data Enhancement using Monte Carlo Tree Search), a scalable framework for efficiently synthesizing instructions.
With tree search and evaluation models, it can efficiently guide each instruction to evolve into a high-quality form, aiding in instruction fine-tuning.
Experimental results show that IDEA-MCTS significantly enhances the seed instruction data, raising the average evaluation scores of quality, diversity, and complexity from 2.19 to 3.81.
arXiv Detail & Related papers (2024-10-14T11:28:30Z) - Efficacy of Synthetic Data as a Benchmark [3.2968976262860408]
We investigate the effectiveness of generating synthetic data through large language models (LLMs)
Our experiments show that while synthetic data can effectively capture performance of various methods for simpler tasks, it falls short for more complex tasks like named entity recognition.
We propose a new metric called the bias factor, which evaluates the biases introduced when the same LLM is used to both generate benchmarking data and to perform the tasks.
arXiv Detail & Related papers (2024-09-18T13:20:23Z) - Hacking Task Confounder in Meta-Learning [18.179340061914708]
We propose a plug-and-play Meta-learning Causal Representation (MetaCRL) to eliminate task confounders.
Our work achieves state-of-the-art (SOTA) performance on benchmark datasets.
arXiv Detail & Related papers (2023-12-10T05:33:40Z) - D3A-TS: Denoising-Driven Data Augmentation in Time Series [0.0]
This work focuses on studying and analyzing the use of different techniques for data augmentation in time series for classification and regression problems.
The proposed approach involves the use of diffusion probabilistic models, which have recently achieved successful results in the field of Image Processing.
The results highlight the high utility of this methodology in creating synthetic data to train classification and regression models.
arXiv Detail & Related papers (2023-12-09T11:37:07Z) - Robust Learning with Progressive Data Expansion Against Spurious
Correlation [65.83104529677234]
We study the learning process of a two-layer nonlinear convolutional neural network in the presence of spurious features.
Our analysis suggests that imbalanced data groups and easily learnable spurious features can lead to the dominance of spurious features during the learning process.
We propose a new training algorithm called PDE that efficiently enhances the model's robustness for a better worst-group performance.
arXiv Detail & Related papers (2023-06-08T05:44:06Z) - The NCI Imaging Data Commons as a platform for reproducible research in
computational pathology [0.0773931605896092]
Reproducibility is a major challenge in developing machine learning (ML)-based solutions in computational pathology (CompPath)
The NCI Imaging Data Commons (IDC) provides >120 cancer image collections according to the FAIR principles and is designed to be used with cloud ML services.
We implement two experiments in which a representative ML-based method for classifying lung tumor tissue was trained and/or evaluated on different datasets.
arXiv Detail & Related papers (2023-03-16T14:32:50Z) - Automatic Data Augmentation via Invariance-Constrained Learning [94.27081585149836]
Underlying data structures are often exploited to improve the solution of learning tasks.
Data augmentation induces these symmetries during training by applying multiple transformations to the input data.
This work tackles these issues by automatically adapting the data augmentation while solving the learning task.
arXiv Detail & Related papers (2022-09-29T18:11:01Z) - TRoVE: Transforming Road Scene Datasets into Photorealistic Virtual
Environments [84.6017003787244]
This work proposes a synthetic data generation pipeline to address the difficulties and domain-gaps present in simulated datasets.
We show that using annotations and visual cues from existing datasets, we can facilitate automated multi-modal data generation.
arXiv Detail & Related papers (2022-08-16T20:46:08Z) - Multi-Domain Joint Training for Person Re-Identification [51.73921349603597]
Deep learning-based person Re-IDentification (ReID) often requires a large amount of training data to achieve good performance.
It appears that collecting more training data from diverse environments tends to improve the ReID performance.
We propose an approach called Domain-Camera-Sample Dynamic network (DCSD) whose parameters can be adaptive to various factors.
arXiv Detail & Related papers (2022-01-06T09:20:59Z) - Unsupervised Domain Adaptive Learning via Synthetic Data for Person
Re-identification [101.1886788396803]
Person re-identification (re-ID) has gained more and more attention due to its widespread applications in video surveillance.
Unfortunately, the mainstream deep learning methods still need a large quantity of labeled data to train models.
In this paper, we develop a data collector to automatically generate synthetic re-ID samples in a computer game, and construct a data labeler to simultaneously annotate them.
arXiv Detail & Related papers (2021-09-12T15:51:41Z) - Shared Space Transfer Learning for analyzing multi-site fMRI data [83.41324371491774]
Multi-voxel pattern analysis (MVPA) learns predictive models from task-based functional magnetic resonance imaging (fMRI) data.
MVPA works best with a well-designed feature set and an adequate sample size.
Most fMRI datasets are noisy, high-dimensional, expensive to collect, and with small sample sizes.
This paper proposes the Shared Space Transfer Learning (SSTL) as a novel transfer learning approach.
arXiv Detail & Related papers (2020-10-24T08:50:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.