Data-Centric Machine Learning in Quantum Information Science
- URL: http://arxiv.org/abs/2201.09134v1
- Date: Sat, 22 Jan 2022 21:36:54 GMT
- Title: Data-Centric Machine Learning in Quantum Information Science
- Authors: Sanjaya Lohani, Joseph M. Lukens, Ryan T. Glasser, Thomas A. Searles,
Brian T. Kirby
- Abstract summary: In particular, we consider how systematic engineering of training sets can significantly enhance the accuracy of pre-trained neural networks.
We show that it is not always optimal to engineer training sets to exactly match the expected distribution of a target scenario, and instead, performance can be further improved by biasing the training set to be slightly more mixed than the target.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose a series of data-centric heuristics for improving the performance
of machine learning systems when applied to problems in quantum information
science. In particular, we consider how systematic engineering of training sets
can significantly enhance the accuracy of pre-trained neural networks used for
quantum state reconstruction without altering the underlying architecture. We
find that it is not always optimal to engineer training sets to exactly match
the expected distribution of a target scenario, and instead, performance can be
further improved by biasing the training set to be slightly more mixed than the
target. This is due to the heterogeneity in the number of free variables
required to describe states of different purity, and as a result, overall
accuracy of the network improves when training sets of a fixed size focus on
states with the least constrained free variables. For further clarity, we also
include a "toy model" demonstration of how spurious correlations can
inadvertently enter synthetic data sets used for training, how the performance
of systems trained with these correlations can degrade dramatically, and how
the inclusion of even relatively few counterexamples can effectively remedy
such problems.
Related papers
- Transferable Post-training via Inverse Value Learning [83.75002867411263]
We propose modeling changes at the logits level during post-training using a separate neural network (i.e., the value network)
After training this network on a small base model using demonstrations, this network can be seamlessly integrated with other pre-trained models during inference.
We demonstrate that the resulting value network has broad transferability across pre-trained models of different parameter sizes.
arXiv Detail & Related papers (2024-10-28T13:48:43Z) - Sobolev neural network with residual weighting as a surrogate in linear and non-linear mechanics [0.0]
This paper investigates the improvement of the training process by including sensitivity information.
In computational mechanics, sensitivities can be applied to neural networks by expanding the training loss function.
This improvement is shown in two examples of linear and non-linear material behavior.
arXiv Detail & Related papers (2024-07-23T13:28:07Z) - Optimal design of experiments in the context of machine-learning inter-atomic potentials: improving the efficiency and transferability of kernel based methods [0.7234862895932991]
Data-driven, machine learning (ML) models of atomistic interactions can relate nuanced aspects of atomic arrangements into predictions of energies and forces.
The main challenge stems from the fact that descriptors of chemical environments are often sparse high-dimensional objects without a well-defined continuous metric.
We will demonstrate that classical concepts of statistical planning of experiments and optimal design can help to mitigate such problems at a relatively low computational cost.
arXiv Detail & Related papers (2024-05-14T14:14:23Z) - A didactic approach to quantum machine learning with a single qubit [68.8204255655161]
We focus on the case of learning with a single qubit, using data re-uploading techniques.
We implement the different proposed formulations in toy and real-world datasets using the qiskit quantum computing SDK.
arXiv Detail & Related papers (2022-11-23T18:25:32Z) - Efficient Augmentation for Imbalanced Deep Learning [8.38844520504124]
We study a convolutional neural network's internal representation of imbalanced image data.
We measure the generalization gap between a model's feature embeddings in the training and test sets, showing that the gap is wider for minority classes.
This insight enables us to design an efficient three-phase CNN training framework for imbalanced data.
arXiv Detail & Related papers (2022-07-13T09:43:17Z) - ClusterQ: Semantic Feature Distribution Alignment for Data-Free
Quantization [111.12063632743013]
We propose a new and effective data-free quantization method termed ClusterQ.
To obtain high inter-class separability of semantic features, we cluster and align the feature distribution statistics.
We also incorporate the intra-class variance to solve class-wise mode collapse.
arXiv Detail & Related papers (2022-04-30T06:58:56Z) - CMW-Net: Learning a Class-Aware Sample Weighting Mapping for Robust Deep
Learning [55.733193075728096]
Modern deep neural networks can easily overfit to biased training data containing corrupted labels or class imbalance.
Sample re-weighting methods are popularly used to alleviate this data bias issue.
We propose a meta-model capable of adaptively learning an explicit weighting scheme directly from data.
arXiv Detail & Related papers (2022-02-11T13:49:51Z) - Deep invariant networks with differentiable augmentation layers [87.22033101185201]
Methods for learning data augmentation policies require held-out data and are based on bilevel optimization problems.
We show that our approach is easier and faster to train than modern automatic data augmentation techniques.
arXiv Detail & Related papers (2022-02-04T14:12:31Z) - Benchmarking quantum tomography completeness and fidelity with machine
learning [0.0]
We train convolutional neural networks to predict whether or not a set of measurements is informationally complete to uniquely reconstruct any given quantum state with no prior information.
Networks are trained to recognize the fidelity and a reliable measure for informational completeness.
arXiv Detail & Related papers (2021-03-02T07:30:32Z) - Using Data Assimilation to Train a Hybrid Forecast System that Combines
Machine-Learning and Knowledge-Based Components [52.77024349608834]
We consider the problem of data-assisted forecasting of chaotic dynamical systems when the available data is noisy partial measurements.
We show that by using partial measurements of the state of the dynamical system, we can train a machine learning model to improve predictions made by an imperfect knowledge-based model.
arXiv Detail & Related papers (2021-02-15T19:56:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.