Enhanced Genetic Programming Models with Multiple Equations for Accurate
Semi-Autogenous Grinding Mill Throughput Prediction
- URL: http://arxiv.org/abs/2401.05382v2
- Date: Mon, 29 Jan 2024 02:01:02 GMT
- Title: Enhanced Genetic Programming Models with Multiple Equations for Accurate
Semi-Autogenous Grinding Mill Throughput Prediction
- Authors: Zahra Ghasemi, Mehdi Nesht, Chris Aldrich, John Karageorgos, Max
Zanin, Frank Neumann, Lei Chen
- Abstract summary: This study introduces an enhanced GP approach entitled multi-equation GP (MEGP) for more accurate prediction of SAG mill throughput.
To assess the effect of distance measures, four different distance measures are employed in MEGP method.
- Score: 11.462441722546428
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Semi-autogenous grinding (SAG) mills play a pivotal role in the grinding
circuit of mineral processing plants. Accurate prediction of SAG mill
throughput as a crucial performance metric is of utmost importance. The
potential of applying genetic programming (GP) for this purpose has yet to be
thoroughly investigated. This study introduces an enhanced GP approach entitled
multi-equation GP (MEGP) for more accurate prediction of SAG mill throughput.
In the new proposed method multiple equations, each accurately predicting mill
throughput for specific clusters of training data are extracted. These
equations are then employed to predict mill throughput for test data using
various approaches. To assess the effect of distance measures, four different
distance measures are employed in MEGP method. Comparative analysis reveals
that the best MEGP approach achieves an average improvement of 10.74% in
prediction accuracy compared with standard GP. In this approach, all extracted
equations are utilized and both the number of data points in each data cluster
and the distance to clusters are incorporated for calculating the final
prediction. Further investigation of distance measures indicates that among
four different metrics employed including Euclidean, Manhattan, Chebyshev, and
Cosine distance, the Euclidean distance measure yields the most accurate
results for the majority of data splits.
Related papers
- Computation-Aware Gaussian Processes: Model Selection And Linear-Time Inference [55.150117654242706]
We show that model selection for computation-aware GPs trained on 1.8 million data points can be done within a few hours on a single GPU.
As a result of this work, Gaussian processes can be trained on large-scale datasets without significantly compromising their ability to quantify uncertainty.
arXiv Detail & Related papers (2024-11-01T21:11:48Z) - Aggregation Models with Optimal Weights for Distributed Gaussian Processes [6.408773096179187]
We propose a novel approach for aggregated prediction in distributed GPs.
The proposed method incorporates correlations among experts, leading to better prediction accuracy with manageable computational requirements.
As demonstrated by empirical studies, the proposed approach results in more stable predictions in less time than state-of-the-art consistent aggregation models.
arXiv Detail & Related papers (2024-08-01T23:32:14Z) - Joint Prediction Regions for time-series models [0.0]
It is an easy task to compute Joint Prediction regions (JPR) when the data is IID.
This project aims to implement Wolf and Wunderli's method for constructing JPRs and compare it with other methods.
arXiv Detail & Related papers (2024-05-14T02:38:49Z) - Minimally Supervised Learning using Topological Projections in
Self-Organizing Maps [55.31182147885694]
We introduce a semi-supervised learning approach based on topological projections in self-organizing maps (SOMs)
Our proposed method first trains SOMs on unlabeled data and then a minimal number of available labeled data points are assigned to key best matching units (BMU)
Our results indicate that the proposed minimally supervised model significantly outperforms traditional regression techniques.
arXiv Detail & Related papers (2024-01-12T22:51:48Z) - Iterative Methods for Vecchia-Laplace Approximations for Latent Gaussian Process Models [11.141688859736805]
We introduce and analyze several preconditioners, derive new convergence results, and propose novel methods for accurately approxing predictive variances.
In particular, we obtain a speed-up of an order of magnitude compared to Cholesky-based calculations.
All methods are implemented in a free C++ software library with high-level Python and R packages.
arXiv Detail & Related papers (2023-10-18T14:31:16Z) - Exogenous Data in Forecasting: FARM -- A New Measure for Relevance
Evaluation [62.997667081978825]
We introduce a new approach named FARM - Forward Relevance Aligned Metric.
Our forward method relies on an angular measure that compares changes in subsequent data points to align time-warped series.
As a first validation step, we present the application of our FARM approach to synthetic but representative signals.
arXiv Detail & Related papers (2023-04-21T15:22:33Z) - Gaussian Processes for Missing Value Imputation [0.0]
We present a hierarchical composition of sparse GPs that is used to predict missing values at each dimension using all the variables from the other dimensions.
The approach missing GP (MGP) can be trained simultaneously to impute all observed missing values.
We evaluate MGP in one private clinical data set and four UCI datasets with a different percentage of missing values.
arXiv Detail & Related papers (2022-04-10T10:46:26Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z) - MuyGPs: Scalable Gaussian Process Hyperparameter Estimation Using Local
Cross-Validation [1.2233362977312945]
We present MuyGPs, a novel efficient GP hyper parameter estimation method.
MuyGPs builds upon prior methods that take advantage of the nearest neighbors structure of the data.
We show that our method outperforms all known competitors both in terms of time-to-solution and the root mean squared error of the predictions.
arXiv Detail & Related papers (2021-04-29T18:10:21Z) - Surface Warping Incorporating Machine Learning Assisted Domain
Likelihood Estimation: A New Paradigm in Mine Geology Modelling and
Automation [68.8204255655161]
A Bayesian warping technique has been proposed to reshape modeled surfaces based on geochemical and spatial constraints imposed by newly acquired blasthole data.
This paper focuses on incorporating machine learning in this warping framework to make the likelihood generalizable.
Its foundation is laid by a Bayesian computation in which the geological domain likelihood given the chemistry, p(g|c) plays a similar role to p(y(c)|g.
arXiv Detail & Related papers (2021-02-15T10:37:52Z) - Machine learning for causal inference: on the use of cross-fit
estimators [77.34726150561087]
Doubly-robust cross-fit estimators have been proposed to yield better statistical properties.
We conducted a simulation study to assess the performance of several estimators for the average causal effect (ACE)
When used with machine learning, the doubly-robust cross-fit estimators substantially outperformed all of the other estimators in terms of bias, variance, and confidence interval coverage.
arXiv Detail & Related papers (2020-04-21T23:09:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.