Heuristic Strategies for Solving Complex Interacting Stockpile Blending
Problem with Chance Constraints
- URL: http://arxiv.org/abs/2102.05303v1
- Date: Wed, 10 Feb 2021 07:56:18 GMT
- Title: Heuristic Strategies for Solving Complex Interacting Stockpile Blending
Problem with Chance Constraints
- Authors: Yue Xie, Aneta Neumann, Frank Neumann
- Abstract summary: In this paper, we consider the uncertainty in material grades and introduce chance constraints that are used to ensure the constraints with high confidence.
To address the stockpile blending problem with chance constraints, we propose a differential evolution algorithm combining two repair operators.
- Score: 14.352521012951865
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Heuristic algorithms have shown a good ability to solve a variety of
optimization problems. Stockpile blending problem as an important component of
the mine scheduling problem is an optimization problem with continuous search
space containing uncertainty in the geologic input data. The objective of the
optimization process is to maximize the total volume of materials of the
operation and subject to resource capacities, chemical processes, and customer
requirements. In this paper, we consider the uncertainty in material grades and
introduce chance constraints that are used to ensure the constraints with high
confidence. To address the stockpile blending problem with chance constraints,
we propose a differential evolution algorithm combining two repair operators
that are used to tackle the two complex constraints. In the experiment section,
we compare the performance of the approach with the deterministic model and
stochastic models by considering different chance constraints and evaluate the
effectiveness of different chance constraints.
Related papers
- Differentiation of Multi-objective Data-driven Decision Pipeline [34.577809430781144]
Real-world scenarios frequently involve multi-objective data-driven optimization problems.
Traditional two-stage methods apply a machine learning model to estimate problem coefficients, followed by invoking a solver to tackle the predicted optimization problem.
Recent efforts have focused on end-to-end training of predictive models that use decision loss derived from the downstream optimization problem.
arXiv Detail & Related papers (2024-06-02T15:42:03Z) - End-to-End Learning for Fair Multiobjective Optimization Under
Uncertainty [55.04219793298687]
The Predict-Then-Forecast (PtO) paradigm in machine learning aims to maximize downstream decision quality.
This paper extends the PtO methodology to optimization problems with nondifferentiable Ordered Weighted Averaging (OWA) objectives.
It shows how optimization of OWA functions can be effectively integrated with parametric prediction for fair and robust optimization under uncertainty.
arXiv Detail & Related papers (2024-02-12T16:33:35Z) - Optimizing Chance-Constrained Submodular Problems with Variable
Uncertainties [12.095075636344536]
We study chance-constrained submodular optimization problems, which capture a wide range of problems with constraints.
We present greedy algorithms that can obtain a high-quality solution, i.e., a constant approximation ratio to the given optimal solution.
arXiv Detail & Related papers (2023-09-23T04:48:49Z) - Bayesian Quality-Diversity approaches for constrained optimization
problems with mixed continuous, discrete and categorical variables [0.3626013617212667]
A new Quality-Diversity methodology based on mixed variables is proposed in the context of limited simulation budget.
The proposed approach provides valuable trade-offs for decision-markers for complex system design.
arXiv Detail & Related papers (2023-09-11T14:29:47Z) - Learning to Optimize with Stochastic Dominance Constraints [103.26714928625582]
In this paper, we develop a simple yet efficient approach for the problem of comparing uncertain quantities.
We recast inner optimization in the Lagrangian as a learning problem for surrogate approximation, which bypasses apparent intractability.
The proposed light-SD demonstrates superior performance on several representative problems ranging from finance to supply chain management.
arXiv Detail & Related papers (2022-11-14T21:54:31Z) - On data-driven chance constraint learning for mixed-integer optimization
problems [0.0]
We develop a Chance Constraint Learning (CCL) methodology with a focus on mixed-integer linear optimization problems.
CCL makes use of linearizable machine learning models to estimate conditional quantiles of the learned variables.
An open-access software has been developed to be used by practitioners.
arXiv Detail & Related papers (2022-07-08T11:54:39Z) - Modeling the Second Player in Distributionally Robust Optimization [90.25995710696425]
We argue for the use of neural generative models to characterize the worst-case distribution.
This approach poses a number of implementation and optimization challenges.
We find that the proposed approach yields models that are more robust than comparable baselines.
arXiv Detail & Related papers (2021-03-18T14:26:26Z) - Offline Model-Based Optimization via Normalized Maximum Likelihood
Estimation [101.22379613810881]
We consider data-driven optimization problems where one must maximize a function given only queries at a fixed set of points.
This problem setting emerges in many domains where function evaluation is a complex and expensive process.
We propose a tractable approximation that allows us to scale our method to high-capacity neural network models.
arXiv Detail & Related papers (2021-02-16T06:04:27Z) - Combining Deep Learning and Optimization for Security-Constrained
Optimal Power Flow [94.24763814458686]
Security-constrained optimal power flow (SCOPF) is fundamental in power systems.
Modeling of APR within the SCOPF problem results in complex large-scale mixed-integer programs.
This paper proposes a novel approach that combines deep learning and robust optimization techniques.
arXiv Detail & Related papers (2020-07-14T12:38:21Z) - Differentially Private Convex Optimization with Feasibility Guarantees [44.36831037077509]
This paper develops a novel differentially private framework to solve convex optimization problems.
The proposed framework provides a trade-off between the expected optimality loss and the variance of optimization results.
arXiv Detail & Related papers (2020-06-22T15:30:52Z) - Automatically Learning Compact Quality-aware Surrogates for Optimization
Problems [55.94450542785096]
Solving optimization problems with unknown parameters requires learning a predictive model to predict the values of the unknown parameters and then solving the problem using these values.
Recent work has shown that including the optimization problem as a layer in a complex training model pipeline results in predictions of iteration of unobserved decision making.
We show that we can improve solution quality by learning a low-dimensional surrogate model of a large optimization problem.
arXiv Detail & Related papers (2020-06-18T19:11:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.