ICNN-enhanced 2SP: Leveraging input convex neural networks for solving two-stage stochastic programming
- URL: http://arxiv.org/abs/2505.05261v1
- Date: Thu, 08 May 2025 14:06:38 GMT
- Title: ICNN-enhanced 2SP: Leveraging input convex neural networks for solving two-stage stochastic programming
- Authors: Yu Liu, Fabricio Oliveira,
- Abstract summary: Two-stage programming (2SP) offers a basic framework for modelling decision-making under uncertainty.<n>Existing learning-based methods like Neural Two-Stage Programming (Neur2SP) employ neural networks (NNs) as recourse function surrogates.<n>We propose ICNN-enhanced 2SP, a method that leverages ICNNs to exploit linear programming representability in convex 2SP problems.
- Score: 3.528295407065282
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Two-stage stochastic programming (2SP) offers a basic framework for modelling decision-making under uncertainty, yet scalability remains a challenge due to the computational complexity of recourse function evaluation. Existing learning-based methods like Neural Two-Stage Stochastic Programming (Neur2SP) employ neural networks (NNs) as recourse function surrogates but rely on computationally intensive mixed-integer programming (MIP) formulations. We propose ICNN-enhanced 2SP, a method that leverages Input Convex Neural Networks (ICNNs) to exploit linear programming (LP) representability in convex 2SP problems. By architecturally enforcing convexity and enabling exact inference through LP, our approach eliminates the need for integer variables inherent to the conventional MIP-based formulation while retaining an exact embedding of the ICNN surrogate within the 2SP framework. This results in a more computationally efficient alternative that maintains solution quality. Comprehensive experiments reveal that ICNNs incur only marginally longer training times while achieving validation accuracy on par with their MIP-based counterparts. Across benchmark problems, ICNN-enhanced 2SP often exhibits considerably faster solution times than the MIP-based formulations while preserving solution quality, with these advantages becoming significantly more pronounced as problem scale increases. For the most challenging instances, the method achieves speedups of up to 100$\times$ and solution quality superior to MIP-based formulations.
Related papers
- Neural Two-Stage Stochastic Optimization for Solving Unit Commitment Problem [0.8848340429852071]
This paper proposes a neural optimization method for efficiently solving the two-stage unit commitment (2S-SUC) problem under high-dimensional uncertainty scenarios.<n>The proposed method approximates the second-stage recourse problem using a deep neural network trained to map commitment decisions and uncertainty features to recourse costs.
arXiv Detail & Related papers (2025-07-13T05:55:25Z) - Performance Analysis of Convolutional Neural Network By Applying Unconstrained Binary Quadratic Programming [0.0]
Convolutional Neural Networks (CNNs) are pivotal in computer vision and Big Data analytics but demand significant computational resources when trained on large-scale datasets.<n>We propose a hybrid optimization method that combines Unconstrained Binary Quadratic Programming (UBQP) with Gradient Descent (SGD) to accelerate CNN training.<n>Our approach achieves a 10--15% accuracy improvement over a standard BP-CNN baseline while maintaining similar execution times.
arXiv Detail & Related papers (2025-05-30T21:25:31Z) - Verifying Properties of Binary Neural Networks Using Sparse Polynomial Optimization [8.323690755070123]
This paper explores methods for verifying the properties of Binary Neural Networks (BNNs)<n>BNNs, like their full-precision counterparts, are also sensitive to input perturbations.<n>We introduce an alternative approach using Semidefinite Programming relaxations derived from sparse Polynomial Optimization.
arXiv Detail & Related papers (2024-05-27T11:03:48Z) - A Multi-Head Ensemble Multi-Task Learning Approach for Dynamical
Computation Offloading [62.34538208323411]
We propose a multi-head ensemble multi-task learning (MEMTL) approach with a shared backbone and multiple prediction heads (PHs)
MEMTL outperforms benchmark methods in both the inference accuracy and mean square error without requiring additional training data.
arXiv Detail & Related papers (2023-09-02T11:01:16Z) - Neur2SP: Neural Two-Stage Stochastic Programming [3.5417030119735045]
We develop Neur2SP, a new method that approximates the expected value function via a neural network.
Neur2SP takes less than 1.66 seconds across all problems, achieving high-quality solutions even as the number of scenarios increases.
arXiv Detail & Related papers (2022-05-20T22:09:22Z) - Comparative Analysis of Interval Reachability for Robust Implicit and
Feedforward Neural Networks [64.23331120621118]
We use interval reachability analysis to obtain robustness guarantees for implicit neural networks (INNs)
INNs are a class of implicit learning models that use implicit equations as layers.
We show that our approach performs at least as well as, and generally better than, applying state-of-the-art interval bound propagation methods to INNs.
arXiv Detail & Related papers (2022-04-01T03:31:27Z) - Neural Stochastic Dual Dynamic Programming [99.80617899593526]
We introduce a trainable neural model that learns to map problem instances to a piece-wise linear value function.
$nu$-SDDP can significantly reduce problem solving cost without sacrificing solution quality.
arXiv Detail & Related papers (2021-12-01T22:55:23Z) - A generic physics-informed neural network-based framework for
reliability assessment of multi-state systems [1.6440434996206623]
We develop a generic PINN-based framework to assess the reliability of multi-state systems (MSSs)
We tackle the problem of high imbalance in the magnitude of the back-propagated gradients in PINN from a multi-task learning perspective.
The proposed PINN-based framework shows generic and remarkable performance in MSS reliability assessment.
arXiv Detail & Related papers (2021-12-01T01:43:57Z) - Learning from Images: Proactive Caching with Parallel Convolutional
Neural Networks [94.85780721466816]
A novel framework for proactive caching is proposed in this paper.
It combines model-based optimization with data-driven techniques by transforming an optimization problem into a grayscale image.
Numerical results show that the proposed scheme can reduce 71.6% computation time with only 0.8% additional performance cost.
arXiv Detail & Related papers (2021-08-15T21:32:47Z) - Combining Deep Learning and Optimization for Security-Constrained
Optimal Power Flow [94.24763814458686]
Security-constrained optimal power flow (SCOPF) is fundamental in power systems.
Modeling of APR within the SCOPF problem results in complex large-scale mixed-integer programs.
This paper proposes a novel approach that combines deep learning and robust optimization techniques.
arXiv Detail & Related papers (2020-07-14T12:38:21Z) - Iterative Algorithm Induced Deep-Unfolding Neural Networks: Precoding
Design for Multiuser MIMO Systems [59.804810122136345]
We propose a framework for deep-unfolding, where a general form of iterative algorithm induced deep-unfolding neural network (IAIDNN) is developed.
An efficient IAIDNN based on the structure of the classic weighted minimum mean-square error (WMMSE) iterative algorithm is developed.
We show that the proposed IAIDNN efficiently achieves the performance of the iterative WMMSE algorithm with reduced computational complexity.
arXiv Detail & Related papers (2020-06-15T02:57:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.