An Offline Deep Reinforcement Learning for Maintenance Decision-Making
- URL: http://arxiv.org/abs/2109.15050v1
- Date: Tue, 28 Sep 2021 03:40:55 GMT
- Title: An Offline Deep Reinforcement Learning for Maintenance Decision-Making
- Authors: Hamed Khorasgani, Haiyan Wang, Chetan Gupta, and Ahmed Farahat
- Abstract summary: We present a maintenance framework based on offline supervised deep reinforcement learning.
Using offline reinforcement learning makes it possible to learn the optimum maintenance policy from historical data.
- Score: 10.244120641608447
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Several machine learning and deep learning frameworks have been proposed to
solve remaining useful life estimation and failure prediction problems in
recent years. Having access to the remaining useful life estimation or
likelihood of failure in near future helps operators to assess the operating
conditions and, therefore, provides better opportunities for sound repair and
maintenance decisions. However, many operators believe remaining useful life
estimation and failure prediction solutions are incomplete answers to the
maintenance challenge. They argue that knowing the likelihood of failure in the
future is not enough to make maintenance decisions that minimize costs and keep
the operators safe. In this paper, we present a maintenance framework based on
offline supervised deep reinforcement learning that instead of providing
information such as likelihood of failure, suggests actions such as
"continuation of the operation" or "the visitation of the repair shop" to the
operators in order to maximize the overall profit. Using offline reinforcement
learning makes it possible to learn the optimum maintenance policy from
historical data without relying on expensive simulators. We demonstrate the
application of our solution in a case study using the NASA C-MAPSS dataset.
Related papers
- Towards Robust Continual Learning with Bayesian Adaptive Moment Regularization [51.34904967046097]
Continual learning seeks to overcome the challenge of catastrophic forgetting, where a model forgets previously learnt information.
We introduce a novel prior-based method that better constrains parameter growth, reducing catastrophic forgetting.
Results show that BAdam achieves state-of-the-art performance for prior-based methods on challenging single-headed class-incremental experiments.
arXiv Detail & Related papers (2023-09-15T17:10:51Z) - Bayesian Inverse Transition Learning for Offline Settings [30.10905852013852]
Reinforcement learning is commonly used for sequential decision-making in domains such as healthcare and education.
We propose a new constraint-based approach that captures our desiderata for reliably learning a posterior distribution of the transition dynamics $T$.
Our results demonstrate that by using our constraints, we learn a high-performing policy, while considerably reducing the policy's variance over different datasets.
arXiv Detail & Related papers (2023-08-09T17:08:29Z) - Reinforcement and Deep Reinforcement Learning-based Solutions for
Machine Maintenance Planning, Scheduling Policies, and Optimization [1.6447597767676658]
This paper presents a literature review on the applications of reinforcement and deep reinforcement learning for maintenance planning and optimization problems.
By leveraging the condition monitoring data of systems and machines with reinforcement learning, smart maintenance planners can be developed, which is a precursor to achieving a smart factory.
arXiv Detail & Related papers (2023-07-07T22:47:29Z) - Resilient Constrained Learning [94.27081585149836]
This paper presents a constrained learning approach that adapts the requirements while simultaneously solving the learning task.
We call this approach resilient constrained learning after the term used to describe ecological systems that adapt to disruptions by modifying their operation.
arXiv Detail & Related papers (2023-06-04T18:14:18Z) - Re-thinking Data Availablity Attacks Against Deep Neural Networks [53.64624167867274]
In this paper, we re-examine the concept of unlearnable examples and discern that the existing robust error-minimizing noise presents an inaccurate optimization objective.
We introduce a novel optimization paradigm that yields improved protection results with reduced computational time requirements.
arXiv Detail & Related papers (2023-05-18T04:03:51Z) - Prescriptive maintenance with causal machine learning [4.169130102668252]
We learn the effect of maintenance conditional on a machine's characteristics from observational data on similar machines.
We validate our proposed approach using real-life data on more than 4,000 maintenance contracts from an industrial partner.
arXiv Detail & Related papers (2022-06-03T13:35:57Z) - Predictive Maintenance using Machine Learning [0.0]
Predictive maintenance (PdM) is implemented to effectively manage maintenance plans of the assets.
Data is collected over a certain period of time to monitor the state of equipment.
arXiv Detail & Related papers (2022-05-19T09:05:37Z) - Uncertainty-aware Remaining Useful Life predictor [57.74855412811814]
Remaining Useful Life (RUL) estimation is the problem of inferring how long a certain industrial asset can be expected to operate.
In this work, we consider Deep Gaussian Processes (DGPs) as possible solutions to the aforementioned limitations.
The performance of the algorithms is evaluated on the N-CMAPSS dataset from NASA for aircraft engines.
arXiv Detail & Related papers (2021-04-08T08:50:44Z) - VisioRed: A Visualisation Tool for Interpretable Predictive Maintenance [5.845912816093006]
Using machine learning, predictive and prescriptive maintenance attempt to anticipate and prevent eventual system failures.
This paper introduces a visualisation tool incorporating interpretations to display information derived from predictive maintenance models, trained on time-series data.
arXiv Detail & Related papers (2021-03-31T11:35:51Z) - Right Decisions from Wrong Predictions: A Mechanism Design Alternative
to Individual Calibration [107.15813002403905]
Decision makers often need to rely on imperfect probabilistic forecasts.
We propose a compensation mechanism ensuring that the forecasted utility matches the actually accrued utility.
We demonstrate an application showing how passengers could confidently optimize individual travel plans based on flight delay probabilities.
arXiv Detail & Related papers (2020-11-15T08:22:39Z) - Chance-Constrained Trajectory Optimization for Safe Exploration and
Learning of Nonlinear Systems [81.7983463275447]
Learning-based control algorithms require data collection with abundant supervision for training.
We present a new approach for optimal motion planning with safe exploration that integrates chance-constrained optimal control with dynamics learning and feedback control.
arXiv Detail & Related papers (2020-05-09T05:57:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.