The Sigma-max System Induced from Randomness & Fuzziness and its Application in Time Series Prediction
- URL: http://arxiv.org/abs/2110.07722v2
- Date: Thu, 02 Jan 2025 02:24:58 GMT
- Title: The Sigma-max System Induced from Randomness & Fuzziness and its Application in Time Series Prediction
- Authors: Wei Mei, Ming Li, Yuanzeng Cheng, Limin Liu,
- Abstract summary: We focus on why the key axiom of "maxitivity" is adopted for possibility measure.<n>Our work provides a physical foundation for the axiomatic definition of possibility for the measure of fuzziness.
- Score: 5.648717826360932
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper managed to induce probability theory (sigma system) and possibility theory (max system) respectively from the clearly-defined randomness and fuzziness, while focusing the question why the key axiom of "maxitivity" is adopted for possibility measure. Such an objective is achieved by following three steps: a) the establishment of mathematical definitions of randomness and fuzziness; b) the development of intuitive definition of possibility as measure of fuzziness based on compatibility interpretation; c) the abstraction of the axiomatic definitions of probability/ possibility from their intuitive definitions, by taking advantage of properties of the well-defined randomness and fuzziness. We derived the conclusion that "max" is the only but un-strict disjunctive operator that is applicable across the fuzzy event space, and is an exact operator for extracting the value from the fuzzy sample space that leads to the largest possibility of one. Then a demonstration example of stock price prediction is presented, which confirms that max inference indeed exhibits distinctive performance, with an improvement up to 18.99%, over sigma inference for the investigated application. Our work provides a physical foundation for the axiomatic definition of possibility for the measure of fuzziness, which hopefully would facilitate wider adoption of possibility theory in practice.
Related papers
- Model uncertainty quantification using feature confidence sets for outcome excursions [0.0]
This paper introduces a novel, model-agnostic framework for quantifying uncertainty in continuous and binary outcomes.
It is validated through simulations and applied to real-world datasets in contexts such as housing price prediction and time to sepsis diagnosis in healthcare.
arXiv Detail & Related papers (2025-04-28T04:08:07Z) - Barycentric bounds on the error exponents of quantum hypothesis exclusion [7.812210699650153]
We study the optimal error probability of quantum state exclusion from an information-theoretic perspective.
We extend our analysis to the more complicated task of quantum channel exclusion.
arXiv Detail & Related papers (2024-07-18T17:27:36Z) - Probabilistic Conformal Prediction with Approximate Conditional Validity [81.30551968980143]
We develop a new method for generating prediction sets that combines the flexibility of conformal methods with an estimate of the conditional distribution.
Our method consistently outperforms existing approaches in terms of conditional coverage.
arXiv Detail & Related papers (2024-07-01T20:44:48Z) - On The Statistical Representation Properties Of The Perturb-Softmax And The Perturb-Argmax Probability Distributions [17.720298535412443]
Gumbel-Softmax and Gumbel-Argmax probability distributions are useful in learning discrete structures in discriminative learning.
Despite the efforts invested in optimizing these probability models, their statistical properties are under-explored.
We investigate their representation properties and determine for which families of parameters these probability distributions are complete.
We conclude the analysis by identifying two sets of parameters that satisfy these assumptions and thus admit a complete and minimal representation.
arXiv Detail & Related papers (2024-06-04T10:22:12Z) - Calibrating Neural Simulation-Based Inference with Differentiable
Coverage Probability [50.44439018155837]
We propose to include a calibration term directly into the training objective of the neural model.
By introducing a relaxation of the classical formulation of calibration error we enable end-to-end backpropagation.
It is directly applicable to existing computational pipelines allowing reliable black-box posterior inference.
arXiv Detail & Related papers (2023-10-20T10:20:45Z) - A Semantic Approach to Decidability in Epistemic Planning (Extended
Version) [72.77805489645604]
We use a novel semantic approach to achieve decidability.
Specifically, we augment the logic of knowledge S5$_n$ and with an interaction axiom called (knowledge) commutativity.
We prove that our framework admits a finitary non-fixpoint characterization of common knowledge, which is of independent interest.
arXiv Detail & Related papers (2023-07-28T11:26:26Z) - What is \textit{Quantum} in Probabilistic Explanations of the Sure Thing
Principle Violation? [0.0]
The Prisoner's Dilemma game (PDG) is one of the simple test-beds for the probabilistic nature of the human decision-making process.
Quantum probabilistic models can explain this violation as a second-order interference effect.
We discuss the role of other quantum information-theoretical quantities, such as quantum entanglement, in the decision-making process.
arXiv Detail & Related papers (2023-06-21T00:01:01Z) - A Measure-Theoretic Axiomatisation of Causality [55.6970314129444]
We argue in favour of taking Kolmogorov's measure-theoretic axiomatisation of probability as the starting point towards an axiomatisation of causality.
Our proposed framework is rigorously grounded in measure theory, but it also sheds light on long-standing limitations of existing frameworks.
arXiv Detail & Related papers (2023-05-19T13:15:48Z) - $\omega$PAP Spaces: Reasoning Denotationally About Higher-Order,
Recursive Probabilistic and Differentiable Programs [64.25762042361839]
$omega$PAP spaces are spaces for reasoning denotationally about expressive differentiable and probabilistic programming languages.
Our semantics is general enough to assign meanings to most practical probabilistic and differentiable programs.
We establish the almost-everywhere differentiability of probabilistic programs' trace density functions.
arXiv Detail & Related papers (2023-02-21T12:50:05Z) - Relative Probability on Finite Outcome Spaces: A Systematic Examination
of its Axiomatization, Properties, and Applications [0.0]
This work proposes a view of probability as a relative measure rather than an absolute one.
We focus on finite outcome spaces and develop three fundamental axioms that establish requirements for relative probability functions.
arXiv Detail & Related papers (2022-12-30T05:16:57Z) - Bounding Counterfactuals under Selection Bias [60.55840896782637]
We propose a first algorithm to address both identifiable and unidentifiable queries.
We prove that, in spite of the missingness induced by the selection bias, the likelihood of the available data is unimodal.
arXiv Detail & Related papers (2022-07-26T10:33:10Z) - The Unreasonable Effectiveness of Deep Evidential Regression [72.30888739450343]
A new approach with uncertainty-aware regression-based neural networks (NNs) shows promise over traditional deterministic methods and typical Bayesian NNs.
We detail the theoretical shortcomings and analyze the performance on synthetic and real-world data sets, showing that Deep Evidential Regression is a quantification rather than an exact uncertainty.
arXiv Detail & Related papers (2022-05-20T10:10:32Z) - Robustness Guarantees for Credal Bayesian Networks via Constraint
Relaxation over Probabilistic Circuits [16.997060715857987]
We develop a method to quantify the robustness of decision functions with respect to credal Bayesian networks.
We show how to obtain a guaranteed upper bound on MARmax in linear time in the size of the circuit.
arXiv Detail & Related papers (2022-05-11T22:37:07Z) - A Logic-based Tractable Approximation of Probability [0.0]
We identify the conditions under which propositional probability functions can be approximated by a hierarchy of depth-bounded Belief functions.
We show that our approximations of probability lead to uncertain reasoning which, under the usual assumptions in the field, qualifies as tractable.
arXiv Detail & Related papers (2022-05-06T13:25:12Z) - Logical Credal Networks [87.25387518070411]
This paper introduces Logical Credal Networks, an expressive probabilistic logic that generalizes many prior models that combine logic and probability.
We investigate its performance on maximum a posteriori inference tasks, including solving Mastermind games with uncertainty and detecting credit card fraud.
arXiv Detail & Related papers (2021-09-25T00:00:47Z) - Maximum Entropy competes with Maximum Likelihood [0.0]
Max entropy (MAXENT) method has a large number of applications in theoretical and applied machine learning.
We show that MAXENT applies in sparse data regimes, but needs specific types of prior information.
In particular, MAXENT can outperform the optimally regularized ML provided that there are prior rank correlations between the estimated random quantity and its probabilities.
arXiv Detail & Related papers (2020-12-17T07:44:22Z) - Amortized Conditional Normalized Maximum Likelihood: Reliable Out of
Distribution Uncertainty Estimation [99.92568326314667]
We propose the amortized conditional normalized maximum likelihood (ACNML) method as a scalable general-purpose approach for uncertainty estimation.
Our algorithm builds on the conditional normalized maximum likelihood (CNML) coding scheme, which has minimax optimal properties according to the minimum description length principle.
We demonstrate that ACNML compares favorably to a number of prior techniques for uncertainty estimation in terms of calibration on out-of-distribution inputs.
arXiv Detail & Related papers (2020-11-05T08:04:34Z) - A Weaker Faithfulness Assumption based on Triple Interactions [89.59955143854556]
We propose a weaker assumption that we call $2$-adjacency faithfulness.
We propose a sound orientation rule for causal discovery that applies under weaker assumptions.
arXiv Detail & Related papers (2020-10-27T13:04:08Z) - Distributionally Robust Bayesian Quadrature Optimization [60.383252534861136]
We study BQO under distributional uncertainty in which the underlying probability distribution is unknown except for a limited set of its i.i.d. samples.
A standard BQO approach maximizes the Monte Carlo estimate of the true expected objective given the fixed sample set.
We propose a novel posterior sampling based algorithm, namely distributionally robust BQO (DRBQO) for this purpose.
arXiv Detail & Related papers (2020-01-19T12:00:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.