Development of fully intuitionistic fuzzy data envelopment analysis
model with missing data: an application to Indian police sector
- URL: http://arxiv.org/abs/2208.02675v1
- Date: Wed, 27 Jul 2022 18:20:13 GMT
- Title: Development of fully intuitionistic fuzzy data envelopment analysis
model with missing data: an application to Indian police sector
- Authors: Anjali Sonkariya, Awadh Pratap Singh, Shiv Prasad Yadav
- Abstract summary: DEA is a technique used to measure the efficiency of decision-making units (DMUs)
Data is usually collected by humans, machines, or both.
A method is presented that can deal with missing values and inaccuracy in the data.
A real-life application to measure the performance efficiencies of Indian police stations is presented.
- Score: 2.9434930072968584
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Data Envelopment Analysis (DEA) is a technique used to measure the efficiency
of decision-making units (DMUs). In order to measure the efficiency of DMUs,
the essential requirement is input-output data. Data is usually collected by
humans, machines, or both. Due to human/machine errors, there are chances of
having some missing values or inaccuracy, such as
vagueness/uncertainty/hesitation in the collected data. In this situation, it
will be difficult to measure the efficiencies of DMUs accurately. To overcome
these shortcomings, a method is presented that can deal with missing values and
inaccuracy in the data. To measure the performance efficiencies of DMUs, an
input minimization BCC (IMBCC) model in a fully intuitionistic fuzzy (IF)
environment is proposed. To validate the efficacy of the proposed fully
intuitionistic fuzzy input minimization BCC (FIFIMBCC) model and the technique
to deal with missing values in the data, a real-life application to measure the
performance efficiencies of Indian police stations is presented.
Related papers
- Towards a Theoretical Understanding of Memorization in Diffusion Models [76.85077961718875]
Diffusion probabilistic models (DPMs) are being employed as mainstream models for Generative Artificial Intelligence (GenAI)
We provide a theoretical understanding of memorization in both conditional and unconditional DPMs under the assumption of model convergence.
We propose a novel data extraction method named textbfSurrogate condItional Data Extraction (SIDE) that leverages a time-dependent classifier trained on the generated data as a surrogate condition to extract training data from unconditional DPMs.
arXiv Detail & Related papers (2024-10-03T13:17:06Z) - Missing Value Imputation for Multi-attribute Sensor Data Streams via
Message Propagation (Extended Version) [25.022656067909523]
We propose a message propagation imputation network (MPIN) that is able to recover the missing values of data instances in a time window.
MPIN can outperform the existing data imputers by wide margins and that the continuous imputation framework is efficient and accurate.
arXiv Detail & Related papers (2023-11-13T14:01:04Z) - Filling the Missing: Exploring Generative AI for Enhanced Federated
Learning over Heterogeneous Mobile Edge Devices [72.61177465035031]
We propose a generative AI-empowered federated learning to address these challenges by leveraging the idea of FIlling the MIssing (FIMI) portion of local data.
Experiment results demonstrate that FIMI can save up to 50% of the device-side energy to achieve the target global test accuracy.
arXiv Detail & Related papers (2023-10-21T12:07:04Z) - Machine Learning Force Fields with Data Cost Aware Training [94.78998399180519]
Machine learning force fields (MLFF) have been proposed to accelerate molecular dynamics (MD) simulation.
Even for the most data-efficient MLFFs, reaching chemical accuracy can require hundreds of frames of force and energy labels.
We propose a multi-stage computational framework -- ASTEROID, which lowers the data cost of MLFFs by leveraging a combination of cheap inaccurate data and expensive accurate data.
arXiv Detail & Related papers (2023-06-05T04:34:54Z) - Conditional expectation with regularization for missing data imputation [19.254291863337347]
Missing data frequently occurs in datasets across various domains, such as medicine, sports, and finance.
We propose a new algorithm named "conditional Distribution-based Imputation of Missing Values with Regularization" (DIMV)
DIMV operates by determining the conditional distribution of a feature that has missing entries, using the information from the fully observed features as a basis.
arXiv Detail & Related papers (2023-02-02T06:59:15Z) - Multiple Imputation with Neural Network Gaussian Process for
High-dimensional Incomplete Data [9.50726756006467]
Imputation is arguably the most popular method for handling missing data, though existing methods have a number of limitations.
We propose two NNGP-based MI methods, namely MI-NNGP, that can apply multiple imputations for missing values from a joint (posterior predictive) distribution.
The MI-NNGP methods are shown to significantly outperform existing state-of-the-art methods on synthetic and real datasets.
arXiv Detail & Related papers (2022-11-23T20:54:26Z) - Calibrating Agent-based Models to Microdata with Graph Neural Networks [1.4911092205861822]
Calibrating agent-based models (ABMs) to data is among the most fundamental requirements to ensure the model fulfils its desired purpose.
We propose to learn parameter posteriors associated with granular microdata directly using temporal graph neural networks.
arXiv Detail & Related papers (2022-06-15T14:41:43Z) - MissDAG: Causal Discovery in the Presence of Missing Data with
Continuous Additive Noise Models [78.72682320019737]
We develop a general method, which we call MissDAG, to perform causal discovery from data with incomplete observations.
MissDAG maximizes the expected likelihood of the visible part of observations under the expectation-maximization framework.
We demonstrate the flexibility of MissDAG for incorporating various causal discovery algorithms and its efficacy through extensive simulations and real data experiments.
arXiv Detail & Related papers (2022-05-27T09:59:46Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z) - Imputation-Free Learning from Incomplete Observations [73.15386629370111]
We introduce the importance of guided gradient descent (IGSGD) method to train inference from inputs containing missing values without imputation.
We employ reinforcement learning (RL) to adjust the gradients used to train the models via back-propagation.
Our imputation-free predictions outperform the traditional two-step imputation-based predictions using state-of-the-art imputation methods.
arXiv Detail & Related papers (2021-07-05T12:44:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.