Efficient Fraud Detection Using Deep Boosting Decision Trees
- URL: http://arxiv.org/abs/2302.05918v2
- Date: Thu, 18 May 2023 11:19:24 GMT
- Title: Efficient Fraud Detection Using Deep Boosting Decision Trees
- Authors: Biao Xu, Yao Wang, Xiuwu Liao, Kaidong Wang
- Abstract summary: Fraud detection is to identify, monitor, and prevent potentially fraudulent activities from complex data.
Recent development and success in AI, especially machine learning, provides a new data-driven way to deal with fraud.
Deep boosting decision trees (DBDT) is a novel approach for fraud detection based on gradient boosting and neural networks.
- Score: 8.941773715949697
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Fraud detection is to identify, monitor, and prevent potentially fraudulent
activities from complex data. The recent development and success in AI,
especially machine learning, provides a new data-driven way to deal with fraud.
From a methodological point of view, machine learning based fraud detection can
be divided into two categories, i.e., conventional methods (decision tree,
boosting...) and deep learning, both of which have significant limitations in
terms of the lack of representation learning ability for the former and
interpretability for the latter. Furthermore, due to the rarity of detected
fraud cases, the associated data is usually imbalanced, which seriously
degrades the performance of classification algorithms. In this paper, we
propose deep boosting decision trees (DBDT), a novel approach for fraud
detection based on gradient boosting and neural networks. In order to combine
the advantages of both conventional methods and deep learning, we first
construct soft decision tree (SDT), a decision tree structured model with
neural networks as its nodes, and then ensemble SDTs using the idea of gradient
boosting. In this way we embed neural networks into gradient boosting to
improve its representation learning capability and meanwhile maintain the
interpretability. Furthermore, aiming at the rarity of detected fraud cases, in
the model training phase we propose a compositional AUC maximization approach
to deal with data imbalances at algorithm level. Extensive experiments on
several real-life fraud detection datasets show that DBDT can significantly
improve the performance and meanwhile maintain good interpretability. Our code
is available at https://github.com/freshmanXB/DBDT.
Related papers
- Comprehensive Botnet Detection by Mitigating Adversarial Attacks, Navigating the Subtleties of Perturbation Distances and Fortifying Predictions with Conformal Layers [1.6001193161043425]
Botnets are computer networks controlled by malicious actors that present significant cybersecurity challenges.
This research addresses the sophisticated adversarial manipulations posed by attackers, aiming to undermine machine learning-based botnet detection systems.
We introduce a flow-based detection approach, leveraging machine learning and deep learning algorithms trained on the ISCX and ISOT datasets.
arXiv Detail & Related papers (2024-09-01T08:53:21Z) - Development of Multistage Machine Learning Classifier using Decision Trees and Boosting Algorithms over Darknet Network Traffic [0.0]
This study proposes a robust solution for network traffic classification.
By leveraging boosting algorithms like AdaBoost and Gradient Boosting coupled with decision trees, this study proposes a robust solution for network traffic classification.
arXiv Detail & Related papers (2024-07-22T17:10:26Z) - The Cascaded Forward Algorithm for Neural Network Training [61.06444586991505]
We propose a new learning framework for neural networks, namely Cascaded Forward (CaFo) algorithm, which does not rely on BP optimization as that in FF.
Unlike FF, our framework directly outputs label distributions at each cascaded block, which does not require generation of additional negative samples.
In our framework each block can be trained independently, so it can be easily deployed into parallel acceleration systems.
arXiv Detail & Related papers (2023-03-17T02:01:11Z) - Backdoor Attack Detection in Computer Vision by Applying Matrix
Factorization on the Weights of Deep Networks [6.44397009982949]
We introduce a novel method for backdoor detection that extracts features from pre-trained DNN's weights.
In comparison to other detection techniques, this has a number of benefits, such as not requiring any training data.
Our method outperforms the competing algorithms in terms of efficiency and is more accurate, helping to ensure the safe application of deep learning and AI.
arXiv Detail & Related papers (2022-12-15T20:20:18Z) - Can pruning improve certified robustness of neural networks? [106.03070538582222]
We show that neural network pruning can improve empirical robustness of deep neural networks (NNs)
Our experiments show that by appropriately pruning an NN, its certified accuracy can be boosted up to 8.2% under standard training.
We additionally observe the existence of certified lottery tickets that can match both standard and certified robust accuracies of the original dense models.
arXiv Detail & Related papers (2022-06-15T05:48:51Z) - Neurosymbolic hybrid approach to driver collision warning [64.02492460600905]
There are two main algorithmic approaches to autonomous driving systems.
Deep learning alone has achieved state-of-the-art results in many areas.
But sometimes it can be very difficult to debug if the deep learning model doesn't work.
arXiv Detail & Related papers (2022-03-28T20:29:50Z) - Multigoal-oriented dual-weighted-residual error estimation using deep
neural networks [0.0]
Deep learning is considered as a powerful tool with high flexibility to approximate functions.
Our approach is based on a posteriori error estimation in which the adjoint problem is solved for the error localization.
An efficient and easy to implement algorithm is developed to obtain a posteriori error estimate for multiple goal functionals.
arXiv Detail & Related papers (2021-12-21T16:59:44Z) - Anomaly Detection in Cybersecurity: Unsupervised, Graph-Based and
Supervised Learning Methods in Adversarial Environments [63.942632088208505]
Inherent to today's operating environment is the practice of adversarial machine learning.
In this work, we examine the feasibility of unsupervised learning and graph-based methods for anomaly detection.
We incorporate a realistic adversarial training mechanism when training our supervised models to enable strong classification performance in adversarial environments.
arXiv Detail & Related papers (2021-05-14T10:05:10Z) - Learning a Domain-Agnostic Visual Representation for Autonomous Driving
via Contrastive Loss [25.798361683744684]
Domain-Agnostic Contrastive Learning (DACL) is a two-stage unsupervised domain adaptation framework with cyclic adversarial training and contrastive loss.
Our proposed approach achieves better performance in the monocular depth estimation task compared to previous state-of-the-art methods.
arXiv Detail & Related papers (2021-03-10T07:06:03Z) - Adversarial Self-Supervised Contrastive Learning [62.17538130778111]
Existing adversarial learning approaches mostly use class labels to generate adversarial samples that lead to incorrect predictions.
We propose a novel adversarial attack for unlabeled data, which makes the model confuse the instance-level identities of the perturbed data samples.
We present a self-supervised contrastive learning framework to adversarially train a robust neural network without labeled data.
arXiv Detail & Related papers (2020-06-13T08:24:33Z) - BiDet: An Efficient Binarized Object Detector [96.19708396510894]
We propose a binarized neural network learning method called BiDet for efficient object detection.
Our BiDet fully utilizes the representational capacity of the binary neural networks for object detection by redundancy removal.
Our method outperforms the state-of-the-art binary neural networks by a sizable margin.
arXiv Detail & Related papers (2020-03-09T08:16:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.