New AI model detects credit card fraud with 99% accuracy and zero black box

The core problem tackled by this research is the opaque nature of many high-performing AI models used in fraud detection. Traditional models like deep neural networks can accurately flag fraudulent transactions but often do so as “black boxes,” offering no insight into how decisions are made - posing compliance and ethical issues for financial institutions. The proposed solution overcomes this by integrating explainable artificial intelligence (XAI) techniques into a powerful stacking ensemble model.


CO-EDP, VisionRICO-EDP, VisionRI | Updated: 17-05-2025 10:10 IST | Created: 17-05-2025 10:10 IST
New AI model detects credit card fraud with 99% accuracy and zero black box
Representative Image. Credit: ChatGPT

The rise in credit card fraud has not only tested the resilience of banks but also exposed the ethical pitfalls of high-performing yet opaque AI systems. Now, researchers from Taif University have developed a cutting-edge artificial intelligence (AI) framework for detecting credit card fraud that achieves near-perfect accuracy while also remaining fully interpretable.

The study, titled "Financial Fraud Detection Using Explainable AI and Stacking Ensemble Methods", was published on arXiv and addresses long-standing trade-offs between predictive performance and transparency in financial AI systems.

Can machine learning be both accurate and explainable?

The core problem tackled by this research is the opaque nature of many high-performing AI models used in fraud detection. Traditional models like deep neural networks can accurately flag fraudulent transactions but often do so as “black boxes,” offering no insight into how decisions are made - posing compliance and ethical issues for financial institutions. The proposed solution overcomes this by integrating explainable artificial intelligence (XAI) techniques into a powerful stacking ensemble model.

The authors combined three state-of-the-art gradient boosting algorithms - XGBoost, LightGBM, and CatBoost - into a stacking ensemble, using an XGBoost model as a meta-learner. This setup was then enhanced with SHAP (SHapley Additive Explanations), LIME (Local Interpretable Model-Agnostic Explanations), Partial Dependence Plots (PDP), and Permutation Feature Importance (PFI) tools, allowing both global and local interpretability. SHAP was also used for feature selection, ensuring that the model not only achieved high accuracy but did so with a reduced and meaningful feature set.

Trained on the IEEE-CIS Fraud Detection dataset, which includes over 590,000 real-world transaction records, the model attained 99% accuracy and an AUC-ROC score of 0.998. Crucially, it identified fraudulent transactions with a 98% recall and a 1.00 precision rate, offering an exceptionally low false positive rate - only 481 out of more than 113,000 legitimate transactions were misclassified.

What makes this framework outperform existing models?

The framework’s success is built on rigorous data processing, ensemble learning, and fine-tuned hyperparameter optimization. The researchers used Optuna for Bayesian optimization to find the best parameter combinations for XGBoost, such as the number of estimators, tree depth, learning rate, and class weighting. Furthermore, they tackled class imbalance - a major issue in fraud detection where fraud cases represent a tiny fraction of total transactions - by applying SMOTE (Synthetic Minority Oversampling Technique) to balance the training data.

The stacking ensemble outperformed traditional models such as logistic regression and decision trees by a wide margin. Compared to a logistic regression model, which achieved only 61% accuracy and a 50% recall for fraud, the proposed model demonstrated much higher robustness and practical reliability. Even compared to other recent advanced ensemble methods, such as those using hybrid deep learning or multi-model stacking, the Taif University framework held its ground. Many competing studies either relied on synthetic datasets or delivered lower interpretability scores.

Crucially, the model maintained consistency across 5-fold stratified cross-validation with AUC scores hovering around 0.998, indicating strong generalizability. The performance remained robust despite variations in training and test splits, further solidifying its potential for real-world deployment.

How does explainability improve trust and compliance?

Beyond performance, explainability is the hallmark of this research. Using SHAP, the team identified the top 30 most influential features, such as “C14” and “V12,” which showed strong predictive power in both global and localized contexts. LIME was employed to generate user-friendly, per-transaction explanations, helping to decode why specific transactions were classified as fraudulent or legitimate. For instance, a highlighted LIME case showed how features pushed a prediction toward non-fraud with a 98% confidence level - an output that could be shown to an auditor or customer to justify the decision.

Permutation Feature Importance and PDPs provided additional insight into how the model’s fraud predictions changed with variations in specific features, offering transparency into nonlinear and threshold-based behavior of the algorithm. These tools ensure the model’s decisions can be scrutinized and justified, making it more acceptable to regulators, auditors, and users.

This emphasis on transparency aligns with growing global regulatory expectations, particularly under frameworks such as the EU’s General Data Protection Regulation (GDPR), which mandates “meaningful information about the logic involved” in automated decisions. The model thus provides a path for financial institutions to meet both performance demands and regulatory requirements in a cost-effective and ethically sound manner.

Looking ahead: Toward ethical, scalable and real-time solutions

The research marks a critical turning point in the development of fraud detection systems. While previous models offered either high accuracy or interpretability, this framework successfully combines both. Its real-world applicability is further strengthened by using a publicly available, large-scale dataset rather than a synthetic or curated one, reinforcing its practicality.

Yet, the study also acknowledges that further work remains. Real-time deployment, continual learning for evolving fraud patterns, and fairness audits are the next frontiers. The authors propose exploring federated learning systems, where models are trained across decentralized data sources without compromising privacy. They also highlight the importance of integrating fairness metrics to ensure the model does not introduce or exacerbate bias, particularly when used in high-stakes environments like finance.

  • FIRST PUBLISHED IN:
  • Devdiscourse
Give Feedback