The row-level extraction dataset, the data-extraction codebook, screening log, including a list of all full-text articles excluded with reasons, and executable analysis notebooks are available at OSF: [https://osf.io/9nykd/](https:/osf.io/9nykd).
* Abudurexiti Y, Han G, Zhang F, Liu L (2025) An explainable unsupervised anomaly detection framework for Industrial Internet of Things. Comput Secur 148:104130. https://doi.org/10.1016/j.cose.2024.104130
Google Scholar
* Acharya DB, Divya B, Kuppan K (2024) Explainable and fair AI: balancing performance in financial and real estate machine learning models. IEEE Access 12:154022-154034. https://doi.org/10.1109/ACCESS.2024.3484409
Google Scholar
* Antwarg L, Miller RM, Shapira B, Rokach L (2021) Explaining anomalies detected by autoencoders using Shapley Additive Explanations. Expert Syst Appl 186:115736. https://doi.org/10.1016/j.eswa.2021.115736
Google Scholar
* Arsenault P-D, Wang S, Patenaude J-M (2025) A survey of explainable artificial intelligence (XAI) in financial time series forecasting. ACM Comput Surv 57:1-37. https://doi.org/10.1145/3729531
Google Scholar
* Awosika T, Shukla RM, Pranggono B (2024) Transparency and privacy: the role of explainable AI and federated learning in financial fraud detection. IEEE Access 12:64551-64560. https://doi.org/10.1109/ACCESS.2024.3394528
Google Scholar
* Ballegeer M, Bogaert M, Benoit DF (2025) Evaluating the stability of model explanations in instance-dependent cost-sensitive credit scoring. Eur J Oper Res 326:630-640. https://doi.org/10.1016/j.ejor.2025.05.039
Google Scholar
* Barredo Arrieta A, Díaz-Rodríguez N, Del Ser J et al (2020) Explainable artificial intelligence (XAI): concepts, taxonomies, opportunities and challenges toward responsible AI. Inf Fusion 58:82-115. https://doi.org/10.1016/j.inffus.2019.12.012
Google Scholar
* Benito D, Quintero C, Aguilar J et al (2024) Explainability analysis: an in-depth comparison between Fuzzy Cognitive Maps and LAMDA. Appl Soft Comput 164:111940. https://doi.org/10.1016/j.asoc.2024.111940
Google Scholar
* Bennetot A, Donadello I, El Qadi El Haouari A et al (2025) A practical tutorial on explainable AI techniques. ACM Comput Surv 57:1-44. https://doi.org/10.1145/3670685
Google Scholar
* Bhakte A, Chakane M, Srinivasan R (2023) Alarm-based explanations of process monitoring results from deep neural networks. Comput Chem Eng 179:108442. https://doi.org/10.1016/j.compchemeng.2023.108442
Google Scholar
* Blagus R, Lusa L (2013) SMOTE for high-dimensional class-imbalanced data. BMC Bioinformatics 14:106. https://doi.org/10.1186/1471-2105-14-106
Google Scholar
* Capuano N, Fenza G, Loia V, Stanzione C (2022) Explainable artificial intelligence in cybersecurity: a survey. IEEE Access 10:93575-93600. https://doi.org/10.1109/ACCESS.2022.3204171
Google Scholar
* Cemernek D, Siddiqi S, Kern R (2024) Effects of class imbalance countermeasures on interpretability. IEEE Access 12:45342-45358. https://doi.org/10.1109/ACCESS.2024.3381536
Google Scholar
* Černevičienė J, Kabašinskas A (2024) Explainable artificial intelligence (XAI) in finance: a systematic literature review. Artif Intell Rev 57:216. https://doi.org/10.1007/s10462-024-10854-8
Google Scholar
* Chadaga K, Prabhu S, Sampathila N, Chadaga R (2023) A machine learning and explainable artificial intelligence approach for predicting the efficacy of hematopoietic stem cell transplant in pediatric patients. Healthcare Analytics 3:100170. https://doi.org/10.1016/j.health.2023.100170
Google Scholar
* Chalapathy R, Chawla S (2019) Deep Learning for Anomaly Detection: A Survey. arXiv preprint
* Chen Yisong Z, Chuqing X Yixin, et al (2024) Year-over-Year developments in financial fraud detection via deep learning: A systematic literature review. https://doi.org/10.48550/arXiv.2502.00201. arXiv:250200201
* Damanik N, Liu C-M (2025) Advanced fraud detection: leveraging K-SMOTEENN and stacking ensemble to tackle data imbalance and extract insights. IEEE Access 13:10356-10370. https://doi.org/10.1109/ACCESS.2025.3528079
Google Scholar
* Darvish E, Jahangoshai Rezaee M, Abbaspour Onari M (2025) A hybrid error correction approach for prediction of credit approval: an explainable artificial intelligence approach. Eng Appl Artif Intell 144:110140. https://doi.org/10.1016/j.engappai.2025.110140
Google Scholar
* Dentamaro V, Giglio P, Impedovo D, Pirlo G (2025) EVolutionary independent DEtermiNistiC explanation. Eng Appl Artif Intell 156:111008. https://doi.org/10.1016/j.engappai.2025.111008
Google Scholar
* Doshi-Velez F, Kim B (2017) Towards a rigorous science of interpretable machine learning. arXiv preprint arXiv:1702.08608. https://doi.org/10.48550/arXiv.1702.08608
* Dritsas E, Trigka M (2025) Machine learning in E-commerce: trends, applications, and future challenges. IEEE Access 13:99048-99067. https://doi.org/10.1109/ACCESS.2025.3572865
Google Scholar
* Dwi Hartomo K, Arthur C, Nataliani Y (2025) A novel weighted loss tabtransformer integrating explainable AI for imbalanced credit risk datasets. IEEE Access 13:31045-31056. https://doi.org/10.1109/ACCESS.2025.3541878
Google Scholar
* Gao Y, Chen R, Qin W et al (2023) Learning from explainable data-driven tunneling graphs: a spatio-temporal graph convolutional network for clogging detection. Autom Constr 147:104741. https://doi.org/10.1016/j.autcon.2023.104741
Google Scholar
* Gao Y, Gu S, Jiang J et al (2024) Going beyond XAI: a systematic survey for explanation-guided learning. ACM Comput Surv 56:1-39. https://doi.org/10.1145/3644073
Google Scholar
* Ghosh S (2025) A novel framework for financial cybersecurity and fraud detection using XAI-RNN-SGRU. IEEE Access 13:88134-88155. https://doi.org/10.1109/ACCESS.2025.3570216
Google Scholar
* Guidotti R, Monreale A, Ruggieri S et al (2019) A survey of methods for explaining black box models. ACM Comput Surv 51:1-42. https://doi.org/10.1145/3236009
Google Scholar
* Hasan M, Rahman MS, Janicke H, Sarker IH (2024) Detecting anomalies in blockchain transactions using machine learning classifiers and explainability analysis. Blockchain: Research and Applications 5:100207. https://doi.org/10.1016/j.bcra.2024.100207
Google Scholar
* Hasan M, Rahman MS, Morshed Chowdhury MJ, Sarker IH (2025) CNN based deep learning modeling with explainability analysis for detecting fraudulent blockchain transactions. Cyber Security and Applications 3:100101. https://doi.org/10.1016/j.csa.2025.100101
Google Scholar
* Hashemi M, Darejeh A, Cruz F (2025) Understanding user preferences in explainable artificial intelligence: a mapping function proposal. ACM Trans Intell Syst Technol 16:1-37. https://doi.org/10.1145/3733837
Google Scholar
* Jairi I, Rekbi A, Ben-Othman S et al (2025) Enhancing particulate matter risk assessment with novel machine learning-driven toxicity threshold prediction. Eng Appl Artif Intell 139:109531. https://doi.org/10.1016/j.engappai.2024.109531
Google Scholar
* Jin C, Zhou J, Xie C et al (2025) Enhancing Ethereum fraud detection via generative and contrastive self-supervision. IEEE Trans Inf Forensics Secur 20:839-853. https://doi.org/10.1109/TIFS.2024.3521611
Google Scholar
* Karnavou E, Cascavilla G, Marcelino G, Geradts Z (2025) I know you’re a fraud: uncovering illicit activity in a Greek bank transactions with unsupervised learning. Expert Syst Appl 288:128148. https://doi.org/10.1016/j.eswa.2025.128148
Google Scholar
* Kumar D, Bakariya B, Verma C, Illes Z (2025) LivXAI-Net: an explainable AI framework for liver disease diagnosis with IoT-based real-time monitoring support. Comput Methods Programs Biomed 270:108950. https://doi.org/10.1016/j.cmpb.2025.108950
Google Scholar
* Lee J, Jung D, Moon J, Rho S (2025) Advanced R-GAN: generating anomaly data for improved detection in imbalanced datasets using regularized generative adversarial networks. Alex Eng J 111:491-510. https://doi.org/10.1016/j.aej.2024.10.084
Google Scholar
* Leuthe D, Mirlach J, Wenninger S, Wiethe C (2024) Leveraging explainable AI for informed building retrofit decisions: insights from a survey. Energy Build 318:114426. https://doi.org/10.1016/j.enbuild.2024.114426
Google Scholar
* Li Z, Zhu Y, Van Leeuwen M (2024) A survey on explainable anomaly detection. ACM Trans Knowl Discov Data 18:1-54. https://doi.org/10.1145/3609333
Google Scholar
* Lin S, Song D, Cao B et al (2025) Credit risk assessment of automobile loans using machine learning-based SHapley additive explanations approach. Eng Appl Artif Intell 147:110236. https://doi.org/10.1016/j.engappai.2025.110236
Google Scholar
* Lundberg SM, Lee SI (2017) A unified approach to interpreting model predictions. In: Advances in neural information processing systems
* Martens D, Hinns J, Dams C et al (2025) Tell me a story! Narrative-driven XAI with Large Language Models. Decis Support Syst 191:114402. https://doi.org/10.1016/j.dss.2025.114402
Google Scholar
* Martins T, de Almeida AM, Cardoso E, Nunes L (2024) Explainable Artificial Intelligence (XAI): a systematic literature review on taxonomies and applications in finance. IEEE Access 12:618-629. https://doi.org/10.1109/ACCESS.2023.3347028
Google Scholar
* Mayrhofer M, Filzmoser P (2023) Multivariate outlier explanations using Shapley values and Mahalanobis distances. https://doi.org/10.1016/j.ecosta.2023.04.003. Econom Stat
* Medianovskyi K, Malakauskas A, Lakstutiene A, Yahia S, Ben (2023) Interpretable machine learning for heterogeneous treatment effect estimators with double ML: a case of access to credit for SMEs. Procedia Comput Sci 225:2163-2172. https://doi.org/10.1016/j.procs.2023.10.207
Google Scholar
* Meza Martínez MA, Nadj M, Langner M et al (2023) Does this explanation help? Designing local model-agnostic explanation representations and an experimental evaluation using eye-tracking technology. ACM Trans Interact Intell Syst 13:1-47. https://doi.org/10.1145/3607145
Google Scholar
* Mohseni S, Zarei N, Ragan ED (2021) A multidisciplinary survey and framework for design and evaluation of explainable AI systems. ACM Trans Interact Intell Syst 11:1-45. https://doi.org/10.1145/3387166
Google Scholar
* Mohsin MT, Nasim NB (2025) Explaining the unexplainable: A systematic review of explainable AI in finance. ArXiv Preprint. https://doi.org/10.48550/arXiv.2503.05966
Google Scholar
* Mozumder MSA, Sakil MBH, Hasan MR et al (2025) Hybrid contrastive learning with attention-based neural networks for robust fraud detection in digital payment systems. IEEE Open Journal of the Computer Society 6:1053-1064. https://doi.org/10.1109/OJCS.2025.3581950
Google Scholar
* Niaz A, Soomro S, Zia H, Choi KN (2024) Increment-CAM: incrementally-weighted class activation maps for better visual explanations. IEEE Access 12:88829-88840. https://doi.org/10.1109/ACCESS.2024.3413859
Google Scholar
* Page MJ, McKenzie JE, Bossuyt PM et al (2021) The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. https://doi.org/10.1136/bmj.n71. BMJ n71
* Pourhabibi T, Ong K-L, Kam BH, Boo YL (2020) Fraud detection: a systematic literature review of graph-based anomaly detection approaches. Decis Support Syst 133:113303. https://doi.org/10.1016/j.dss.2020.113303
Google Scholar
* Sambasivan N, Kapania S, Highfill H et al (2021) Everyone wants to do the model work, not the data work: Data Cascades in High-Stakes AI. In: Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. ACM, New York, NY, USA, pp 1-15
* Samek W, Wiegand T, Müller K-R (2017) Explainable artificial intelligence: understanding, visualizing and interpreting deep learning models. ITU J (Geneva) 1:39-48
Google Scholar
* Sasikumar M, Raj ED (2025) Investigating explainability of deep learning models for sequential data on stock price prediction. Procedia Comput Sci 258:4190-4201. https://doi.org/10.1016/j.procs.2025.04.669
Google Scholar
* Shi J, Siebes APJM, Mehrkanoon S (2025) TransCORALNet: a two-stream transformer CORAL networks for supply chain credit assessment cold start. Expert Syst Appl 282:127581. https://doi.org/10.1016/j.eswa.2025.127581
Google Scholar
* Slack D, Hilgard S, Jia E et al (2020) Fooling LIME and SHAP. In: Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society. ACM, New York, NY, USA, pp 180-186
* Sokolovsky A, Arnaboldi L, Bacardit J, Gross T (2023) Interpretable trading pattern designed for machine learning applications. Machine Learning with Applications 11:100448. https://doi.org/10.1016/j.mlwa.2023.100448
Google Scholar
* Tan Y, Lu Y, Wang L (2025) Flight delay dynamics: unraveling the impact of airport-network-spilled propagation on airline on-time performance. Decis Support Syst 196:114494. https://doi.org/10.1016/j.dss.2025.114494
Google Scholar
* Temraz M, Keane MT (2022) Solving the class imbalance problem using a counterfactual method for data augmentation. Machine Learning with Applications 9:100375. https://doi.org/10.1016/j.mlwa.2022.100375
Google Scholar
* Thuy A, Benoit DF (2024) Explainability through uncertainty: trustworthy decision-making with neural networks. Eur J Oper Res 317:330-340. https://doi.org/10.1016/j.ejor.2023.09.009
Google Scholar
* Vorobyev I, Krivitskaya A (2022) Reducing false positives in bank anti-fraud systems based on rule induction in distributed tree-based models. Comput Secur 120:102786. https://doi.org/10.1016/j.cose.2022.102786
Google Scholar
* Xu C, Zhu P, Fan Z, Ma Y (2025) Evaluating example-based explainable artificial intelligence methods for breast cancer diagnosis. Appl Soft Comput 177:113222. https://doi.org/10.1016/j.asoc.2025.113222
Google Scholar
* Yousefimehr B, Ghatee M (2025) A distribution-preserving method for resampling combined with LightGBM-LSTM for sequence-wise fraud detection in credit card transactions. Expert Systems with Applications 262:125661. https://doi.org/10.1016/j.eswa.2024.125661
Google Scholar
* Zhou Y, Li H, Xiao Z, Qiu J (2023) A user-centered explainable artificial intelligence approach for financial fraud detection. Fin Res Lett 58:104309. https://doi.org/10.1016/j.frl.2023.104309
Google Scholar

