Explainable Deep Learning Models for Trustworthy Decision Support in High-Stakes Data Science Applications
DOI:
https://doi.org/10.61453/jods.v20260102Keywords:
Explainable AI, Deep Learning, Trustworthy AI, Decision Support Systems, Model InterpretabilityAbstract
Deep learning models are increasingly deployed in high-stakes domains such as healthcare, finance, and public decision systems, where predictive errors and opaque reasoning can lead to significant societal consequences. Despite their superior predictive capabilities, most deep learning systems remain black-box models, limiting transparency, regulatory compliance, and user trust. Existing explainable artificial intelligence (XAI) approaches often function as post-hoc add-ons and rarely integrate explanation stability into the model optimization process. To address this gap, this study proposes a unified explainable deep learning framework that embeds model-agnostic and model-specific interpretability techniques directly into a multi-objective optimization pipeline. The framework jointly optimizes predictive performance, computational efficiency, and explanation stability under predefined deployability constraints. Experiments were conducted on benchmark datasets representing high-stakes risk assessment and resource allocation scenarios using MLP and attention-based architectures. Results show that explainability-integrated models achieved a stability score of 0.89 (vs. 0.72 baseline) and reduced representation shift by 39%, while maintaining competitive predictive performance (ROC-AUC up to 0.901, <1.2% degradation). Human-centered evaluation further demonstrated a significant increase in trust scores (4.18 vs. 3.12, p < 0.001). These findings indicate that embedding explainability as a structural design principle enhances robustness and trustworthiness without sacrificing accuracy. The study contributes a deployable framework for responsible AI in high-stakes decision support systems
References
Abbas, M. J., Khan, M. A., Hamza, A., Alsenan, S., Rehman, A., Baili, J., & Zhang, Y. (2025). C3BAM-XAI: Convolutional Block Attention Module Enhanced Explainable Artificial Intelligence-Based Parkinson’s Disease Stage Classification. Cognitive Computation 2025 17:3, 17(3), 111-. https://doi.org/10.1007/s12559-025-10472-8
Anand, S., Sharma, A., Natarajan, B., Slathia, A. S., Rathi, A., Behara, K. P., & Elakkiya, R. (2025). CHASHNIt for enhancing skin disease classification using GAN augmented hybrid model with LIME and SHAP based XAI heatmaps. Scientific Reports 2025 15:1, 15(1), 31138-. https://doi.org/10.1038/s41598-025-13647-3
Assis, A., Dantas, J., & Andrade, E. (2024). The performance-interpretability trade-off: a comparative study of machine learning models. Journal of Reliable Intelligent Environments 2024 11:1, 11(1), 1-. https://doi.org/10.1007/s40860-024-00240-0
Azizi, M., Aickelin, U., A. Khorshidi, H., & Baghalzadeh Shishehgarkhaneh, M. (2023). Energy valley optimizer: a novel metaheuristic algorithm for global and engineering optimization. Scientific Reports 2023 13:1, 13(1), 226-. https://doi.org/10.1038/s41598-022-27344-y
Azhar, M., Amjad, A., Dewi, D. A., & Kasim, S. (2025). A Systematic Review and Experimental Evaluation of Classical and Transformer-Based Models for Urdu Abstractive Text Summarization. Information, 16(9). https://doi.org/10.3390/info16090784
Guan, Y., Liao, H., Li, Z., Hu, J., Yuan, R., Li, Y., Zhang, G., & Xu, C. (2024). World Models for Autonomous Driving: An Initial Survey. IEEE Transactions on Intelligent Vehicles. https://doi.org/10.1109/TIV.2024.3398357
Gupta, C., Gill, N. S., Gulia, P., Kumar, A., Karamti, H., & Moges, D. M. (2025). An optimized YOLO NAS based framework for realtime object detection. Scientific Reports 2025 15:1, 15(1), 32903-. https://doi.org/10.1038/s41598-025-17919-w
Ibrahim, R., & Omair Shafiq, M. (2023). Explainable Convolutional Neural Networks: A Taxonomy, Review, and Future Directions. ACM Computing Surveys, 55(10). https://doi.org/10.1145/3563691
Kosasih, E. E., Papadakis, E., Baryannis, G., & Brintrup, A. (2024). A review of explainable artificial intelligence in supply chain management using neurosymbolic approaches. International Journal of Production Research, 62(4), 1510–1540. https://doi.org/10.1080/00207543.2023.2281663
Kruschel, Sven, Hambauer, Nico, Weinzierl, Sven, Zilker, Sandra, Kraus, Mathias, Zschech, Patrick, Kruschel, S, Hambauer, Á. N., Kraus, Á. M., Hambauer, N, Kraus, M, Weinzierl, S, Zilker, Á. S., Zilker, S, & Zschech, P. (2025). Challenging the Performance-Interpretability Trade-Off: An Evaluation of Interpretable Machine Learning Models. Business & Information Systems Engineering 2025, 1–25. https://doi.org/10.1007/s12599-024-00922-2
Masud, M. T., Keshk, M., Moustafa, N., Linkov, I., & Emge, D. K. (2025). Explainable Artificial Intelligence for Resilient Security Applications in the Internet of Things. IEEE Open Journal of the Communications Society, 6, 2877–2906. https://doi.org/10.1109/OJCOMS.2024.3413790
Ning, E., Wang, Y., Wang, C., Zhang, H., & Ning, X. (2024). Enhancement, integration, expansion: Activating representation of detailed features for occluded person re-identification. Neural Networks, 169, 532–541. https://doi.org/10.1016/j.neunet.2023.11.003
Payrovnaziri, S. N., Chen, Z., Rengifo-Moreno, P., Miller, T., Bian, J., Chen, J. H., Liu, X., & He, Z. (2020). Explainable artificial intelligence models using real-world electronic health record data: A systematic scoping review. Journal of the American Medical Informatics Association, 27(7), 1173–1185. https://doi.org/10.1093/jamia/ocaa053
Sahoh, B., & Choksuriwong, A. (2023). The role of explainable Artificial Intelligence in high-stakes decision-making systems: a systematic review. Journal of Ambient Intelligence and Humanized Computing 2023 14:6, 14(6), 7827–7843. https://doi.org/10.1007/s12652-023-04594-w
Sancar, N., Onakpojeruo, E. P., Inan, D., & Uzun Ozsahin, D. (2023). Adaptive Elastic Net Based on Modified PSO for Variable Selection in Cox Model with High-Dimensional Data: A Comprehensive Simulation Study. IEEE Access, 11, 127302–127316. https://doi.org/10.1109/ACCESS.2023.3329386
Schofield, A., Wu, S., De Volo, T. B., Kuze, T., Gomez, A., & Sultana, S. (2025). “My Very Subjective Human Interpretation”: Domain Expert Perspectives on Navigating the Text Analysis Loop for Topic Models. Proceedings of the ACM on Human-Computer Interaction, 9(1), 30. https://doi.org/10.1145/3701201
Shan, B., Borghetti, A., Zheng, W., & Guo, Q. (2025). Explainable AI-Based Short-Term Voltage Stability Mechanism Analysis: Explainability Measure and Stability-Oriented Preventive Control. CSEE Journal of Power and Energy Systems, 11(6), 2673–2683. https://doi.org/10.17775/CSEEJPES.2025.02850
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2026 Journal of Data Science

This work is licensed under a Creative Commons Attribution 4.0 International License.