International Journal of Innovative Research in Computer and Communication Engineering
ISSN Approved Journal | Impact factor: 8.771 | ESTD: 2013 | Follows UGC CARE Journal Norms and Guidelines
| Monthly, Peer-Reviewed, Refereed, Scholarly, Multidisciplinary and Open Access Journal | High Impact Factor 8.771 (Calculated by Google Scholar and Semantic Scholar | AI-Powered Research Tool | Indexing in all Major Database & Metadata, Citation Generator | Digital Object Identifier (DOI) |
| TITLE | Demystifying the Black Box: A Foundational Survey and Comparative Analysis of Basic Explainable AI (XAI) Techniques |
|---|---|
| ABSTRACT | The unprecedented performance of complex Artificial Intelligence (AI) models, particularly deep neural networks (DNNs), has been accompanied by a critical opacity problem—their decision-making processes are often inscrutable "black boxes." This lack of transparency erodes trust, impedes debugging, and prevents deployment in high-stakes domains like healthcare, finance, and criminal justice, where accountability is paramount. Explainable AI (XAI) has emerged as a vital subfield dedicated to making AI systems more interpretable and understandable to human stakeholders. This research paper provides a comprehensive, pedagogical exploration of fundamental, post-hoc XAI techniques, designed to demystify the core concepts for researchers and practitioners entering the field. We systematically categorize and describe essential methods spanning feature importance techniques (e.g., Permutation Feature Importance, SHAP, LIME), example-based explanations (e.g., Counterfactual Explanations, Prototypes/Criticisms), and visualization techniques for deep learning (e.g., Saliency Maps, Grad-CAM). The study employs a structured methodology of theoretical exposition followed by a consistent empirical evaluation across three benchmark datasets (tabular, image, text) using three common model types (Random Forest, Convolutional Neural Network, BERT). We implemented and compared eight basic XAI techniques, analyzing their outputs not just on predictive accuracy but on key explainability criteria: fidelity (how well the explanation reflects the model's true reasoning), stability (consistency for similar inputs), comprehensibility (ease of human understanding), and actionability. Results revealed a fundamental trade-off: global, model-agnostic methods like SHAP provided robust, consistent feature importance scores but at high computational cost, while local, model-specific methods like Grad-CAM offered intuitive visual explanations for DNNs but were less stable under input perturbations. A critical finding was that no single technique dominated across all criteria, emphasizing the need for a portfolio approach to XAI. Furthermore, we demonstrate that even "basic" techniques, when applied correctly, can reveal model biases, identify spurious correlations, and guide model improvement. The paper concludes that foundational XAI techniques are not merely diagnostic tools but essential components for responsible AI development and deployment. Their mastery is a prerequisite for advancing towards more sophisticated, causally-grounded explanations and for fostering the necessary human-AI collaboration in critical applications. |
| AUTHOR | VEENA MORE, BHARATI H NAIKAWADI, KELLA SOWMYA, SUMITRA M MUDDA Assistant Professor, Department of BCA, A.S.Patil College of Commerce (Autonomous),Vijayapura, Karnataka, India Assistant Professor, Department of Computer Science, A.S.Patil College of Commerce (Autonomous),Vijayapura, Karnataka, India Lecturer, Department of Computer Science, Vijayanagar College, Hosapete, Karnataka, India Assistant Professor, Department of MCA, Guru Nanak Dev Engineering College, BIDAR, Karnataka, India |
| VOLUME | 177 |
| DOI | DOI: 10.15680/IJIRCCE.2025.1312094 |
| pdf/94_Demystifying the Black Box A Foundational Survey and Comparative Analysis of Basic Explainable AI (XAI) Techniques.pdf | |
| KEYWORDS | |
| References | [1] Y. LeCun, Y. Bengio, and G. Hinton, "Deep learning," Nature, vol. 521, no. 7553, pp. 436–444, 2015. [2] F. Doshi-Velez and B. Kim, "Towards a Rigorous Science of Interpretable Machine Learning," arXiv preprint arXiv:1702.08608, 2017. [3] S. Barocas, M. Hardt, and A. Narayanan, Fairness and Machine Learning: Limitations and Opportunities. MIT Press, 2019. [4] D. Gunning, M. Stefik, J. Choi, T. Miller, S. Stumpf, and G. Z. Yang, "XAI—Explainable artificial intelligence," Sci. Robot., vol. 4, no. 37, p. eaay7120, 2019. [5] C. Rudin, "Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead," Nat. Mach. Intell., vol. 1, no. 5, pp. 206–215, 2019. [6] A. B. Arrieta et al., "Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI," Inf. Fusion, vol. 58, pp. 82–115, 2020. [7] T. Miller, "Explanation in artificial intelligence: Insights from the social sciences," Artif. Intell., vol. 267, pp. 1–38, 2019. [8] R. R. Hoffman, S. T. Mueller, G. Klein, and J. Litman, "Metrics for explainable AI: Challenges and prospects," arXiv preprint arXiv:1812.04608, 2018. [9] L. Breiman, "Random Forests," Mach. Learn., vol. 45, no. 1, pp. 5–32, 2001. [10] S. M. Lundberg and S. I. Lee, "A Unified Approach to Interpreting Model Predictions," in Proc. Adv. Neural Inf. Process. Syst. (NeurIPS), 2017, pp. 4765–4774. [11] M. T. Ribeiro, S. Singh, and C. Guestrin, ""Why Should I Trust You?": Explaining the Predictions of Any Classifier," in Proc. ACM SIGKDD Int. Conf. Knowl. Discov. Data Min., 2016, pp. 1135–1144. [12] B. Kim, C. Rudin, and J. A. Shah, "The Bayesian Case Model: A Generative Approach for Case-Based Reasoning and Prototype Classification," in Proc. Adv. Neural Inf. Process. Syst. (NeurIPS), 2014, pp. 1952–1960. [13] S. Wachter, B. Mittelstadt, and C. Russell, "Counterfactual explanations without opening the black box: Automated decisions and the GDPR," Harv. JL & Tech., vol. 31, p. 841, 2017. [14] K. Simonyan, A. Vedaldi, and A. Zisserman, "Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps," in Proc. Int. Conf. Learn. Represent. (ICLR), 2014. [15] R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, and D. Batra, "Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization," in Proc. IEEE Int. Conf. Comput. Vis. (ICCV), 2017, pp. 618–626. [16] A. Vaswani et al., "Attention Is All You Need," in Proc. Adv. Neural Inf. Process. Syst. (NeurIPS), 2017, pp. 5998–6008. [17] C.-K. Yeh, C.-Y. Hsieh, A. Suggala, D. I. Inouye, and P. Ravikumar, "On the (In)fidelity and Sensitivity of Explanations," in Proc. Adv. Neural Inf. Process. Syst. (NeurIPS), 2019, pp. 10965–10976. [18] Q. V. Liao and K. R. Varshney, "Human-Centered Explainable AI (XAI): From Algorithms to User Experiences," arXiv preprint arXiv:2110.10790, 2021. [19] P. Dabkowski and Y. Gal, "Real Time Image Saliency for Black Box Classifiers," in Proc. Adv. Neural Inf. Process. Syst. (NeurIPS), 2017, pp. 6967–6976. [20] M. Sundararajan, A. Taly, and Q. Yan, "Axiomatic Attribution for Deep Networks," in Proc. Int. Conf. Mach. Learn. (ICML), 2017, pp. 3319–3328. |