Developing interpretable models and techniques for explainable AI in decision-making
Downloads
Published
DOI:
https://doi.org/10.58414/SCIENTIFICTEMPER.2023.14.4.39Keywords:
Explainable AI interpretable AI models, Cybersecurity, Attack types, Decision-making, Botanical classification.Dimensions Badge
Issue
Section
License
Copyright (c) 2023 The Scientific Temper

This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.
The rapid proliferation of artificial intelligence (AI) technologies across various industries and decision-making processes has undeniably transformed the way of approaching complex problems and tasks. AI systems have proven their prowess in areas such as healthcare, finance, and autonomous systems, revolutionizing how decisions are made. Nevertheless, this proliferation of AI has raised critical concerns regarding the transparency, accountability, and fairness of these systems, as many of the state-of-the-art AI models often resemble complex black boxes. These intricate models, particularly deep learning neural networks, harbor non-linear relationships that are difficult for human users to decipher, thereby raising concerns about bias, fairness, and overall trustworthiness in AI-driven decisions. The urgency of this issue is underscored by the realization that AI should not merely be accurate; it should also be interpretable. Explainable AI (XAI) has emerged as a vital field of research, emphasizing the development of models and techniques that render AI systems comprehensible and transparent in their decision-making processes. This paper investigates into the relevance and significance of XAI across various domains, including healthcare, finance, and autonomous systems, where the ability to understand the rationale behind AI decisions is paramount. In healthcare, where AI assists in diagnosis and treatment, the interpretability of AI models is crucial for clinicians to make informed decisions. In finance, applications like credit scoring and investment analysis demand transparent AI to ensure fairness and accountability. In the realm of autonomous systems, transparency is indispensable to guarantee safety and compliance with regulations. Moreover, government agencies in areas such as law enforcement and social services require interpretable AI to maintain ethical standards and accountability. This paper also highlights the diverse array of research efforts in the XAI domain, spanning from model-specific interpretability methods to more general approaches aimed at unveiling complex AI models. Interpretable models like decision trees and rule-based systems have gained attention for their inherent transparency, while integrating explanation layers into deep neural networks strives to balance accuracy with interpretability. The study emphasizes the significance of this burgeoning field in bridging the gap between AI's advanced capabilities and human users' need for comprehensible AI systems. It seeks to contribute to this field by exploring the design, development, and practical applications of interpretable AI models and techniques, with the ultimate goal of enhancing the trust and understanding of AI-driven decisions.Abstract
How to Cite
Downloads
Similar Articles
- T. Kanimozhi, V. Rajeswari, R. Suguna, J. Nirmaladevi, P. Prema, B. Janani, R. Gomathi, RWHO: A hybrid of CNN architecture and optimization algorithm to predict basal cell carcinoma skin cancer in dermoscopic images , The Scientific Temper: Vol. 15 No. 02 (2024): The Scientific Temper
- M. A. Shanti, Optimizing predictive accuracy: A comparative study of feature selection strategies in the healthcare domain , The Scientific Temper: Vol. 15 No. spl-1 (2024): The Scientific Temper
- D. Prabakar, Santhosh Kumar D.R., R.S. Kumar, Chitra M., Somasundaram K., S.D.P. Ragavendiran, Narayan K. Vyas, Task offloading and trajectory control techniques in unmanned aerial vehicles with Internet of Things – An exhaustive review , The Scientific Temper: Vol. 14 No. 04 (2023): The Scientific Temper
- Amala Deepa V., T. Lucia Agnes Beena, Enhancing data imputation in complex datasets using Lagrange polynomial interpolation and hot-deck fusion , The Scientific Temper: Vol. 16 No. 01 (2025): The Scientific Temper
- Abhishek Dwivedi, Shekhar Verma, SCNN Based Classification Technique for the Face Spoof Detection Using Deep Learning Concept , The Scientific Temper: Vol. 13 No. 02 (2022): The Scientific Temper
- B. Kalpana, P. Krishnamoorthy, S. Kanageswari, Anitha J. Albert, Machine learning approaches for predicting species interactions in dynamic ecosystems , The Scientific Temper: Vol. 15 No. 03 (2024): The Scientific Temper
- Milindkumar N. Dandale, Amar P. Yadav, P. S. K. Reddy, Seema G. Kadu, Madhusudana T, Manthan S. Manavadaria, Deep learning enhanced drug discovery for novel biomaterials in regenerative medicine utilizing graph neural network approach for predicting cellular responses , The Scientific Temper: Vol. 15 No. 01 (2024): The Scientific Temper
- Shamba Gowda, AR Chethan Kumar, S. Srinivasaragavan, Mapping of research productivity on forestry research in India: A scientometric study , The Scientific Temper: Vol. 15 No. spl-1 (2024): The Scientific Temper
- Rashmika Vaghela, Dileep Labana, Kirit Modi, Efficient I3D-VGG19-based architecture for human activity recognition , The Scientific Temper: Vol. 14 No. 04 (2023): The Scientific Temper
- K. Sreenivasulu, Sampath S, Arepalli Gopi, Deepak Kartikey, S. Bharathidasan, Neelam Labhade Kumar, Advancing device and network security for enhanced privacy , The Scientific Temper: Vol. 14 No. 04 (2023): The Scientific Temper
<< < 2 3 4 5 6 7 8 9 10 11 > >>
You may also start an advanced similarity search for this article.
Most read articles by the same author(s)
- Pravin P. Adivarekar1, Amarnath Prabhakaran A, Sukhwinder Sharma, Divya P, Muniyandy Elangovan, Ravi Rastogi, Automated machine learning and neural architecture optimization , The Scientific Temper: Vol. 14 No. 04 (2023): The Scientific Temper
- Balaji V, Purnendu Bikash Acharjee, Muniyandy Elangovan, Gauri Kalnoor, Ravi Rastogi, Vishnu Patidar, Developing a semantic framework for categorizing IoT agriculture sensor data: A machine learning and web semantics approach , The Scientific Temper: Vol. 14 No. 04 (2023): The Scientific Temper