Academic Journal

Explainability and interpretability are important aspects in ensuring the security of decisions made by intelligent systems (review article)

Bibliographic Details
Title: Explainability and interpretability are important aspects in ensuring the security of decisions made by intelligent systems (review article)
Authors: D. N. Biryukov, A. S. Dudkin
Source: Научно-технический вестник информационных технологий, механики и оптики, Vol 25, Iss 3, Pp 373-386 (2025)
Publisher Information: ITMO University, 2025.
Publication Year: 2025
Collection: LCC:Information technology
Subject Terms: искусственный интеллект, нейронные сети, глубокое обучение, «черный ящик», объяснимость, интерпретируемость, xai, Information technology, T58.5-58.64
Description: The issues of trust in decisions made (formed) by intelligent systems are becoming more and more relevant. A systematic review of Explicable Artificial Intelligence (XAI) methods and tools aimed at bridging the gap between the complexity of neural networks and the need for interpretability of results for end users is presented. A theoretical analysis of the differences between explainability and interpretability in the context of artificial intelligence as well as their role in ensuring the security of decisions made by intelligent systems is carried out. It is shown that explainability implies the ability of a system to generate justifications understandable to humans, whereas interpretability focuses on the passive clarity of internal mechanisms. A classification of XAI methods is proposed based on their approach (preliminary/subsequent analysis: ante hoc/post hoc) and the scale of explanations (local/global). Popular tools, such as Local Interpretable Model Agnostic Explanations, Shapley Values, and integrated gradients, are considered, with an assessment of their strengths and limitations of applicability. Practical recommendations are given on the choice of methods for various fields and scenarios. The architecture of an intelligent system based on the V.K. Finn model and adapted to modern requirements for ensuring “transparency” of solutions, where the key components are the information environment, the problem solver and the intelligent interface, are discussed. The problem of a compromise between the accuracy of models and their explainability is considered: transparent models (“glass boxes”, for example, decision trees) are inferior in performance to deep neural networks, but provide greater certainty of decision-making. Examples of methods and software packages for explaining and interpreting machine learning data and models are provided. It is shown that the development of XAI is associated with the integration of neuro-symbolic approaches combining deep learning capabilities with logical interpretability.
Document Type: article
File Description: electronic resource
Language: English
Russian
ISSN: 2226-1494
2500-0373
Relation: https://ntv.elpub.ru/jour/article/view/461; https://doaj.org/toc/2226-1494; https://doaj.org/toc/2500-0373
DOI: 10.17586/2226-1494-2025-25-3-373-386
Access URL: https://doaj.org/article/1c7a5a2a2bc6417bae0374b8a9a6df6e
Accession Number: edsdoj.1c7a5a2a2bc6417bae0374b8a9a6df6e
Database: Directory of Open Access Journals
Description
ISSN:22261494
25000373
DOI:10.17586/2226-1494-2025-25-3-373-386