The WHY?

Artificial intelligence (AI) is a powerful technology that can perform tasks that normally require human intelligence, such as understanding natural language, recognizing images, playing games, and generating content. However, many AI systems are based on complex algorithms that are difficult to understand and explain, especially for non-experts. These systems are often referred to as “black boxes”, because their inner workings are hidden from view.

This lack of transparency and interpretability can pose challenges for users, developers, regulators, and society at large. For example, how can we trust an AI system that makes decisions that affect our lives, such as diagnosing diseases, approving loans, or driving cars? How can we ensure that an AI system is fair, ethical, and accountable? How can we debug and improve an AI system that makes errors or behaves unexpectedly?

The WHAT?

These questions motivate the need for Explainable AI (XAI), a set of processes and methods that allows human users to comprehend and trust the results and output created by machine learning algorithms. Explainable AI is used to describe an AI model, its expected impact and potential biases. It helps characterize model accuracy, fairness, transparency and outcomes in AI-powered decision making.

Explainable AI is crucial for building trust and confidence in AI systems, especially in domains where human lives or values are at stake. It can also help developers improve and optimize their models, regulators ensure compliance with standards and regulations, and users understand and control their interactions with AI systems.

The WHEN?

The concept of XAI is not new. In fact, it has been around for decades. It was first introduced in 2004 by Van Lent et al.[1], in the context of the non-player characters [NPCs] in video games. The authors proposed a framework to extract key events and decision points from the playback log to allow the NPC AI controlled soldiers to explain their behavior in response to the questions selected from the XAI menu. Of course, this paper was not the first one to motivate this concept, but it was the first one to introduce the term “XAI” as per my knowledge.

The HOW?

Explainable AI can be achieved in different ways, depending on the type and complexity of the AI system, the level and purpose of the explanation, and the target audience. Some common approaches include:

The WHY NOW?

One domain where Explainable AI is particularly relevant is natural language processing (NLP), which is the branch of AI that deals with understanding and generating natural language. NLP has seen remarkable advances in recent years, thanks to the development of deep learning models such as ChatGPT.

ChatGPT is being used for countless applications, such as chatbots, text summarization, text generation, or text completion and so much more - ossinsight has a great list of ChatGPT applications and this is just the tip of the iceberg.

However, ChatGPT is also a black-box model that is hard to interpret and explain. It is not clear how ChatGPT generates its text outputs, what factors influence its choices, what biases or errors it might have, or how it can be improved or controlled. Moreover, ChatGPT can potentially generate harmful or misleading content that can affect users’ emotions, opinions, or actions.

In fact, these concerns are shared by leading tech luminaries such as Elon Musk, Steve Wozniak and others where they have signed an open letter title “Pause Giant AI Experiments: An Open Letter”. The letter calls for a moratorium on the development of AI systems that have “human-competitive intelligence” and “pose profound risks to society and humanity”. On a side note, this has been a very interesting topic of discussion and I recommend reading about it. Some of the good articles I found are:

In the light of all this, tools like Explainable AI can play an important role. By providing explanations for how ChatGPT works and why it produces certain outputs, Explainable AI can help users understand and trust ChatGPT’s capabilities and limitations. It can also help developers monitor and improve ChatGPT’s performance and quality. Furthermore, explainable AI can help regulators ensure that ChatGPT complies with ethical principles and social norms.

P.S. I am not an expert on XAI. This is just what I have been reading in the past few weeks. I am sure there are many more things to be said about XAI and its relevance in the modern AI research. I would love to hear your thoughts on this in the comments below.

References

[1] Van Lent, Michael, William Fisher, and Michael Mancuso. “An explainable artificial intelligence system for small-unit tactical behavior.” Proceedings of the national conference on artificial intelligence. Menlo Park, CA; Cambridge, MA; London; AAAI Press; MIT Press; 1999, 2004.

[2] Minh, Dang, et al. “Explainable artificial intelligence: a comprehensive review.” Artificial Intelligence Review (2022): 1-66.

[3] Schwalbe, Gesina, and Bettina Finzel. “A comprehensive taxonomy for explainable artificial intelligence: a systematic survey of surveys on methods and concepts.” Data Mining and Knowledge Discovery (2023): 1-59.

[4] Das, Arun, and Paul Rad. “Opportunities and challenges in explainable artificial intelligence (xai): A survey.” arXiv preprint arXiv:2006.11371 (2020).