AI Explainability Bridge MachineVision NLP
AI Explainability Bridge MachineVision NLP
net/publication/349467250
CITATIONS READS
13 845
1 author:
Mourad Oussalah
University of Oulu
274 PUBLICATIONS 3,227 CITATIONS
SEE PROFILE
All content following this page was uploaded by Mourad Oussalah on 15 November 2022.
Mourad Oussalah(B)
1 Introduction
Aided by the advances in computer system computational performances and learning
system theory, the success of machine learning methods in the last decade has been
phenomenal in various fields, especially, computer vision and natural language process-
ing, which enhanced the prediction and automated decision-making capabilities. This
has taken machine intelligence and artificial intelligence (AI) to a new frontier that wit-
nessed the emergence of new industry standard (e.g., industry 4.0) and human-computer
interaction modes where a machine guides medical diagnosis systems, creates recom-
mender systems, makes investment decisions and instructs driverless vehicles. On the
other hand, the state-of-the-art systems in many AI applications use ensembles of deep
neural networks that are even more difficult to interpret, even for skilled programmer
users. This negatively impacts trust. For instance, during the PwC’s 2017 Global CEO
Survey [1], although it is acknowledged the substantial increase of AI market to more
than $15 trillion, 67% of the business leaders believe that this will impact negatively
stakeholder trust levels in their industry in the next five years. This fosters the emer-
gence of explainable AI research that seeks to ensure transparency and interpretability
of machine learning and AI based algorithms. Indeed, many applications have seen a
huge increase in demand for transparency from the various stakeholders involved at
various levels of product pipeline. For instance, in precision-medicine, explanation is
required to support system diagnosis outcome and clinical investigation; in finance and
2 Background
We will adopt in this paper Gunning definition of Explainable Artificial Intelligence
(XAI) [12]: “XAI will create a suite of machine learning techniques that enables human
users to understand, appropriately trust, and effectively manage the emerging generation
of artificially intelligent partners”, see Fig. 1.
This definition brings together two concepts; namely, understanding and trust that
need to be addressed. Such concepts are ultimately linked to several other aspects
that overlap with cognitive operations of understanding or comprehension tasks. This
includes for instance causality, transferability, informativeness, fairness and confidence
[13].
Regardless of the type of applications involved or the system inputs, the explanation
differs according to the underlying chosen criterion. Especially, explanation methods
and techniques for ML interpretability can be classified according to different criteria.
• Type of explanation
This includes Feature summary (either through visualization or textual input), Model
internals (model specific), Data point (output data points that make the model inter-
pretable), Surrogate intrinsically interpretable model—through approximation of ML
model either locally or globally with an intrinsically interpretable model.
• Simulatability
This refers to comprehending how the model makes decisions, grounded on a holistic
view of the data features and individual components (e.g., weights, parameters) in order
to simulate the overall system.
AI Explainability 261
on the same setting is growing. This motivates the development of unified frameworks
in XAI to handle both types of inputs as well.
into image using image generator. Methods based on conditional Generative Adversar-
ial Network (GAN) have been employed in several text-to-image synthesis tasks and
competitions [56, 57] and tested on large scale dataset such as birds [58], flowers [59],
MS-COCO [60]. In this regard, the task of image generation is viewed as a problem
of translating semantic labels into pixels. Nevertheless, the complexity of the reason-
ing cannot be ignored. Especially, learning a direct mapping from text to image is not
straightforward and layout generator requires several constraints to enhance its practi-
cality due to the vast amount of possibilities of potential image candidates that fit a given
textual utterance, see, for instance, the example in Figs. 3, 4.
Fig. 4. Architecture for image generator conditioned on the text description and semantic layout
generator in [78].
Since the emergence of Latent Dirichlet Allocation (LDA) [70], the task of automatic
discovery of topics in a textual document has seen a new landmark. In essence, LDA
introduces sparse Dirichlet prior distributions over document-topic and topic-word dis-
tributions, encoding the intuition that documents cover a small number of topics and
that topics often use a small number of words. Topic models are a form of unsupervised
machine learning, in that the topics and mixture parameters are unknown and inferred
solely from the data where each topic is represented by its N most probable words.
Humans can judge whether words of a given topic (cluster) form interpretable concept
(s). Therefore, it is important to seek automatic alternative to measure the interpretabil-
ity of the outputted set of words of each topic. A commonly employed approach is
based on the co-occurrence analysis, stipulating that words that have high frequency
of co-occurrence (either within the document under investigation or in a more wider
corpus) would indicate high coherence and relatedness, as for words caught and fever
for instance [71]. The development of word embedding promoted by Google researchers
has also promoted the so called embedded topic model [72] where each word is mod-
elled as a categorical distribution whose natural parameter is the inner product between
a word embedding and an embedding of its assigned topic. This has shown to discover
interpretable topics even with large vocabulary. On the other hand, the development of
interactive topic modelling [73], where more interaction modes with system output are
enabled, offers a nice setting to apply a range of visualization tools developed in the
context of explainable AI for this purpose.
268 M. Oussalah
5 Conclusion
Explanation methods are a promising approach to leverage hidden knowledge about
the workings of neural networks and black-box systems, promoting transparency and
interpretability of the results in the light of the new data protection EU directive on the
“right of explanation”. This papers attempted to review the state of art of explainability
methods focusing on intertwine between image processing and natural language pro-
cessing fields in a way to promote fruitful development of new explanation framework.
Especially, the paper highlights the implicit link between the two research fields through,
e.g., automatic image annotation, visual question-answer systems, Text-To-Image gen-
eration, multimedia analytics in addition to the overall input-output like system analysis.
On the other hand, this review has also identified several NLP research fields that would
benefit from visual explainability based approach. This includes, wordsense disambigua-
tion, sentiment analysis, argumentation theory, automatic text summarization and topical
modelling.
There are several interesting future research directions to explore further. An inter-
esting direction is semi-supervised learning of the model using a large set of partially
annotated data. For instance, we can exploit a small number of fully annotated images
and a large number of partially annotated images (e.g. images with only text descrip-
tions), which allows the developed model to exploit large-scale datasets, such as the
Google Conceptual Caption dataset. The paper also opens up new research directions in
multimedia analytics, text summarization and abstract argumentation logic.
AI Explainability 269
Acknowledgment. This work is partly supported by the H2020 YoungRes (# 823701) project,
which is gratefully acknowledged.
References
1. Oxborough, C., Cameron, E.: Explainable AI, PWC report (2020). https://www.pwc.co.uk/
services/risk-assurance/insights/explainable-ai.html Accessed July 2020
2. Grice, H.P.: Logic and Conversation. In: Syntax and Semantics 3: Speech arts, pp. 41–58
(1975)
3. Conati, C., Porayska-Pomsta, K., Mavrikis, M.: AI in education needs interpretable machine
learning: lessons from open learner modelling. arXiv preprint arXiv:1807.00154 (2018)
4. Goodman, B., Flaxman, S.: European union regulations on algorithmic decision making and
a “right to explanation.” AI Mag. 38(3), 50 (2017)
5. Abdul, A., Vermeulen, J., Wang, D., Lim, B.Y., Kankanhalli, M.: Trends and trajectories for
explainable, accountable and intelligible systems: an HCI research agenda. In Proceedings of
the 2018 CHI. Association for Computing Machinery, Montreal, Canada (2018). https://doi.
org/10.1145/3173574.3174156
6. Doshi-velez, F., Kim, B.: A Roadmap for a Rigorous Science of Interpretability. CoRR,
abs/1702.08608 (2017). https://arxiv.org/abs/1702.08608
7. Samek, W., Wiegand, T., Müller, K.R.: Explainable artificial intelligence: understanding,
visualizing and interpreting deep learning models. ITU J. ICT Discoveries Spec. (1), 39–48
(2017)
8. Lewis, D.: Causal Explanation. In: Philosophical Papers. Vol II. Oxford University Press,
New York, Chapter Twenty two, pp. 214–240 (1986)
9. Halpern, J.Y., Pearl, J.: Causes and explanations : a structural-model approach. Part II :
Explanations. 56(4), 889–911 (2005). https://doi.org/10.1093/bjps/axi148
10. Miller, T.: Explanation in artificial intelligence: Insights from the social sciences. Artif. Intell.
267, 1–38 (2019). https://doi.org/10.1016/j.artint.2018.07.007
11. Arrieta, Del Ser et al. Explainable artificial intelligence (XAI): concepts, taxonomies,
opportunities and challenges toward responsible AI. Inf. Fusion 58, 82–115 (2019)
12. Gunning, D.: Explainable artificial intelligence (XAI ). Technical report, pp. 1–18 (2017)
13. Lipton, Z.C.: The mythos of model interpretability. Queue 16(3), 30:31–30:57 (2018)
14. Kim, B., Doshi-Velez, F.: Introduction to interpretable machine learning. In Proceedings of
the CVPR 2018 Tutorial on Interpretable Machine Learning for Computer Vision, Salt Lake
City, UT, USA (2018)
15. Jolliffe, I.: Principal component analysis. In International Encyclopedia of Statistical Science,
pp. 1094–1096. Springer, Berlin (2011). https://doi.org/10.1007/978-3-642-04898-2_455
16. Maaten, L.Y.D., Hinton, G.: Visualizing data using t-SNE. J. Mach. Learn. Res. 9, 2579–2605
(2008)
17. Kim, B., Khanna, R., Koyejo, O.O.: Examples are not enough, learn to criticize! Criticism
for interpretability. In: Advances in Neural Information Processing Systems. MIT Press:
Cambridge, pp. 2280–2288 (2016)
18. Hartigan, J.A., Wong, M.A.: A k-means clustering algorithm. J. R. Stat. Soc. Ser. C (Appl.
Stat.) 28, 100–108 (1979)
19. Adadi, A., Berrada, M.: Peeking inside the black-box: a survey on explainable artificial
intelligence (XAI). IEEE Access 6, 52138–52160 (2018)
20. Molnar, C.: Interpretable machine learning (2019). https://christophm.github.io/interpretable-
ml-book/. Accessed July 2020
270 M. Oussalah
21. Yang, C., Rangarajan, A., Ranka, S.: Global model interpretation via recursive partitioning
(2018). arXiv:1802.04253
22. Nguyen, A., Dosovitskiy, A., Yosinski, J., Brox, T., Clune, J.: Synthesizing the preferred
inputs for neurons in neural networks via deep generator networks. Proc. Adv. Neural Inf.
Process. Syst. (NIPS), 3387–3395 (2016)
23. Ribeiro, M.T., Singh, S., Guestrin, C.: Why should I trust you?: Explaining the predic-
tions of any classifier, In: Proceedings of 22nd ACM SIGKDD International Conference
on Knowledge Discovery Data Mining (2016)
24. Lundberg, S.M., Lee, S.L.: A unified approach to interpreting model predictions. Proc. Adv.
Neural Inf. Process. Syst. 4768–4777 (2017)
25. Cortez, P., Embrechts, M.J.: Opening black box data mining models using sensitivity analysis.
In: Proceedings of IEEE Symposium on Computational Intelligence Data Mining (CIDM),
pp. 341–348 (2011)
26. Smilkov, D., Thorat, N., Kim, B., ViØgas F., Wattenberg, M.: SmoothGrad: removing noise
by adding noise (2017). https://arxiv.org/abs/1706.03825
27. Green D.P., Kern, H.L.: Modeling heterogeneous treatment effects in large-scale experiments
using Bayesian additive regression trees. In: Proceedings of Annual Summer Meeting Society
for Political Methodology, pp. 1–40 (2010)
28. Bastani, O., Kim, C., Bastani, H.: Interpretability via model extraction (2017). https://arxiv.
org/abs/1706.09773
29. Thiagarajan, J.J., Kailkhura, B., Sattigeri,P., Ramamurthy, K.N.: TreeView: peeking into deep
neural networks via feature-space partitioning. (2016) https://arxiv.org/abs/1611.07429
30. Goldstein, A., Kapelner, A., Bleich, J., Pitkin, E.: Peeking inside the black box: visualizing
statistical learning with plots of individual conditional expectation, J. Comput. Graph. Statist.
24(1), 44–65 (2015)
31. Frank, E., Witten, I.H.: Generating accurate rule sets without global optimization. In: ICML
1998, pp. 144–151 (1998)
32. Robnik-Šikonja, M., Kononenko, L.: Explaining classifications for individual instances. IEEE
Trans. Knowl. Data Eng. 20(5), 589–600 (2008)
33. Etchells, T.A., Lisboa, P.J.G.: Orthogonal search-based rule extraction (OSRE) for trained
neural networks: a practical and efficient approach. IEEE Trans. Neural Netw. 17(2), 374–384
(2006)
34. Tan, S., Caruana, R., Hooker, G., Lou, Y.: Detecting bias in black-box models using transparent
model distillation. https://arxiv.org/abs/1710.06169 (2018)
35. Xu, K., Park, D.H., Yi, D.H., Sutton, C.: Interpreting deep classifier by visual distillation of
dark knowledge (2018). https://arxiv.org/abs/1803.04042
36. Cortez, P., Embrechts, M.J.: Using sensitivity analysis and visualization techniques to open
black box data mining models. Inf. Sci. 225, 1–7 (2013)
37. Fisher, A., Rudin, C., Dominici, F.: Model class reliance: variable importance measures for
any machine learning model class, from the ‘Rashomon’ perspective (2018). https://arxiv.org/
abs/1801.01489
38. Casalicchio, G., Molnar, C., Bischl, B.: Visualizing the feature importance for black box
models (2018). https://arxiv.org/abs/1804.06620
39. Schetinin, V., et al.: Confident interpretation of Bayesian decision tree ensembles for clinical
applications. IEEE Trans. Inf. Technol. Biomed. 11(3), 312 (2007)
40. Hara, S., Hayashi, K.: Making tree ensembles interpretable (2016). https://arxiv.org/abs/1606.
05390
41. Ribeiro, M.T., Singh, S., Guestrin, C.: Anchors: high-precision model-agnostic explanations.
In: Proceedings of AAAI Conference on Artificial Intelligence, pp. 1–9 (2018)
AI Explainability 271
42. García, S., Fernández, A., Herrera, F.: Enhancing the effectiveness and interpretability of deci-
sion tree and rule induction classifiers with evolutionary training set selection over imbalanced
problems. Appl. Soft Comput. 9(4), 1304–1314 (2009)
43. Wang, F., Rudin, C.: Falling rule lists. In: Proceedings of 18th International Confrence
on Artificial Intelligence on Statistics (AISTATS), San Diego, CA, USA: JMLR W&CP,
pp. 1013–1022 (2015)
44. Ras, G., Van Gerven, M., Haselager, P.: Explanation methods in deep learning: Users, values,
concerns and challenges (2018). https://arxiv.org/abs/1803.07517
45. Johansson, U., König, R., Niklasson, I.: The truth is in there—rule extraction from opaque
models using genetic programming. In: Proceedings of FLAIRS Conference, pp. 658–663
(2004)
46. Wachter, S., Mittelstadt, B., Russell, C.: Counterfactual explanations without opening the
black box: automated decisions and the GDPR (2017). https://arxiv.org/abs/1711.00399
47. Bach, S., Binder, A., Montavon, G., Klauschen, F., Müller, K.R., Samek, W.: On pixel-wise
explanations for non-linear classifier decisions by layer-wise relevance propagation. PLoS
ONE 10(7), e0130140 (2015)
48. Thiagarajan, J.J., Kailkhura, B., Sattigeri,P., Ramamurthy, K.N.: TreeView: peeking into deep
neural networks via feature-space partitioning (2016). arXiv preprint arXiv:1611.07429
49. Wang, T., Rudin,C., Velez-Doshi, F., Liu, Y., Klamp, E., MacNeille, P.: Bayesian rule sets for
interpretable classification. In: IEEE 16th International Conference on Data Mining (ICDM),
pp. 1269–1274 (2016)
50. Boz, O.: Extracting decision trees from trained neural networks. In: Proceedings of the Eighth
ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 456–
461. ACM (2002)
51. Shrikumar, A., Greenside, P., Shcherbina, A., Kundaje, A.: Not just a black box: learning
important features through propagating activation differences (2016). arXiv:1605.01713
52. Zhou, A.M., Gan, J.Q.: Low-level interpretability and high-level interpretability: a unified
view of data-driven interpretable fuzzy system modelling. Fuzzy Sets Syst. 159(23), 3091–
3131 (2008)
53. https://cloud.google.com/vision
54. Szegedy, C., et al.: Going deeper with convolutions. In: CPRV’15 (2004). arXiv preprint
arxiv:1409.4842
55. Xu, H., Saenko, K.: Ask, attend and answer: exploring question-guided spatial attention for
visual question answering. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016.
LNCS, vol. 9911, pp. 451–466. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-
46478-7_28
56. Lu, J., Yang, J., Batra, D.: Parikh. Hierarchical question image co-attention for visual question
answering. In: Advances in Neural Information Processing Systems (NIPS2016), pp 289–297
(2016)
57. Reed S. et al.: Generative adversarial text to image synthesis. In: ICML 2016, pp. 1060–1069
(2016)
58. Reed, S.E., Akata, Z., Mohan, S., Tenka, S., Schiele, B., Lee, H.: Learning what and where
to draw. In: NIPS, pp. 217–225 (2016)
59. Welinder, P., et al.: Caltech-UCSD Birds 200. Technical report. CNS-TR-2010- 001,
California Institute of Technology (2010)
60. Nilsback, M.E., Zisserman, A.: Automated flower classification over a large number of
classes. In: Proceedings of the Indian Conference on Computer Vision, Graphics and Image
Processing, pp. 722–729 (2008)
272 M. Oussalah
61. Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Dollár, P., Zitnick,
C.L.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B.,
Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 740–755. Springer, Cham (2014).
https://doi.org/10.1007/978-3-319-10602-1_48
62. Lesk, M.: Automatic sense disambiguation using machine readable dictionaries: how to tell
a pine cone from an ice cream cone. In: Proceedings of SIGDOC, pp. 24–26 (1986)
63. Mihalcea, R.: Knowledge-based methods for WSD. In: Word Sense Disambiguation: Algo-
rithms and Applications, Text, Speech and Language Technology, pp. 107–132. Springer,
Dordrecht (2006). https://doi.org/10.1007/978-1-4020-4809-8_5
64. Navigli, R., Lapata, M.: Graph connectivity measures for unsupervised word sense disam-
biguation. In: IJCAI International Joint Conference on Artificial Intelligence, Hyderabad,
India, pp. 1683–1688 (2007)
65. Dung, P.M.: On the acceptability of arguments and its fundamental role in nonmonotonic
reasoning, logic programming and n-person games. Artif. Intell. 77(2), 321–357 (1995)
66. Cocarascu, O., Stylianou, A., Cyras K., Toni, F.: Data-empowered argumentation for dialecti-
cally explainable predictions. In: 24th European Conference on Artificial Intelligence – ECAI
(2020)
67. Tsytsarau, M., Palpanas, T.: Survey on mining subjective data on the web. Data Min. Knowl.
Discov. 24, 478–514 (2012)
68. Arras, L., Horn, F., Montavon, G., Muller, K.R., Samek W.: Explaining predictions of non-
linear classifiers in NLP. arXiv preprint arXiv:1606.07298 (2016)
69. Zhang Y., Lai, G., Zhang, M., Zhang, Y., Liu, Y., Ma, S.: Explicit factor models for explainable
recommendation based on phrase-level sentiment analysis. In: Proceedings of the 37th ACM
SIGIR, pp. 83–92 (2014)
70. Sherstov, A.A., Stone, P.: Improving action selection in MDP’s via knowledge transfer. AAAI
5, 1024–1029 (2005)
71. Blei, D.M., Lafferty, J.D.: TopicMmodels. Chapman & Hall/CRC (2009)
72. Lau, H.J., Newman, D., Baldwin, T.: Machine reading tea leaves: automatically evaluating
topic coherence and topic model quality. In: EAC (2014)
73. Dieng, A.B., Ruiz, F.R., Blei, D.M.: Topic Modelling in Embedding Spaces (2019). arXiv:
1907.04907v1 cs.IR
74. Hu, Y., Boyd-Graber, J., Satinoff, B., Smith, A.: Interactive topic modeling. Mach. Learn. 95,
423–469 (2013)
75. Nenkova A., McKeown K.: A survey of text summarization techniques. In: Aggarwal, C.,
Zhai, C. (eds.) Mining Text Data. Springer, Boston, MA (2012). https://doi.org/10.1007/978-
1-4614-3223-4_3
76. Mihalcea, R., Tarau, P.: Textrank: bringing order into text. In: Proceedings of the conference
on empirical methods in natural language processing (2004)
77. Gong, Y., Liu, X.: Generic text summarization using relevance measure and latent semantic
analysis. In: Proceedings of the 24th ACM SIGIR, pp. 19– 25 (2001)
78. Steinberger, J., Jezek, K.: Using latent semantic analysis in text summarization and summary
evaluation. Proc. ISIM 4, 93–100 (2004)
79. Hong, S., Yang, D., Choi, J., Lee, H.: Interpretable text-to-image synthesis with hierarchical
semantic layout generation. In: Samek, et al. (eds.): Explainable AI: Interpreting, Explaining
and Visualizing Deep Learning. Springer, Germany (2019). https://doi.org/10.1007/978-3-
030-28954-6
80. Mohamed, M., Oussalah, M.: A hybrid approach for paraphrase identification based on
knowledge-enriched semantic heuristics. Lang. Resour. Eval. 54(2), 457–485 (2019). https://
doi.org/10.1007/s10579-019-09466-4
AI Explainability 273
81. Mohamed, M., Oussalah, M.: SRL-ESA-TextSum: a text summarization approach based on
semantic role labeling and explicit semantic analysis. Inf. Process. Manage. (2020). https://
doi.org/10.1016/j.ipm.2019.04.003
82. Buchanan, B.G., Shortliffe, E.H.: Rule Based Expert Systems: The MYCIN Experiment of
the Stanford Heuristic Programming Project. Addison-Wesley, Reading, MA (1984)