Abstract
Nowadays, Artificial Intelligence (AI) is proving to be successful for solving complex problems in various application domains. However, despite the numerous success stories of AI-systems, one challenge that characterizes these systems is that they often lack transparency in terms of understandability and explainability. In this study, we propose to address this challenge from the design thinking lens as a way to amplify human understanding of ML (Machine Learning) and AI algorithms. We exemplify our proposed approach by depicting a case based on a conventional ML algorithm applied on sentiment analysis of students’ feedback. This paper aims to contribute to the overall discourse of a need of innovation when it comes to the understandability and explainability of ML and AI solutions, especially since innovation is an inherent feature of design thinking.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Arrieta, A.B., Diaz-Rodrigues, N., Del Ser, J., Bennetot, A., Tabik, S., Barbado, A., Garcia, S., Gli-Lopez, S., Molina, D., Benjamins, R., Chatila, R.: Explainable artificial intelligence (XAI): concepts, taxonomies, opportunities and challenges towards responsible AI. Inf. Fusion 58, 82–115 (2020)
Hendricks, L.A., Akata, Z., Rohrbach, M., Donahue, J., Schiele, B., Darrell, T.: Generating visual explanations. In: European Conference on Computer Vision, pp. 3–19. Springer Cham (2016)
Fulton, L.B., Lee, J.Y., Wang, Q., Yuan, Z., Hammer, J., Perer, A.: Getting playful with explainable AI: games with a purpose to improve human understanding of AI. In: Extended Abstracts of the CHI Conference on Human Factors in Computing Systems, pp. 1–8. Association for Computing Machinery (ACM), New York (2020)
Adadi, A., Berrada, M.: Peeking inside the black-box: a survey on explainable artificial intelligence (XAI). IEEE Access 6, 52138–52160 (2018)
Liao, Q.V., Gruen, D., Miller, S.: Questioning the AI: informing design practices for explainable AI user experiences. In: Proceedings of the CHI Conference on Human Factors in Computing Systems, p. 15. Association for Computing Machinery (ACM), New York (2020)
Wolf, C.T.: Explainability scenarios: towards scenario-based XAI design. In: Proceedings of the 24th International Conference on Intelligent User Interfaces, pp. 252–257. Association for Computing Machinery (ACM), New York (2019)
Wolf, C.T., Ringland, K.E.: Designing accessible, explainable AI (XAI) experiences. ACM SIGACCESS Accessibility Comput. (125) (2020)
Guidotti, R., Monreale, A., Ruggieri, S., Turini, F., Giannotti, F., Pedreschi, D.: A survey of methods for explaining black box models. ACM Comput. Surv. (CSUR) 51(5), 1–42 (2018)
Brown, T.: Design thinking. Harv. Bus. Rev. 86(6), 84 (2008)
Kastrati, Z., Arifaj, B., Lubishtani, A., Gashi, F., Nishliu, E.: Aspect-based opinion mining of students’ reviews on online courses. In: Proceedings of the 6th International Conference on Computing and Artificial Intelligence, pp. 510–514. Association for Computing Machinery, New York (2020)
Kastrati, Z., Imran, A.S., Kurti, A.: Weakly supervised framework for aspect-based sentiment analysis on students’ reviews of MOOCs . IEEE Access 8, 106799–106810 (2020)
Lissack, M.: Understanding is a design problem: cognizing from a designerly thinking perspective. Part 1. She Ji: J. Des. Econ. Innov. 5(3), 231–246 (2019)
Lissack, M.: Understanding is a design problem: cognizing from a designerly thinking perspective. Part 2. She Ji: J. Des. Econ. Innov. 5(4), 327–342 (2019)
Weller, A.J.: Design thinking for a user-centered approach to artificial intelligence. She Ji: J. Des. Econ. Innov. 5(4), 394–396 (2019)
Miller, G.A.: The magical number seven, plus or minus two: some limits on our capacity for processing information. Psychol. Rev. 63(2), 81 (1956)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Kurti, A., Dalipi, F., Ferati, M., Kastrati, Z. (2021). Increasing the Understandability and Explainability of Machine Learning and Artificial Intelligence Solutions: A Design Thinking Approach. In: Ahram, T., Taiar, R., Groff, F. (eds) Human Interaction, Emerging Technologies and Future Applications IV. IHIET-AI 2021. Advances in Intelligent Systems and Computing, vol 1378. Springer, Cham. https://doi.org/10.1007/978-3-030-74009-2_5
Download citation
DOI: https://doi.org/10.1007/978-3-030-74009-2_5
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-73270-7
Online ISBN: 978-3-030-74009-2
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)