• search hit 7 of 8
Back to Result List

A nascent design theory for explainable intelligent systems

Please always quote using this URN: urn:nbn:de:bvb:20-opus-323809
  • Due to computational advances in the past decades, so-called intelligent systems can learn from increasingly complex data, analyze situations, and support users in their decision-making to address them. However, in practice, the complexity of these intelligent systems renders the user hardly able to comprehend the inherent decision logic of the underlying machine learning model. As a result, the adoption of this technology, especially for high-stake scenarios, is hampered. In this context, explainable artificial intelligence offers numerousDue to computational advances in the past decades, so-called intelligent systems can learn from increasingly complex data, analyze situations, and support users in their decision-making to address them. However, in practice, the complexity of these intelligent systems renders the user hardly able to comprehend the inherent decision logic of the underlying machine learning model. As a result, the adoption of this technology, especially for high-stake scenarios, is hampered. In this context, explainable artificial intelligence offers numerous starting points for making the inherent logic explainable to people. While research manifests the necessity for incorporating explainable artificial intelligence into intelligent systems, there is still a lack of knowledge about how to socio-technically design these systems to address acceptance barriers among different user groups. In response, we have derived and evaluated a nascent design theory for explainable intelligent systems based on a structured literature review, two qualitative expert studies, a real-world use case application, and quantitative research. Our design theory includes design requirements, design principles, and design features covering the topics of global explainability, local explainability, personalized interface design, as well as psychological/emotional factors.show moreshow less

Download full text files

Export metadata

Additional Services

Share in Twitter Search Google Scholar Statistics
Metadaten
Author: Lukas-Valentin HermORCiD, Theresa Steinbach, Jonas WannerORCiD, Christian Janiesch
URN:urn:nbn:de:bvb:20-opus-323809
Document Type:Journal article
Faculties:Wirtschaftswissenschaftliche Fakultät / Betriebswirtschaftliches Institut
Language:English
Parent Title (English):Electronic Markets
ISSN:1019-6781
Year of Completion:2022
Volume:32
Issue:4
Pagenumber:2185-2205
Source:Electronic Markets (2022) 32:4, 2185-2205. DOI: 10.1007/s12525-022-00606-3
DOI:https://doi.org/10.1007/s12525-022-00606-3
Dewey Decimal Classification:3 Sozialwissenschaften / 38 Handel, Kommunikation, Verkehr / 380 Handel, Kommunikation, Verkehr
6 Technik, Medizin, angewandte Wissenschaften / 65 Management, Öffentlichkeitsarbeit / 650 Management und unterstützende Tätigkeiten
Tag:XAI; artificial intelligence; design science research; design theory; explainable artificial intelligence; intelligent systems
JEL-Classification:C Mathematical and Quantitative Methods / C6 Mathematical Methods and Programming
C Mathematical and Quantitative Methods / C8 Data Collection and Data Estimation Methodology; Computer Programs
C Mathematical and Quantitative Methods / C9 Design of Experiments
M Business Administration and Business Economics; Marketing; Accounting / M1 Business Administration / M15 IT Management
Release Date:2024/01/17
Licence (German):License LogoCC BY: Creative-Commons-Lizenz: Namensnennung 4.0 International