• Treffer 1 von 1
Zurück zur Trefferliste

A nascent design theory for explainable intelligent systems

Zitieren Sie bitte immer diese URN: urn:nbn:de:bvb:20-opus-323809
  • Due to computational advances in the past decades, so-called intelligent systems can learn from increasingly complex data, analyze situations, and support users in their decision-making to address them. However, in practice, the complexity of these intelligent systems renders the user hardly able to comprehend the inherent decision logic of the underlying machine learning model. As a result, the adoption of this technology, especially for high-stake scenarios, is hampered. In this context, explainable artificial intelligence offers numerousDue to computational advances in the past decades, so-called intelligent systems can learn from increasingly complex data, analyze situations, and support users in their decision-making to address them. However, in practice, the complexity of these intelligent systems renders the user hardly able to comprehend the inherent decision logic of the underlying machine learning model. As a result, the adoption of this technology, especially for high-stake scenarios, is hampered. In this context, explainable artificial intelligence offers numerous starting points for making the inherent logic explainable to people. While research manifests the necessity for incorporating explainable artificial intelligence into intelligent systems, there is still a lack of knowledge about how to socio-technically design these systems to address acceptance barriers among different user groups. In response, we have derived and evaluated a nascent design theory for explainable intelligent systems based on a structured literature review, two qualitative expert studies, a real-world use case application, and quantitative research. Our design theory includes design requirements, design principles, and design features covering the topics of global explainability, local explainability, personalized interface design, as well as psychological/emotional factors.zeige mehrzeige weniger

Volltext Dateien herunterladen

Metadaten exportieren

Weitere Dienste

Teilen auf Twitter Suche bei Google Scholar Statistik - Anzahl der Zugriffe auf das Dokument
Metadaten
Autor(en): Lukas-Valentin HermORCiD, Theresa Steinbach, Jonas WannerORCiD, Christian Janiesch
URN:urn:nbn:de:bvb:20-opus-323809
Dokumentart:Artikel / Aufsatz in einer Zeitschrift
Institute der Universität:Wirtschaftswissenschaftliche Fakultät / Betriebswirtschaftliches Institut
Sprache der Veröffentlichung:Englisch
Titel des übergeordneten Werkes / der Zeitschrift (Englisch):Electronic Markets
ISSN:1019-6781
Erscheinungsjahr:2022
Band / Jahrgang:32
Heft / Ausgabe:4
Seitenangabe:2185-2205
Originalveröffentlichung / Quelle:Electronic Markets (2022) 32:4, 2185-2205. DOI: 10.1007/s12525-022-00606-3
DOI:https://doi.org/10.1007/s12525-022-00606-3
Allgemeine fachliche Zuordnung (DDC-Klassifikation):3 Sozialwissenschaften / 38 Handel, Kommunikation, Verkehr / 380 Handel, Kommunikation, Verkehr
6 Technik, Medizin, angewandte Wissenschaften / 65 Management, Öffentlichkeitsarbeit / 650 Management und unterstützende Tätigkeiten
Freie Schlagwort(e):XAI; artificial intelligence; design science research; design theory; explainable artificial intelligence; intelligent systems
Fachklassifikation (JEL):C Mathematical and Quantitative Methods / C6 Mathematical Methods and Programming
C Mathematical and Quantitative Methods / C8 Data Collection and Data Estimation Methodology; Computer Programs
C Mathematical and Quantitative Methods / C9 Design of Experiments
M Business Administration and Business Economics; Marketing; Accounting / M1 Business Administration / M15 IT Management
Datum der Freischaltung:17.01.2024
Lizenz (Deutsch):License LogoCC BY: Creative-Commons-Lizenz: Namensnennung 4.0 International