Refine
Has Fulltext
- yes (2)
Is part of the Bibliography
- yes (2)
Year of publication
- 2022 (2)
Document Type
- Journal article (2)
Language
- English (2)
Keywords
- embodiment (2) (remove)
Institute
- Institut Mensch - Computer - Medien (2) (remove)
When interacting with sophisticated digital technologies, people often fall back on the same interaction scripts they apply to the communication with other humans—especially if the technology in question provides strong anthropomorphic cues (e.g., a human-like embodiment). Accordingly, research indicates that observers tend to interpret the body language of social robots in the same way as they would with another human being. Backed by initial evidence, we assumed that a humanoid robot will be considered as more dominant and competent, but also as more eerie and threatening once it strikes a so-called power pose. Moreover, we pursued the research question whether these effects might be accentuated by the robot’s body size. To this end, the current study presented 204 participants with pictures of the robot NAO in different poses (expansive vs. constrictive), while also manipulating its height (child-sized vs. adult-sized). Our results show that NAO’s posture indeed exerted strong effects on perceptions of dominance and competence. Conversely, participants’ threat and eeriness ratings remained statistically independent of the robot’s depicted body language. Further, we found that the machine’s size did not affect any of the measured interpersonal perceptions in a notable way. The study findings are discussed considering limitations and future research directions.
Introduction
Modern digital devices, such as conversational agents, simulate human–human interactions to an increasing extent. However, their outward appearance remains distinctly technological. While research revealed that mental representations of technology shape users' expectations and experiences, research on technology sending ambiguous cues is rare.
Methods
To bridge this gap, this study analyzes drawings of the outward appearance participants associate with voice assistants (Amazon Echo or Google Home).
Results
Human beings and (humanoid) robots were the most frequent associations, which were rated to be rather trustworthy, conscientious, agreeable, and intelligent. Drawings of the Amazon Echos and Google Homes differed marginally, but “human,” “robotic,” and “other” associations differed with respect to the ascribed humanness, consciousness, intellect, affinity to technology, and innovation ability.
Discussion
This study aims to further elaborate on the rather unconscious cognitive and emotional processes elicited by technology and discusses the implications of this perspective for developers, users, and researchers.