TY - JOUR AU - Fiorini, Laura AU - D'Onofrio, Grazia AU - Sorrentino, Alessandra AU - Cornacchia Loizzo, Federica Gabriella AU - Russo, Sergio AU - Ciccone, Filomena AU - Giuliani, Francesco AU - Sancarlo, Daniele AU - Cavallo, Filippo PY - 2024 DA - 2024/1/26 TI - The Role of Coherent Robot Behavior and Embodiment in Emotion Perception and Recognition During Human-Robot Interaction: Experimental Study JO - JMIR Hum Factors SP - e45494 VL - 11 KW - social robot KW - emotion recognition KW - human emotion perception KW - human-robot interaction KW - robot cospeech gestures evaluation AB - Background: Social robots are becoming increasingly important as companions in our daily lives. Consequently, humans expect to interact with them using the same mental models applied to human-human interactions, including the use of cospeech gestures. Research efforts have been devoted to understanding users’ needs and developing robot’s behavioral models that can perceive the user state and properly plan a reaction. Despite the efforts made, some challenges regarding the effect of robot embodiment and behavior in the perception of emotions remain open. Objective: The aim of this study is dual. First, it aims to assess the role of the robot’s cospeech gestures and embodiment in the user’s perceived emotions in terms of valence (stimulus pleasantness), arousal (intensity of evoked emotion), and dominance (degree of control exerted by the stimulus). Second, it aims to evaluate the robot’s accuracy in identifying positive, negative, and neutral emotions displayed by interacting humans using 3 supervised machine learning algorithms: support vector machine, random forest, and K-nearest neighbor. Methods: Pepper robot was used to elicit the 3 emotions in humans using a set of 60 images retrieved from a standardized database. In particular, 2 experimental conditions for emotion elicitation were performed with Pepper robot: with a static behavior or with a robot that expresses coherent (COH) cospeech behavior. Furthermore, to evaluate the role of the robot embodiment, the third elicitation was performed by asking the participant to interact with a PC, where a graphical interface showed the same images. Each participant was requested to undergo only 1 of the 3 experimental conditions. Results: A total of 60 participants were recruited for this study, 20 for each experimental condition for a total of 3600 interactions. The results showed significant differences (P<.05) in valence, arousal, and dominance when stimulated with the Pepper robot behaving COH with respect to the PC condition, thus underlying the importance of the robot’s nonverbal communication and embodiment. A higher valence score was obtained for the elicitation of the robot (COH and robot with static behavior) with respect to the PC. For emotion recognition, the K-nearest neighbor classifiers achieved the best accuracy results. In particular, the COH modality achieved the highest level of accuracy (0.97) when compared with the static behavior and PC elicitations (0.88 and 0.94, respectively). Conclusions: The results suggest that the use of multimodal communication channels, such as cospeech and visual channels, as in the COH modality, may improve the recognition accuracy of the user’s emotional state and can reinforce the perceived emotion. Future studies should investigate the effect of age, culture, and cognitive profile on the emotion perception and recognition going beyond the limitation of this work. SN - 2292-9495 UR - https://humanfactors.jmir.org/2024/1/e45494 UR - https://doi.org/10.2196/45494 UR - http://www.ncbi.nlm.nih.gov/pubmed/38277201 DO - 10.2196/45494 ID - info:doi/10.2196/45494 ER -