Believing in BERT: Using expressive communication to enhance trust and counteract operational error in physical Human-Robot Interaction

Reading time: 5 minute
...

📝 Abstract

Strategies are necessary to mitigate the impact of unexpected behavior in collaborative robotics, and research to develop solutions is lacking. Our aim here was to explore the benefits of an affective interaction, as opposed to a more efficient, less error prone but non-communicative one. The experiment took the form of an omelet-making task, with a wide range of participants interacting directly with BERT2, a humanoid robot assistant. Having significant implications for design, results suggest that efficiency is not the most important aspect of performance for users; a personable, expressive robot was found to be preferable over a more efficient one, despite a considerable trade off in time taken to perform the task. Our findings also suggest that a robot exhibiting human-like characteristics may make users reluctant to ‘hurt its feelings’; they may even lie in order to avoid this.

💡 Analysis

Strategies are necessary to mitigate the impact of unexpected behavior in collaborative robotics, and research to develop solutions is lacking. Our aim here was to explore the benefits of an affective interaction, as opposed to a more efficient, less error prone but non-communicative one. The experiment took the form of an omelet-making task, with a wide range of participants interacting directly with BERT2, a humanoid robot assistant. Having significant implications for design, results suggest that efficiency is not the most important aspect of performance for users; a personable, expressive robot was found to be preferable over a more efficient one, despite a considerable trade off in time taken to perform the task. Our findings also suggest that a robot exhibiting human-like characteristics may make users reluctant to ‘hurt its feelings’; they may even lie in order to avoid this.

📄 Content

Abstract— Strategies are necessary to mitigate the impact of unexpected behavior in collaborative robotics, and research to develop solutions is lacking. Our aim here was to explore the benefits of an affective interaction, as opposed to a more efficient, less error prone but non-communicative one. The experiment took the form of an omelet-making task, with a wide range of participants interacting directly with BERT2, a humanoid robot assistant. Having significant implications for design, results suggest that efficiency is not the most important aspect of performance for users; a personable, expressive robot was found to be preferable over a more efficient one, despite a considerable trade off in time taken to perform the task. Our findings also suggest that a robot exhibiting human-like characteristics may make users reluctant to ‘hurt its feelings’; they may even lie in order to avoid this.
I. INTRODUCTION The new arena of collaborative robotics is sorely in need of strategies to deal with the challenges that arise where robots and humans work in close proximity. The human- populated world contains an infinite number of unknowns. Thus, while collaborative robotics is still in its nascent phase, designing for the initial mistakes, misunderstandings and failures likely to arise between human and robot is crucial. How does a robot recover a user’s trust after an error? How effective is an attempt to rectify the situation, or an apology, in mitigating dissatisfaction caused by unpredictable behavior? To what extent can a machine displaying human- like attributes soften displeasure? These are the central questions this paper seeks to begin to address. Typically, evaluation has tended to emphasize the efficiency of collaboration under varying conditions, as outlined in [27]. This (simulated) study suggests that efficiency is not the most important aspect of performance for users and seeks to demonstrate that embodied emotional expressiveness improves the integration of human-robot activity. Taking this one step further, we explored whether expressiveness - verbal and non-verbal - could mitigate any

  • This work was supported in part by the EPSRC grants EP/K006320/1 and EP/K006223/1, as part of the project “Trustworthy Robotic Assistants.”
    a A. Hamacher is with the UCL Interaction Centre, University College London, UK. She is also a freelance journalist (+44 7808 719739; email: adriana.hamacher.11@ucl.ac.uk, adahamacher@gmail.com).
    b N. Bianchi-Berthouze is with the UCL Interaction Centre, University College London, UK (email: n.berthouze@ucl.ac.uk).
    c A.G. Pipe is with the Bristol Robotics Laboratory, Bristol, UK (email: Tony.Pipe@brl.ac.uk). d K. Eder is with the Department of Computer Science, University of Bristol, UK (email: kerstin.eder@bristol.ac.uk) and leads the Verification and Validation for Safety in Robots research theme at the Bristol Robotics Laboratory, Bristol, UK.
    dissatisfaction caused by erroneous or unexpected behavior and positively affect the participants’ experience.

Figure 1: The BERT2 platform with neutral expression (left) and BERT C’s facial expression on egg drop (right). Our study took the form of a real-life task, with participants spanning a range of ages and levels of experience (with robots), working directly with a humanoid robot assistant (Fig. 1) in an omelet-making task. A total of 15 of the 21 participants preferred the communicative, personable robot over a more efficient, less error prone, but non- communicative one. Satisfaction was significantly increased in the communicative condition and participants were particularly responsive to this robot’s apparent awareness of its error and expression of regret. For the majority, personable, transparent behavior appeared to negate the fact that the interaction took 50 per cent longer than in the non- communicative conditions. Our results suggest that users are likely to prefer an expressive and personable robot, even if it is less efficient and more error prone, than a non-communicative one. Our study furthermore offers fresh insight into how mistakes made by a robot affect its trustworthiness and acceptance in human-robot collaboration. It suggests that a robot effectively demonstrating apparent emotions, such as regret and enthusiasm, and awareness of its error, influences the user experience in such a way that dissatisfaction with its erroneous behavior is significantly tempered, if not forgiven, with a corresponding effect on trust. In fact, human-like characteristics may make users reluctant to hurt a robot’s ‘feelings’ and they may even lie in order to avoid this. II. LITERATURE REVIEW As robots are increasingly developed for use in social settings, acceptance, persuasiveness and likability are key and Believing in BERT:
Using expressive communication to enhance trust and counteract operation

This content is AI-processed based on ArXiv data.

Start searching

Enter keywords to search articles

↑↓
ESC
⌘K Shortcut