Nonverbal Gestures for Improving Joint Attention in Human-Robot Interactions
Joint attention is one of the main facets of social interaction (and a deficit in autism). To enable socially assistive robots to engage in richer social interactions with children, we are developing models of attention acquisition techniques. We model joint attention as a series of three steps: (1) attention acquisition, (2) attention direction, and (3) attention sharing.
Since socially assistive robots are often meant to interact with children in conjunction with other interfaces such as tablets or physical artifacts (toys, learning tools), the robotic often needs to first acquire the child’s attention before being able to direct it. We use nonverbal cues, primarily non-speech sounds and physical motion, to present a continuum of gesturing signals to acquire child attention based on measures of engagement.
The first validation study for the approach is under development in multi-party interactions with a convenience population (college students and families). After the culmination of the first data collection, we will use the data to develop a model of engagement for deployment in our upcoming long-term in-home deployments.
Two of the foundational research questions in this space are:
(1) What available channels of communication are best for generating minimally disruptive, attention acquisition gestures?
(2) In different interaction contexts, how are different magnitudes of gestures in each of these channels related to the level of engagement or “interruptability” of the user?
Our general approach is to experimentally find how sets of different magnitudes of gestures in the selected channels related to different levels of user engagement. By systematically generating different gestures while a user is performing a task (,i.e. On-screen puzzles or GRE questions), we can check if the generated gesture is sufficient by checking whether or not there is an observed response. From the collected data we can theoretically form an n-dimensional manifold that relates the different combinations of gestures to user engagement.
This adaptive, n-dimensional manifold in combination with real-time observations on user engagement will allow us to deliver appropriate multi-modal, nonverbal gestures for attention acquisition, enhancing the interaction experience by minimizing the perceived invasiveness and rudeness of the robotic agent.