The future of robots: A psychological approach to human-automation interaction

(Nanowerk News) Itís called the uncanny valley. Those who are fans of the HBO show Westworld or who have seen the movie "Ex Machina" may already be familiar with the phenomenon. But for those who are not, itís essentially the idea that humans are comfortable with robots who have humanoid features, but become very uncomfortable when the robot looks almost but not exactly like a human.
For Dr. Nathan Tenhundfeld, however, the uncanny valley is just one of many factors he must take into account while researching human-automation interaction as an assistant professor in the Department of Psychology at The University of Alabama in Huntsville (UAH).
Dr. Nathan Tenhundfeld
Assistant professor of psychology Dr. Nathan Tenhundfeld, left, recently established the Advanced Teaming, Technology, Automation, and Computing Lab to study human-machine teaming. (Image: Michael Mercier, UAH)
"Weíre at a unique point with the development of the technology where automated systems or platforms are no longer a tool but a teammate that is incorporated into our day-to-day experiences," he says. "So weíre looking at commercial platforms that offer the same systems but in different forms to see whether a certain appearance or characteristic affects the user and in what way."
Take for example, the recent push by the U.S. Department of Defense to incorporate automation into warfighting. As a concept, it makes sense: the more robots we have fighting wars, the less cost there is to human life. But in practice, itís a little more complex. What should a warfighting robot look like? A person? A machine?
To answer these questions, Dr. Tenhundfeld has partnered with a colleague at the U.S. Air Force Academy, where he conducted research as a postdoctoral fellow, to use "a massive database of robots" (the ABOT Database) so that they determine how various components might affect the perception of a robotís capabilities. "We want to know things like, does a robot with wheels or a track fit better with our expectation of what we should be sending to war versus a humanoid robot?" he says. "And, does having a face on the robot affect whether we want to put it in harmís way?"
Even if there were easy answers Ė which there arenít Ė thereís another equally important factor to consider beyond the robotís user interface: trust. For a robot to be effective, the user must trust the information that it is providing. To explain, Dr. Tenhundfeld points to research he conducted on the Tesla Model X while at the Academy ("Calibrating Trust in Automation Through Familiarity With the Autoparking Feature of a Tesla Model X"). Looking at the carís autoparking feature specifically, he and his team wanted to determine the userís willingness to let the car complete its task as a function of their risk-taking preference or confidence in their abilities.
"The data suggest automated vehicles tend to be safer than humans, but humans donít like to relinquish control," he says with a laugh. "So we had this pattern where there were high intervention rates at first, but as they developed trust in the system Ė after it wasnít so novel and it started to meet their expectations Ė they began to trust it more and the intervention rates went down."
The flip side of that coin, however, is the potential for empathy in, or attachment to, a particular automated system users may have developed trust in. To illustrate this concept, he recounts a case study of explosive-ordinance disposal teams who employ robots to safely blow up bombs. "When they have to send the robots back to get repaired, they have an issue when theyíre given a different robot," he says. "So theyíve placed this trust in a specific robot even though the intelligence/capability is the same across all of the robots."
And lest it start to sound like there is already more than enough for Dr. Tenhundfeld to factor in, there is also situational trust, which sits somewhere between trust and overtrust. In this scenario, a user may develop a certain level of trust as a whole over time, but then realize they donít trust some aspects as much as much as others. "Say I have an automated system, or robot, providing intelligence in a mission-planning environment, and it screws that up," he says. "I might not trust it in a different environment, such as on the battlefield, even though it has a different physical embodiment for use in that environment, and may be distinctly capable on the battlefield."
In short, the increasingly digital nature of our world introduces a seemingly endless list of considerations when it comes to ensuring automated systems can successfully meet human needs Ė all of which Dr. Tenhundfeld must take into account with the research he is doing in his Advanced Teaming, Technology, Automation, and Computing Lab, or ATTAC Lab. But given UAHís role as an academic partner to this emerging industry, itís a challenge that he and his fellow researchers have embraced. "Businesses are focused on being first to market with a product," he says. "We help them improve the product so that it works well for the user."
Source: University of Alabama Huntsville
Subscribe to a free copy of one of our daily
Nanowerk Newsletter Email Digests
with a compilation of all of the day's news.
These articles might interest you as well: