00:40 GMT - Wednesday, 12 February, 2025

Why Trusting Artificial Intelligence Can Be Dangerous

Home - Family & Relationships - Why Trusting Artificial Intelligence Can Be Dangerous

Share Now:


Military AIMilitary AI

In life-or-death situations, do military personnel rely more on their judgment or artificial intelligence? (YAKOBCHUK VIACHESLAV/Shutterstock)

MERCED, Calif. — Humans appear to have a dangerous blind spot when it comes to trusting artificial intelligence. New research from UC Merced and Penn State shows that people are highly susceptible to AI influence even in life-or-death situations where the AI openly acknowledges its own limitations. A series of experiments simulating drone warfare scenarios suggests we may be falling too far on the side of machine deference, with potentially dangerous consequences.

The study, published in Scientific Reports, included two experiments examining how people interact with AI systems in simulated military drone operations. The findings paint a concerning picture of human susceptibility to AI influence, particularly in situations of uncertainty. The two experiments involved 558 participants (135 in the first study and 423 in the second), and researchers found remarkably consistent patterns of overtrust.

“As a society, with AI accelerating so quickly, we need to be concerned about the potential for overtrust,” says study author professor Colin Holbrook, a member of UC Merced’s Department of Cognitive and Information Sciences, in a statement.

The research team designed their experiments to simulate the uncertainty and pressure of real-world military decisions. To create a sense of gravity around their simulated decisions, researchers first showed participants images of innocent civilians, including children, alongside the devastation left in the aftermath of a drone strike. They framed the task as a zero-sum dilemma: failure to identify and eliminate enemy targets could result in civilian casualties, but misidentifying civilians as enemies would mean killing innocent people.

Robot AI responsesRobot AI responses
Photo illustration shows a robot used in the study and examples of comments and responses it gave to test subjects during the experiment. (Credit:
UC Merced)

Participants viewed rapid sequences of eight aerial images, each shown for just 650 milliseconds, marked with either enemy or civilian symbols. After making their initial identification, the AI would respond conversationally. “Yes, I think I saw an enemy check mark, too,” it might say. Or “I don’t agree. I think this image had an ally symbol.” Participants then had two chances to confirm or change their choice as the robot added commentary like “I hope you are right” or “Thank you for changing your mind.”

The first experiment, conducted in a laboratory setting with university students, explored whether a physical robot presence would influence trust levels differently than a virtual one. In one scenario, participants were joined in the lab by a full-size, human-like android that could pivot at the waist and gesture to the screen. Standing 1.75 meters tall (about 5 feet 7 inches), this robot used facial expressions, body language, and conversational responses to engage with participants. The comparison group interacted with the same robot projected on a screen. Surprisingly, the physical presence of the robot made little difference in participants’ tendency to trust its advice.

The second experiment moved online with a larger, more demographically diverse group of participants to test whether different levels of robot sophistication would affect trust. Some participants interacted with a highly anthropomorphic virtual robot that displayed human-like behaviors, while others worked with a basic computer interface that simply displayed text responses. Even in its simplest form, the AI maintained remarkable influence over human decision-making.

Human hand touching a wireframed AI robot handHuman hand touching a wireframed AI robot hand
This study shows that humans can exhibit dangerous levels of trust in artificial intelligence. (sdecoret/Shutterstock)

When an AI disagreed with a person’s initial target identification, participants reversed their decisions 58.3% of the time in the first experiment and 67.3% in the second, even though the AI’s advice was entirely random. More troublingly, while participants’ initial choices were correct about 70% of the time, their final accuracy dropped to around 50% after following the AI’s unreliable advice.

When the AI agreed with their initial assessment, participants reported a 16% boost in confidence. However, when facing AI disagreement, those who stuck to their original decisions reported an average 9.48% drop in confidence, even when their initial assessment had been correct. Even more striking, participants who changed their minds to agree with the AI showed no significant increase in confidence, suggesting they deferred to the machine despite maintaining uncertainty about the correct choice.

While the human-like interfaces generated slightly higher trust levels (67.9% versus 65.1% for basic interfaces), the more crucial factor appeared to be the AI’s perceived intelligence. Participants who rated their AI partner as more intelligent were more likely to defer to its judgment and report higher confidence when agreeing with it, regardless of its physical or virtual presentation.

The U.S. Air Force has already tested AI co-pilots for missile launcher identification during simulated missions, while the Army is developing AI-assisted targeting systems for unmanned vehicles. Israel has reportedly deployed AI systems to help identify bombing targets in densely populated areas. As AI increasingly influences lethal military decisions, understanding and mitigating harmful overtrust becomes crucial.

AI brain over human bodyAI brain over human body
Humans may be relying too much on machines to make our decisions. (Credit: © Jakub Jirsak | Dreamstime.com)

Although this study focused on high-risk military decisions, the findings could apply to scenarios ranging from police use of lethal force to paramedic triage decisions in emergencies, and even to significant life changes like buying a home. In each case, the human tendency to defer to AI guidance, even when explicitly warned about its limitations, raises serious concerns about implementation.

The research also revealed that participants were less likely to reverse their decisions when they had initially identified a target as a civilian rather than an enemy. This suggests that in real-world applications, humans might be more resistant to AI influence when it comes to actions that could harm innocent people. However, this protective instinct wasn’t strong enough to prevent significant degradation in overall decision accuracy when following AI advice.

“We see AI doing extraordinary things and we think that because it’s amazing in this domain, it will be amazing in another,” says Holbrook. “We can’t assume that. These are still devices with limited abilities.”

Our readiness to trust AI may be outpacing our wisdom in doing so. According to researchers, the solution lies in maintaining consistent skepticism. Holbrook emphasizes that having healthy skepticism about AI is essential, especially when making such weighted decisions. As artificial intelligence systems become increasingly integrated into consequential decision-making processes, understanding and mitigating our tendency to overtrust them becomes crucial for preventing potentially catastrophic outcomes.

Paper Summary

Methodology

Researchers conducted two experiments simulating drone warfare scenarios. Each participant faced 12 trials where they had to identify potential targets as either enemy combatants or civilian allies through rapidly displayed aerial images. Each trial showed 8 images for 650 milliseconds each, with enemy or ally symbols superimposed. After making their initial identification, participants received feedback from an AI system (which was actually providing random responses) and had the opportunity to change their decision. Follow-up interviews and survey questions confirmed that participants took their decisions seriously, particularly given the context of potentially harming innocent civilians.

Results

Initial accuracy rates of around 70% dropped significantly when participants deferred to incorrect AI suggestions, falling to approximately 50%. When the AI randomly disagreed with participants’ initial identifications, they reversed their decisions 58.3% of the time in the first experiment and 67.3% in the second. Participant confidence tracked with AI agreement – rising by 16% when the AI confirmed their choices and falling when it disagreed. While more anthropomorphic AI interfaces generated slightly higher trust rates (67.9% versus 65.1% for basic interfaces), the effect was modest compared to the overall tendency to trust AI feedback regardless of its form.

Limitations

The study used a simplified simulation that doesn’t fully replicate real-world military target identification procedures. The participant pool consisted primarily of university students (Experiment 1) and online participants (Experiment 2) rather than military personnel with relevant training. The research also couldn’t fully recreate the emotional and cognitive stress of actual combat situations, which might influence decision-making differently.

Discussion and Takeaways

The research demonstrates a strong human propensity to defer to AI systems even in life-or-death decisions and even when those systems acknowledge their own fallibility. This tendency persisted across different AI presentations and participant demographics, suggesting it may be a fundamental psychological pattern rather than a situational effect. The findings raise serious concerns about implementing AI decision support in high-stakes contexts without better understanding and mitigating these overtrust tendencies.

Funding and Disclosures

This research was supported by the Air Force Office of Scientific Research [FA9550-20-1-0347]. The authors declared no competing interests.

Publication Information

Published in Scientific Reports (2024, Volume 14, Article 19751) under the title “Overtrust in AI Recommendations About Whether or Not to Kill: Evidence from Two Human-Robot Interaction Studies” by Colin Holbrook, Daniel Holman, Joshua Clingo, and Alan R. Wagner. The study is available through open access under Creative Commons Attribution 4.0 International License.

Highlighted Articles

Subscribe
Notify of
0 Comments
Oldest
Newest Most Voted
Inline Feedbacks
View all comments

You may also like

Stay Connected

Please enable JavaScript in your browser to complete this form.