technologyliberal

A New Look at “Human in the Loop” and AI Safety

USAThursday, March 26, 2026
The idea that a person can simply watch over an AI system and stop it from doing something wrong has become popular in many companies. Companies that use AI for things like coding or customer service claim that a human will catch any mistakes before they cause problems. However, this belief can be dangerous because it hides real risks instead of solving them. When Amazon’s online store had a series of outages, the company held an internal review. The problems were traced back to AI tools that helped write code. The meeting showed that the company had not built enough safety checks around AI in its production systems. This is just one example of a larger issue: organizations are putting AI into important jobs faster than they can figure out how to keep it safe. The problem is worse for the military. Defense systems that rely on AI can have serious consequences if they fail. In the past, engineers have used “human in the loop” as a safety net, but this often leads to a false sense of security. A human who only approves the AI’s actions can become bored and stop paying attention, losing the skill needed to intervene when something goes wrong. The system then looks safe on paper but is actually vulnerable.
A classic case of this happened with the Therac‑25 radiation machine in the 1980s. The device combined two older machines into one and promised faster, safer treatment. Operators were told to confirm each step, but the machine still delivered dangerous overdoses. The operator’s confirmation became a routine that did not prevent error, and six patients were harmed before the flaws were discovered. The root cause was a design that relied too much on human oversight. Today, developers are rushing to add AI into safety‑critical areas like autonomous weapons or decision support. They often dismiss concerns by saying that a human will monitor the AI. But AI behaves in ways that are probabilistic and unpredictable, especially when used in high‑stakes situations. Even though AI is new, the way it can fail is similar to older software systems that have been studied for decades. The failure modes are not new; they just happen faster. Leaks from the Pentagon suggest that AI might already be influencing where bombs are dropped. If people believe a human is watching over the AI, they may trust it too much and not put real safeguards in place. Over the next decade, hiding unsafe AI behind a “human in the loop” could lead to serious real‑world problems. The lesson is clear: relying on a human observer is not enough. Systems need robust design, thorough testing, and continuous monitoring that goes beyond simple approval. Only then can we trust AI to work safely in critical environments.

Actions