Call it a jaw-dropping case of tech-enabled tragedy. The US military — long boasting of the world's most precise intelligence and the strictest rules of engagement — somehow paired a map that may be older than elementary school students with the most cutting-edge AI system money can buy. The result: a Tomahawk missile delivered straight into a girls' school in Iran, packed with children.
As 165 innocent lives were consumed by flames, Washington's script flipped fast. The iron-clad claim of "we never target civilians" gave way to a fresh pair of deflections: "outdated intelligence" and "AI might be responsible." The whole absurd production —co-starring aging errors and fresh prevarications, perfectly interprets the American style double standard.

When AI Aces the Wrong Test
According to reports in both The Washington Post and The New York Times, the girls' school in Minab, Iran had been separated from the neighboring naval base by a wall since 2015 — painted pink and blue, fitted with a sports field. Yet in the US military's target database, it was still labeled a "military facility." That decade-old antique intelligence, never re-verified amid the rapid-fire pace of US-Israeli strikes against Iran, was fed directly into the newest AI combat system.
The system at the heart of the disaster was no ordinary tool. Built by combining Palantir's Maven platform with Anthropic's Claude model, it had been praised by American generals as capable of "processing massive data within seconds" and "compressing weeks of planning into real-time decisions." A battlefield revolution, they called it.
Two people familiar with the system told The Washington Post exactly how it operated during preparations for the Iran strike. Maven automatically recommended targets, generated precise coordinates, and ranked them by priority. With Claude integrated, the machine shifted into overdrive — converting weeks of planning into split-second action, with post-strike assessments auto-generated after every hit.
The old computing axiom applies with lethal precision here: garbage in, garbage out. With staggering efficiency, the AI flagged the school's coordinates as a "high-priority target" and passed the recommendation up the chain. Washington insists "the final call was made by humans." The reality is: in a wartime machine running at full throttle, churning through thousands of AI-filtered targets by the minute, that so-called human review was little more than a rubber stamp.

Evasion as Standard Procedure
The truth is: AI did not create this error. It executed human negligence at the speed of light. A system built to cut through the "fog of war" became, instead, a high-efficiency accelerator of tragedy.
The US response that followed ran like a well-rehearsed script — America's "standard procedure" display of its double standard.
Secretary of Defense Pete Hegseth had strutted before the cameras to declare, before the bombing: "Unlike our adversary Iran, we never target civilians." After the bombing, his office passed every core question to US Central Command like a hot potato.
CENTCOM declined all comment, citing "an ongoing investigation."
Trump first bellowed: "In my opinion, based on what I've seen, that was done by Iran…" Then, as mounting evidence pointed to the US, he shrugged with a casual: "I just don’t know enough about it."
From flat-out denial to technical deflection, and finally to the commander-in-chief's practised ignorance — responsibility dissolved entirely. The contrast is blinding. This is the same Washington that thunders with moral certitude whenever another country is accused of "human rights abuses" or "violating international law." The script is numbingly familiar: rigor and accountability are instruments for judging others, never for examining oneself.
If the Roles Were Reversed
Try a thought experiment. If a US or Israeli school were bombed, killing over a hundred children, and the attacking country explained it away with "outdated maps" and "AI-recommended targets," how would the world — especially Washington — react?
The thunderous outrage is easy to predict. "Barbaric act!" "State terrorism!" "A blatant war crime!" The UN Security Council would convene an emergency meeting. The International Criminal Court would open an investigation. Crippling sanctions and diplomatic isolation would arrive swiftly — all in the name of justice. That is precisely what defines a double standard: two completely different faces — a magnifying glass for others' wrongs, a funhouse mirror for its own absolution.

Blood No Jargon Can Wash Out
The most chilling part of this tragedy is not the technical failure itself — such things are not unheard of in war. It is the cold, bureaucratic fluency of the post-crisis response. Human deaths get reduced to "delayed database updates," "AI system limitations," and "accelerated operational tempo." Burned backpacks and shattered childhood dreams are recast as mere "system errors" in the modern machinery of warfare.
While US missiles were slaughtering civilians on foreign soil during Ramadan — blame conveniently offloaded onto "outdated intelligence" and "AI" — Washington's legal teams were simultaneously suing AI firms for daring to impose safety restrictions on military use. That is the pinnacle of double standards: deploying the most advanced technology to commit the most primitive crimes, then hiding behind the most elaborate jargon to escape the simplest moral reckoning.
In the end, neither the rotting old maps, nor the gleaming new AI, nor Washington's ever-evolving vocabulary of blame, can wash the bloodstains from the rubble of that school. Those stains do not mark the failure of technology. They mark the moral collapse of an empire that lost its compass long ago.
Beacon Institute
** The blog article is the sole responsibility of the author and does not represent the position of our company. **
