ChatGPT maker OpenAI this week quietly removed language from its usage policy that prohibited military use of its technology, a move with serious implications given the increase use of artificial intelligence on battlefields including Gaza.
ChatGPT is a free tool that lets users enter prompts to receive text or images generated by AI. The Intercept’s Sam Biddle reported Friday that prior to Wednesday, OpenAI’s permissible uses page banned “activity that has high risk of physical harm, including,” specifically, “weapons development” and “military and warfare.”
Although the company’s new policy stipulates that users should not harm human beings or “develop or use weapons,” experts said the removal of the “military and warfare” language leaves open the door for lucrative contracts with U.S. and other militaries.
“Given the use of AI systems in the targeting of civilians in Gaza, it’s a notable moment to make the decision to remove the words ‘military and warfare’ from OpenAI’s permissible use policy,” Sarah Myers West, managing director of the AI Now Institute and a former AI policy analyst at the Federal Trade Commission, told The Intercept.
“The language that is in the policy remains vague and raises questions about how OpenAI intends to approach enforcement,” she added.
OpenAI spokesperson Niko Felix told The Intercept that the company “aimed to create a set of universal principles that are both easy to remember and apply, especially as our tools are now globally used by everyday users who can now also build GPTs.”
“A principle like ‘don’t harm others’ is broad yet easily grasped and relevant in numerous contexts,” Felix added. “Additionally, we specifically cited weapons and injury to others as clear examples.”
As AI advances, so does its weaponization. Experts warn that AI applications including lethal autonomous weapons systems, commonly called “killer robots,” could pose a potentially existential threat to humanity that underscores the imperative of arms control measures to slow the pace of weaponization.
That’s the goal of nuclear weapons legislation introduced last year in the U.S. Congress. The bipartisan Block Nuclear Launch by Autonomous Artificial Intelligence Act — introduced by Sen. Ed Markey (D-Mass.) and Reps. Ted Lieu (D-Calif.), Don Beyer (D-Va.), and Ken Buck (R-Colo.) — asserts that “any decision to launch a nuclear weapon should not be made” by AI.
Angry, shocked, overwhelmed? Take action: Support independent media.
We’ve borne witness to a chaotic first few months in Trump’s presidency.
Over the last months, each executive order has delivered shock and bewilderment — a core part of a strategy to make the right-wing turn feel inevitable and overwhelming. But, as organizer Sandra Avalos implored us to remember in Truthout last November, “Together, we are more powerful than Trump.”
Indeed, the Trump administration is pushing through executive orders, but — as we’ve reported at Truthout — many are in legal limbo and face court challenges from unions and civil rights groups. Efforts to quash anti-racist teaching and DEI programs are stalled by education faculty, staff, and students refusing to comply. And communities across the country are coming together to raise the alarm on ICE raids, inform neighbors of their civil rights, and protect each other in moving shows of solidarity.
It will be a long fight ahead. And as nonprofit movement media, Truthout plans to be there documenting and uplifting resistance.
As we undertake this life-sustaining work, we appeal for your support. We have 3 days left in our fundraiser: Please, if you find value in what we do, join our community of sustainers by making a monthly or one-time gift.