SUBSCRIBE TO OUR FREE NEWSLETTER
Daily news & progressive opinion—funded by the people, not the corporations—delivered straight to your inbox.
5
#000000
#FFFFFF
To donate by check, phone, or other method, see our More Ways to Give page.
Daily news & progressive opinion—funded by the people, not the corporations—delivered straight to your inbox.
"As we live in an increasingly digital age, we need to ensure that humans hold the power alone to command, control, and launch nuclear weapons—not robots," said co-sponsor Sen. Ed Markey.
In the name of "protecting future generations from potentially devastating consequences," a bipartisan group of U.S. lawmakers on Wednesday introduced legislation meant to prevent artificial intelligence from launching nuclear weapons without meaningful human control.
The Block Nuclear Launch by Autonomous Artificial Intelligence Act—introduced by Sen. Ed Markey (D-Mass.) and Reps. Ted Lieu (D-Calif.), Don Beyer (D-Va.), and Ken Buck (R-Colo.)—asserts that "any decision to launch a nuclear weapon should not be made" by AI.
The proposed legislation acknowledges that the Pentagon's 2022 Nuclear Posture Review states that current U.S. policy is to "maintain a human 'in the loop' for all actions critical to informing and executing decisions by the president to initiate and terminate nuclear weapon employment."
The bill would codify that policy so that no federal funds could be used "to launch a nuclear weapon [or] select or engage targets for the purposes of launching" nukes.
"As we live in an increasingly digital age, we need to ensure that humans hold the power alone to command, control, and launch nuclear weapons—not robots," Markey asserted in a statement. "We need to keep humans in the loop on making life-or-death decisions to use deadly force, especially for our most dangerous weapons."
\u201cAI is amazing and has made our lives better. It can also kill us. No matter how smart AI becomes, it can never have control over nuclear weapons.\n\nI introduced bipartisan legislation with @RepKenBuck @RepDonBeyer and Sen @EdMarkey to require a human to launch any nuclear weapon.\u201d— Ted Lieu (@Ted Lieu) 1682544333
Buck argued that "while U.S. military use of AI can be appropriate for enhancing national security purposes, use of AI for deploying nuclear weapons without a human chain of command and control is reckless, dangerous, and should be prohibited."
According to the 2023 AI Index Report—an annual assessment published earlier this month by the Stanford Institute for Human-Centered Artificial Intelligence—36% of surveyed AI experts worry about the possibility that automated systems "could cause nuclear-level catastrophe."
"Use of AI for deploying nuclear weapons without a human chain of command and control is reckless, dangerous, and should be prohibited."
The report followed a February assessment by the Arms Control Association, an advocacy group, that AI and other emerging technologies including lethal autonomous weapons systems and hypersonic missiles pose a potentially existential threat that underscores the need for measures to slow the pace of weaponization.
"While we all try to grapple with the pace at which AI is accelerating, the future of AI and its role in society remains unclear," Lieu said in a statement introducing the new bill.
"It is our job as members of Congress to have responsible foresight when it comes to protecting future generations from potentially devastating consequences," he continued. "That's why I'm pleased to introduce the bipartisan, bicameral Block Nuclear Launch by Autonomous AI Act, which will ensure that no matter what happens in the future, a human being has control over the employment of a nuclear weapon—not a robot."
"AI can never be a substitute for human judgment when it comes to launching nuclear weapons," Lieu added.
While dozens of countries support the Treaty on the Prohibition of Nuclear Weapons, none of the world's nine nuclear powers, including the United States, have signed on, and Russia's invasion of Ukraine has reawakened fears of nuclear conflict that were largely dormant since the Cold War.
Political revenge. Mass deportations. Project 2025. Unfathomable corruption. Attacks on Social Security, Medicare, and Medicaid. Pardons for insurrectionists. An all-out assault on democracy. Republicans in Congress are scrambling to give Trump broad new powers to strip the tax-exempt status of any nonprofit he doesn’t like by declaring it a “terrorist-supporting organization.” Trump has already begun filing lawsuits against news outlets that criticize him. At Common Dreams, we won’t back down, but we must get ready for whatever Trump and his thugs throw at us. Our Year-End campaign is our most important fundraiser of the year. As a people-powered nonprofit news outlet, we cover issues the corporate media never will, but we can only continue with our readers’ support. By donating today, please help us fight the dangers of a second Trump presidency. |
In the name of "protecting future generations from potentially devastating consequences," a bipartisan group of U.S. lawmakers on Wednesday introduced legislation meant to prevent artificial intelligence from launching nuclear weapons without meaningful human control.
The Block Nuclear Launch by Autonomous Artificial Intelligence Act—introduced by Sen. Ed Markey (D-Mass.) and Reps. Ted Lieu (D-Calif.), Don Beyer (D-Va.), and Ken Buck (R-Colo.)—asserts that "any decision to launch a nuclear weapon should not be made" by AI.
The proposed legislation acknowledges that the Pentagon's 2022 Nuclear Posture Review states that current U.S. policy is to "maintain a human 'in the loop' for all actions critical to informing and executing decisions by the president to initiate and terminate nuclear weapon employment."
The bill would codify that policy so that no federal funds could be used "to launch a nuclear weapon [or] select or engage targets for the purposes of launching" nukes.
"As we live in an increasingly digital age, we need to ensure that humans hold the power alone to command, control, and launch nuclear weapons—not robots," Markey asserted in a statement. "We need to keep humans in the loop on making life-or-death decisions to use deadly force, especially for our most dangerous weapons."
\u201cAI is amazing and has made our lives better. It can also kill us. No matter how smart AI becomes, it can never have control over nuclear weapons.\n\nI introduced bipartisan legislation with @RepKenBuck @RepDonBeyer and Sen @EdMarkey to require a human to launch any nuclear weapon.\u201d— Ted Lieu (@Ted Lieu) 1682544333
Buck argued that "while U.S. military use of AI can be appropriate for enhancing national security purposes, use of AI for deploying nuclear weapons without a human chain of command and control is reckless, dangerous, and should be prohibited."
According to the 2023 AI Index Report—an annual assessment published earlier this month by the Stanford Institute for Human-Centered Artificial Intelligence—36% of surveyed AI experts worry about the possibility that automated systems "could cause nuclear-level catastrophe."
"Use of AI for deploying nuclear weapons without a human chain of command and control is reckless, dangerous, and should be prohibited."
The report followed a February assessment by the Arms Control Association, an advocacy group, that AI and other emerging technologies including lethal autonomous weapons systems and hypersonic missiles pose a potentially existential threat that underscores the need for measures to slow the pace of weaponization.
"While we all try to grapple with the pace at which AI is accelerating, the future of AI and its role in society remains unclear," Lieu said in a statement introducing the new bill.
"It is our job as members of Congress to have responsible foresight when it comes to protecting future generations from potentially devastating consequences," he continued. "That's why I'm pleased to introduce the bipartisan, bicameral Block Nuclear Launch by Autonomous AI Act, which will ensure that no matter what happens in the future, a human being has control over the employment of a nuclear weapon—not a robot."
"AI can never be a substitute for human judgment when it comes to launching nuclear weapons," Lieu added.
While dozens of countries support the Treaty on the Prohibition of Nuclear Weapons, none of the world's nine nuclear powers, including the United States, have signed on, and Russia's invasion of Ukraine has reawakened fears of nuclear conflict that were largely dormant since the Cold War.
In the name of "protecting future generations from potentially devastating consequences," a bipartisan group of U.S. lawmakers on Wednesday introduced legislation meant to prevent artificial intelligence from launching nuclear weapons without meaningful human control.
The Block Nuclear Launch by Autonomous Artificial Intelligence Act—introduced by Sen. Ed Markey (D-Mass.) and Reps. Ted Lieu (D-Calif.), Don Beyer (D-Va.), and Ken Buck (R-Colo.)—asserts that "any decision to launch a nuclear weapon should not be made" by AI.
The proposed legislation acknowledges that the Pentagon's 2022 Nuclear Posture Review states that current U.S. policy is to "maintain a human 'in the loop' for all actions critical to informing and executing decisions by the president to initiate and terminate nuclear weapon employment."
The bill would codify that policy so that no federal funds could be used "to launch a nuclear weapon [or] select or engage targets for the purposes of launching" nukes.
"As we live in an increasingly digital age, we need to ensure that humans hold the power alone to command, control, and launch nuclear weapons—not robots," Markey asserted in a statement. "We need to keep humans in the loop on making life-or-death decisions to use deadly force, especially for our most dangerous weapons."
\u201cAI is amazing and has made our lives better. It can also kill us. No matter how smart AI becomes, it can never have control over nuclear weapons.\n\nI introduced bipartisan legislation with @RepKenBuck @RepDonBeyer and Sen @EdMarkey to require a human to launch any nuclear weapon.\u201d— Ted Lieu (@Ted Lieu) 1682544333
Buck argued that "while U.S. military use of AI can be appropriate for enhancing national security purposes, use of AI for deploying nuclear weapons without a human chain of command and control is reckless, dangerous, and should be prohibited."
According to the 2023 AI Index Report—an annual assessment published earlier this month by the Stanford Institute for Human-Centered Artificial Intelligence—36% of surveyed AI experts worry about the possibility that automated systems "could cause nuclear-level catastrophe."
"Use of AI for deploying nuclear weapons without a human chain of command and control is reckless, dangerous, and should be prohibited."
The report followed a February assessment by the Arms Control Association, an advocacy group, that AI and other emerging technologies including lethal autonomous weapons systems and hypersonic missiles pose a potentially existential threat that underscores the need for measures to slow the pace of weaponization.
"While we all try to grapple with the pace at which AI is accelerating, the future of AI and its role in society remains unclear," Lieu said in a statement introducing the new bill.
"It is our job as members of Congress to have responsible foresight when it comes to protecting future generations from potentially devastating consequences," he continued. "That's why I'm pleased to introduce the bipartisan, bicameral Block Nuclear Launch by Autonomous AI Act, which will ensure that no matter what happens in the future, a human being has control over the employment of a nuclear weapon—not a robot."
"AI can never be a substitute for human judgment when it comes to launching nuclear weapons," Lieu added.
While dozens of countries support the Treaty on the Prohibition of Nuclear Weapons, none of the world's nine nuclear powers, including the United States, have signed on, and Russia's invasion of Ukraine has reawakened fears of nuclear conflict that were largely dormant since the Cold War.