Skip to main content
Clear icon
61Āŗ

Biden wants to move fast on AI safeguards and signs an executive order to address his concerns

1 / 4

Copyright 2023 The Associated Press. All rights reserved

President Joe Biden signs an executive on artificial intelligence in the East Room of the White House, Monday, Oct. 30, 2023, in Washington. Vice President Kamala Harris looks on at right. (AP Photo/Evan Vucci)

WASHINGTON ā€“ President Joe Biden on Monday signed an ambitious executive order on artificial intelligence that seeks to balance the needs of cutting-edge technology companies with national security and consumer rights, creating an early set of guardrails that could be fortified by legislation and global agreements.

Before signing the order, Biden said AI is driving change at ā€œwarp speedā€ and carries tremendous potential as well as perils.

Recommended Videos



ā€œAI is all around us,ā€ Biden said. ā€œTo realize the promise of AI and avoid the risk, we need to govern this technology.ā€

The order is an initial step that is meant to ensure that AI is trustworthy and helpful, rather than deceptive and destructive. The order ā€” which will likely need to be augmented by congressional action ā€” seeks to steer how AI is developed so that companies can profit without putting public safety in jeopardy.

Using the Defense Production Act, the order requires leading AI developers to share safety test results and other information with the government. The National Institute of Standards and Technology is to create standards to ensure AI tools are safe and secure before public release.

The Commerce Department is to issue guidance to label and watermark AI-generated content to help differentiate between authentic interactions and those generated by software. The extensive order touches on matters of privacy, civil rights, consumer protections, scientific research and worker rights.

White House chief of staff Jeff Zients recalled Biden giving his staff a directive when formulating the order to move with urgency.

ā€œWe canā€™t move at a normal government pace,ā€ Zients said the Democratic president told him. ā€œWe have to move as fast, if not faster, than the technology itself.ā€

In Biden's view, the government was late to address the risks of social media and now U.S. youth are grappling with related mental health issues. AI has the positive ability to accelerate cancer research, model the impacts of climate change, boost economic output and improve government services among other benefits. But it could also warp basic notions of truth with false images, deepen racial and social inequalities and provide a tool to scammers and criminals.

With the European Union nearing final passage of a sweeping law to rein in AI harms and Congress still in the early stages of debating safeguards, the Biden administration is ā€œstepping up to use the levers it can control,ā€ said digital rights advocate Alexandra Reeve Givens, president of the Center for Democracy & Technology. "Thatā€™s issuing guidance and standards to shape private sector behavior and leading by example in the federal governmentā€™s own use of AI.ā€

The order builds on voluntary commitments already made by technology companies. It's part of a broader strategy that administration officials say also includes congressional legislation and international diplomacy, a sign of the disruptions already caused by the introduction of new AI tools such as ChatGPT that can generate text, images and sounds.

The guidance within the order is to be implemented and fulfilled over the range of 90 days to 365 days.

Last Thursday, Biden gathered his aides in the Oval Office to review and finalize the executive order, a 30-minute meeting that stretched to 70 minutes, despite other pressing matters, including the mass shooting in Maine, the Israel-Hamas war and the selection of a new House speaker.

Biden was profoundly curious about the technology in the months of meetings that led up to drafting the order. His science advisory council focused on AI at two meetings and his Cabinet discussed it at two meetings. The president also pressed tech executives and civil society advocates about the technology's capabilities at multiple gatherings.

ā€œHe was as impressed and alarmed as anyone,ā€ deputy White House chief of staff Bruce Reed said in an interview. ā€œHe saw fake AI images of himself, of his dog. He saw how it can make bad poetry. And heā€™s seen and heard the incredible and terrifying technology of voice cloning, which can take three seconds of your voice and turn it into an entire fake conversation.ā€

The issue of AI was seemingly inescapable for Biden. At Camp David one weekend, he relaxed by watching the Tom Cruise film ā€œMission: Impossible ā€” Dead Reckoning Part One.ā€ The film's villain is a sentient and rogue AI known as ā€œthe Entityā€ that sinks a submarine and kills its crew in the movie's opening minutes.

ā€œIf he hadnā€™t already been concerned about what could go wrong with AI before that movie, he saw plenty more to worry about,ā€ said Reed, who watched the film with the president.

Governments around the world have raced to establish protections, some of them tougher than Biden's directives. After more than two years of deliberation, the EU is putting the final touches on a comprehensive set of regulations that targets the riskiest applications with the tightest restrictions. China, a key AI rival to the U.S., has also set some rules.

U.K. Prime Minister Rishi Sunak hopes to carve out a prominent role for Britain as an AI safety hub at a summit starting Wednesday that Vice President Kamala Harris plans to attend. And on Monday, officials from the Group of Seven major industrial nations agreed to a set of AI safety principles and a voluntary code of conduct for AI developers.

The U.S., particularly its West Coast, is home to many of the leading developers of cutting-edge AI technology, including tech giants Google, Meta and Microsoft, and AI-focused startups such as OpenAI, maker of ChatGPT. The White House took advantage of that industry weight earlier this year when it secured commitments from those companies to implement safety mechanisms as they build new AI models.

But the White House also faced significant pressure from Democratic allies, including labor and civil rights groups, to make sure its policies reflected their concerns about AIā€™s real-world harms.

Suresh Venkatasubramanian, a former Biden administration official who helped craft principles for approaching AI, said one of the biggest challenges within the federal government has been what to do about law enforcementā€™s use of AI tools, including at U.S. borders.

ā€œThese are all places where we know that the use of automation is very problematic, with facial recognition, drone technology,ā€ Venkatasubramanian said. Facial recognition technology has been shown to perform unevenly across racial groups, and has been tied to mistaken arrests.

While the EUā€™s forthcoming AI law is set to ban real-time facial recognition in public, Bidenā€™s order appears to simply ask for federal agencies to review how theyā€™re using AI in the criminal justice system, falling short of the stronger language sought by some activists.

The American Civil Liberties Union is among the groups that met with the White House to try to ensure ā€œweā€™re holding the tech industry and tech billionaires accountableā€ so that algorithmic tools ā€œwork for all of us and not just a few,ā€ said ReNika Moore, director of the ACLUā€™s racial justice program, who attended Monday's signing.

After seeing the text of the order, Moore applauded how it addressed discrimination and other AI harms in workplaces and housing, but said the administration ā€œessentially kicks the can down the roadā€ in protecting people from law enforcementā€™s growing use of the technology.


Loading...

Recommended Videos