Washington is determined to govern AI, but how?
Send a link to a friend
[May 15, 2023]
By Diane Bartz and Jeffrey Dastin
WASHINGTON (Reuters) - U.S. lawmakers are grappling with what guardrails
to put around burgeoning artificial intelligence, but months after
ChatGPT got Washington's attention, consensus is far from certain.
Interviews with a U.S. senator, congressional staffers, AI companies and
interest groups show there are a number of options under discussion.
Some proposals focus on AI that may put people's lives or livelihoods at
risk, like in medicine and finance. Other possibilities include rules to
ensure AI isn't used to discriminate or violate someone's civil rights.
Another debate is whether to regulate the developer of AI or the company
that uses it to interact with consumers. And OpenAI, the startup behind
the chatbot sensation ChatGPT, has discussed a standalone AI regulator.
It's uncertain which approaches will win out, but some in the business
community, including IBM and the U.S. Chamber of Commerce, favor the
approach that only regulates critical areas like medical diagnoses,
which they call a risk-based approach.
If Congress decides new laws are necessary, the U.S. Chamber's AI
Commission advocates that "risk be determined by impact to individuals,"
said Jordan Crenshaw of the Chamber's Technology Engagement Center. "A
video recommendation may not pose as high of a risk as decisions made
about health or finances."
Surging popularity of so-called generative AI, which uses data to create
new content like ChatGPT's human-sounding prose, has sparked concern the
fast-evolving technology could encourage cheating on exams, fuel
misinformation and lead to a new generation of scams.
The AI hype has led to a flurry of meetings, including a White House
visit this month by the CEOs of OpenAI, its backer Microsoft Corp, and
Alphabet Inc. President Joe Biden met with the CEOs.
Congress is similarly engaged, say congressional aides and tech experts.
"Staff broadly across the House and the Senate have basically woken up
and are all being asked to get their arms around this," said Jack Clark,
co-founder of high-profile AI startup Anthropic, whose CEO also attended
the White House meeting. "People want to get ahead of AI, partly because
they feel like they didn't get ahead of social media."
As lawmakers get up to speed, Big Tech's main priority is to push
against "premature overreaction," said Adam Kovacevich, head of the
pro-tech Chamber of Progress.
And while lawmakers like Senate Majority Leader Chuck Schumer are
determined to tackle AI issues in a bipartisan way, the fact is Congress
is polarized, a Presidential election is next year, and lawmakers are
addressing other big issues, like raising the debt ceiling.
Schumer's proposed plan requires independent experts to test new AI
technologies prior to their release. It also calls for transparency and
providing the government with data it needs to avert harm.
[to top of second column]
|
The U.S. Capitol dome is seen from the
Russell Senate Office Building on Capitol Hill in Washington, U.S.,
April 19, 2023. REUTERS/Sarah Silbiger/File Photo
GOVERNMENT MICROMANAGEMENT
The risk-based approach means AI used to diagnose cancer, for
example, would be scrutinized by the Food and Drug Administration,
while AI for entertainment would not be regulated. The European
Union has moved toward passing similar rules.
But the focus on risks seems insufficient to Democratic Senator
Michael Bennet, who introduced a bill calling for a government AI
task force. He said he advocates for a "values-based approach" to
prioritize privacy, civil liberties and rights.
Risk-based rules may be too rigid and fail to pick up dangers like
AI's use to recommend videos that promote white supremacy, a Bennet
aide added.
Legislators have also discussed how best to ensure AI is not used to
racially discriminate, perhaps in deciding who gets a low-interest
mortgage, according to a person following congressional discussions
who is not authorized to speak to reporters.
At OpenAI, staff have contemplated broader oversight.
Cullen O'Keefe, an OpenAI research scientist, proposed in an April
talk at Stanford University the creation of an agency that would
mandate that companies obtain licenses before training powerful AI
models or operating the data centers that facilitate them. The
agency, O'Keefe said, could be called the Office for AI Safety and
Infrastructure Security, or OASIS.
Asked about the proposal, Mira Murati, OpenAI's chief technology
officer, said a trustworthy body could "hold developers accountable"
to safety standards. But more important than the mechanics was
agreement "on what are the standards, what are the risks that you're
trying to mitigate."
The last major regulator to be created was the Consumer Financial
Protection Bureau, which was set up after the 2007-2008 financial
crisis.
Some Republicans may balk at any AI regulation.
"We should be careful that AI regulatory proposals don't become the
mechanism for government micromanagement of computer code like
search engines and algorithms," a Senate Republican aide told
Reuters.
(Reporting by Diane Bartz and Jeffrey Dastin; Editing by Chris
Sanders and Anna Driver)
[© 2023 Thomson Reuters. All rights
reserved.]This material may not be published,
broadcast, rewritten or redistributed.
Thompson Reuters is solely responsible for this content.
|