Video mimicking Harris' voice raises concerns about AI in politics

Video mimicking Harris' voice raises concerns about AI in politics

Technology

It claims Harris is a “diversity hire” because she is a woman and a person of color

Follow on
Follow us on Google News
 

NEW YORK (AP) - A manipulated video that mimics the voice of Vice President Kamala Harris saying things she did not say is raising concerns about the power of artificial intelligence to mislead with Election Day about three months away.

The video gained attention after tech billionaire Elon Musk shared it on his social media platform X on Friday evening without explicitly noting it was originally released as parody.

The video uses many of the same visuals as a real ad that Harris, the likely Democratic president nominee, released last week launching her campaign. But the video swaps out the voice-over audio with another voice that convincingly impersonates Harris.

“I, Kamala Harris, am your Democrat candidate for president because Joe Biden finally exposed his senility at the debate,” the voice says in the video.

It claims Harris is a “diversity hire” because she is a woman and a person of color, and it says she doesn’t know “the first thing about running the country.” The video retains “Harris for President” branding. It also adds in some authentic past clips of Harris.

Mia Ehrenberg, a Harris campaign spokesperson, said in an email to The Associated Press: “We believe the American people want the real freedom, opportunity and security Vice President Harris is offering; not the fake, manipulated lies of Elon Musk and Donald Trump.” 

The widely shared video is an example of how lifelike AI-generated images, videos or audio clips have been utilized both to poke fun and to mislead about politics as the United States draws closer to the presidential election.

It exposes how, as high-quality AI tools have become far more accessible, there remains a lack of significant federal action so far to regulate their use, leaving rules guiding AI in politics largely to states and social media platforms.

The video also raises questions about how to best handle content that blurs the lines of what is considered an appropriate use of AI, particularly if it falls into the category of satire.

The original user who posted the video, a YouTuber known as Mr Reagan, has disclosed both on YouTube and on X that the manipulated video is a parody.

But Musk’s post, which has been viewed more than 123 million times, according to the platform, only includes the caption “This is amazing” with a laughing emoji.

X users who are familiar with the platform may know to click through Musk’s post to the original user’s post, where the disclosure is visible. Musk’s caption does not direct them to do so.

While some participants in X’s “community note” feature to add context to posts have suggested labeling Musk’s post, no such label had been added to it as of Sunday afternoon.

Some users online questioned whether his post might violate X’s policies, which say users “may not share synthetic, manipulated, or out-of-context media that may deceive or confuse people and lead to harm.”

The policy has an exception for memes and satire as long as they do not cause “significant confusion about the authenticity of the media.”

Musk endorsed former President Donald Trump, the Republican nominee, earlier this month. Neither Mr Reagan nor Musk immediately responded to emailed requests for comment Sunday.

Two experts who specialize in AI-generated media reviewed the fake ad’s audio and confirmed that much of it was generated using AI technology.

One of them, University of California, Berkeley, digital forensics expert Hany Farid, said the video shows the power of generative AI and deepfakes.

“The AI-generated voice is very good,” he said in an email. “Even though most people won’t believe it is VP Harris’ voice, the video is that much more powerful when the words are in her voice.”

He said generative AI companies that make voice-cloning tools and other AI tools available to the public should do better to ensure their services are not used in ways that could harm people or democracy.

Rob Weissman, co-president of the advocacy group Public Citizen, disagreed with Farid, saying he thought many people would be fooled by the video.

“I don’t think that’s obviously a joke,” Weissman said in an interview. “I’m certain that most people looking at it don’t assume it’s a joke.

The quality isn’t great, but it’s good enough. And precisely because it feeds into preexisting themes that have circulated around her, most people will believe it to be real.”

Weissman, whose organization has advocated for Congress, federal agencies and states to regulate generative AI, said the video is “the kind of thing that we’ve been warning about.”

Other generative AI deepfakes in both the U.S. and elsewhere would have tried to influence voters with misinformation, humor or both. In Slovakia in 2023, fake audio clips impersonated a candidate discussing plans to rig an election and raise the price of beer days before the vote.

In Louisiana in 2022, a political action committee’s satirical ad superimposed a Louisiana mayoral candidate’s face onto an actor portraying him as an underachieving high school student.

Congress has yet to pass legislation on AI in politics, and federal agencies have only taken limited steps, leaving most existing U.S. regulation to the states.

More than one-third of states have created their own laws regulating the use of AI in campaigns and elections, according to the National Conference of State Legislatures.

Beyond X, other social media companies also have created policies regarding synthetic and manipulated media shared on their platforms.

Users on the video platform YouTube, for example, must reveal whether they have used generative artificial intelligence to create videos or face suspension.