AI can steal your voice, and there's not much you can do about it

NBC News Clone summarizes the latest on: Ai Voice Cloning Software Flimsy Guardrails Report Finds Rcna195131 - Technology and Innovation | NBC News Clone. This article is rewritten and presented in a simplified tone for a better reader experience.

Voice cloning programs — most of which are free- have flimsy barriers to prevent nonconsensual impersonations, a new report finds.
Man using mobile phone obscured behind glass.
Jasper James / Getty Images

Most leading artificial intelligence voice cloning programs have no meaningful barriers to stop people from nonconsensually impersonating others, a Consumer Reports investigation found.

Voice cloning AI technology has made remarkable strides in recent years, and many services can effectively mimic a person’s cadence with only a few seconds of sample audio. A flashpoint moment came during the Democratic primaries last year, when robocalls of a fake Joe Biden spammed the phones of voters telling them not to vote. The political consultant who admitted to masterminding the scheme was fined $6 million, and the Federal Communications Commission has since banned AI-generated robocalls.

A new survey of the six leading publicly available AI voice cloning tools found that five have easily bypassable safeguards, making it simple to clone a person’s voice without their consent. Deepfake audio detection software often struggles to tell the difference between real and synthetic voices.

Generative AI, which mimics human qualities such as their appearance, writing and voices, is a new and rapidly evolving technology, and the industry has few federal regulations. Most ethical and safety checks in the industry at large are self-imposed. Biden had included some safety demands in his executive order on AI, which he signed in 2023, though President Donald Trump revoked that order when he took office.

Voice cloning technology works by taking an audio sample of a person speaking and then extrapolating that person’s voice into a synthetic audio file. Without safeguards in place, anyone who registers an account can simply upload audio of an individual speaking, such as from a TikTok or YouTube video, and have the service imitate them.

Four of the services — ElevenLabs, Speechify, PlayHT and Lovo — simply require checking a box saying that the person whose voice is being cloned had given authorization.

Another service, Resemble AI, requires recording audio in real time, rather than allowing a person to just upload a recording. But Consumer Reports was able to easily circumvent that restriction by simply playing an audio recording from a computer.

Only the sixth service, Descript, had a somewhat effective safeguard. It requires a would-be cloner to record a specific consent statement, which is difficult to falsify except through cloning through another service.

All six services are available to the public via their websites. Only Eleven Labs and Resemble AI cost money — respectively $5 and $1 — to create a custom voice clone. The others are free.

Some of the companies claim that abuse of their tool can have serious negative consequences.

“We recognize the potential for misuse of this powerful tool and have implemented robust safeguards to prevent the creation of deepfakes and protect against voice impersonation,” a spokesperson for Resemble AI told NBC News in an emailed statement.

There are legitimate uses for AI voice cloning, including helping people with disabilities and creating audio translations of people speaking in different languages. But there is also enormous potential for harm, said Sarah Myers West, co-executive director of the AI Now Institute, a think tank that focuses on the consequences of AI policy.

“This could obviously be used for fraud, scams and disinformation, for example impersonating institutional figures,” West told NBC News.

There is little research on the scope of how often AI is used in audio-based scams. In so-called grandparent scams, a criminal makes a phone call to a person claiming an emergency involving a family member, like they have been kidnapped, arrested or injured. The Federal Trade Commission has warned that such scams may use AI, though the scams predate the technology.

Cloned voices have been used to create music without the depicted artist’s permission, as happened with a viral 2023 song that falsely seemed to be by Drake and the Weeknd, and some musicians have struggled to control their image when other people release music with their voices.

×
AdBlock Detected!
Please disable it to support our content.

Related Articles

Donald Trump Presidency Updates - Politics and Government | NBC News Clone | Inflation Rates 2025 Analysis - Business and Economy | NBC News Clone | Latest Vaccine Developments - Health and Medicine | NBC News Clone | Ukraine Russia Conflict Updates - World News | NBC News Clone | Openai Chatgpt News - Technology and Innovation | NBC News Clone | 2024 Paris Games Highlights - Sports and Recreation | NBC News Clone | Extreme Weather Events - Weather and Climate | NBC News Clone | Hollywood Updates - Entertainment and Celebrity | NBC News Clone | Government Transparency - Investigations and Analysis | NBC News Clone | Community Stories - Local News and Communities | NBC News Clone