Microsoft has developed a groundbreaking artificial intelligence (AI) voice cloning program, VALL-E 2, which can produce speech that sounds identical to a human voice. Despite this technological feat, the company has decided not to make it available to the public because of potential security risks associated with its misuse.
VALL-E 2 has achieved what experts call “human parity,” meaning its generated speech is virtually indistinguishable from that of a real person. This level of sophistication is unprecedented, as the program can mimic voices after hearing just a three-second audio clip. Its ability to replicate the original speaker’s voice with high fidelity marks a significant advancement over previous text-to-speech systems.
However, the potential dangers posed by such a powerful tool have led Microsoft to restrict its use to research purposes only. Researchers expressed concerns that VALL-E 2 could be exploited for malicious activities, including identity theft and impersonation. The program’s capacity to convincingly replicate a person’s voice could be used in scams, fraud, and other deceptive practices.
“There are potential risks in the misuse of the model, such as spoofing voice identification or impersonating a specific speaker,” Microsoft researchers stated. As a result, there are “no plans to incorporate VALL-E 2 into a product or expand access to the public.”
The decision to withhold this technology comes amidst growing concerns about the misuse of AI in various sectors, including politics. Earlier this year, there were incidents of fake robocalls using the voice of President Joe Biden in New Hampshire, highlighting the potential for AI-generated voices to be used in political misinformation campaigns. Some supporters of Biden have suggested using AI to mask his cognitive decline during public appearances.
Microsoft’s cautious approach underscores the ethical and security challenges that accompany advancements in AI. By keeping VALL-E 2 out of public hands, the company aims to prevent potential abuses that could arise from its misuse. This move reflects the broader responsibility of tech companies to balance innovation with the need to protect individuals and society from the unintended consequences of powerful technologies.