Soon You Won’t Be Able to Tell an AI From a Human Voice

The choppy, cybernetic voices of digital assistants like Siri may not sound so mechanical for much longer, thanks to a significant breakthrough in using artificial intelligence to generate realistic human speech. In a new paper, scientists at Google-owned AI shop DeepMind have unveiled WaveNet, a neural network that generates audio waveforms by predicting and adapting to its own output in real-time. The result is dramatically more natural-sounding computerized speech, which the researchers say reduces the perceived gap between human and computer voices speaking both English and Chinese by over 50 percent. The system’s predictive model is a far cry from the synthesized speech systems used by “digital assistant” apps like Siri. Instead of using a “concatenative” speech system that pieces together from a library of speech fragments recorded by one…


Link to Full Article: Soon You Won’t Be Able to Tell an AI From a Human Voice

Pin It on Pinterest

Share This

Join Our Newsletter

Sign up to our mailing list to receive the latest news and updates about homeAI.info and the Informed.AI Network of AI related websites which includes Events.AI, Neurons.AI, Awards.AI, and Vocation.AI

You have Successfully Subscribed!