MIT CSAIL researchers create video-trained AI that produces realistic sounds

As a way to help machines and robots better understand the objects and environment around them, researchers at MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) have created an algorithm that can effectively learn how to predict sound. According to the researchers’ paper, “The algorithm uses a recurrent neural network to predict sound features from videos and then produces a waveform from these features with an example-based synthesis procedure.” The team simply wanted to show that sounds predicted by their model were realistic enough to fool participants in a “real or fake psychophysical experiment.” While the researchers achieved their goal, there is still some room for improvement so it can help machines or robots understand the world around them. (Related: AI app helps developers plan out their day) Abhinav Gupta,…


Link to Full Article: MIT CSAIL researchers create video-trained AI that produces realistic sounds