Speech recognition technology promises to dramatically alter our relationship with machines, making devices and the internet easier to use and more accessible. However, current systems are not sufficiently accurate or natural for users to rely on them. This is especially critical in developing economies where new users will connect to the internet for the first time with a mobile device. As part of our mission to create technologies that let us have an impact on hundreds of millions of people, our team in the Silicon Valley AI Lab at Baidu is creating a new type of speech recognition system, called DeepSpeech, that seeks to significantly improve speech recognition performance and eventually make devices as natural to interact with as other humans. In this talk, I will describe how scalability and Deep Learning are driving progress in AI, enabling powerful end-to-end systems like DeepSpeech to reach new levels of performance.
Adam Coates is Director of the Silicon Valley AI Lab at Baidu Research in Sunnyvale.