Over the past decade or so, advances in machine learning have paved the way for the development of increasingly advanced speech recognition tools. By analyzing audio files of human speech, these tools can learn to identify words and phrases in different languages, converting them into a machine-readable format. While several machine learning-based models have achieved promising results on speech recognition tasks, they do not always perform well in all languages. For instance, when a language has a vocabulary with many similar-sounding words, the performance of speech recognition systems can decline considerably. Researchers at Mahatma Gandhi Mission's College of Engineering & Technology and Jaypee Institute of Information Technology, in India, have developed a speech recognition system to tackle this problem.
Next time you hear a voice generated by Baidu's Deep Voice 2, you might not be able to tell whether it's human. Baidu, the Beijing-based juggernaut that commands 80 percent of the Chinese internet search market, is investing heavily in artificial intelligence. In 2013, it opened the Institute of Deep Learning, an R&D center focused on machine learning. And in May, it took the wraps off the newest version of Deep Voice, its AI-powered text-to-speech engine. Deep Voice 2, which follows on the heels of Deep Voice's public debut earlier this year, can produce real-time speech that's nearly indistinguishable from a human voice.
Co-located in Silicon Valley and Beijing, Baidu Research brings together top talent from around the world to focus on future-looking fundamental research in artificial intelligence. Our research directions include deep learning, computer vision, speech recognition and synthesis, natural language understanding, data mining and knowledge discovery, business intelligence, artificial general intelligence, high performance computing, robotics and autonomous driving. At Baidu Research, we aim to revolutionize human-machine interfaces with the latest artificial intelligence techniques. Our Deep Voice project was [...] The AAAI (Association for the Advancement of Artificial Intelligence) is one of the world's premiere artificial conferences, with annual summits [...] Today, we are excited to announce the hiring of three world-renowned artificial intelligence scientists, Dr. Kenneth Church, Dr. Jun Huan [...]
Tencent's New Lab Shows It's Serious about Mastering AI Tencent has established an AI lab in Seattle, and the company is building a very serious research team back in China. One of China's biggest tech companies, Tencent, is establishing an AI research lab in Seattle, demonstrating a growing determination to master a technology that looks set to define the future of many industries. Tencent is already one of China's dominant tech companies. It operates the hugely successful mobile chat app WeChat--which boasts over 889 million active users in China--along with lots of other social tools, e-commerce services, games, and the like. Based in Shenzhen, a manufacturing hub in the southeastern part of the country, Tencent has the potential to become a key player in the development and commercialization of artificial intelligence.
Daniel Povey, the main developer of the widely used open-source speech recognition toolkit Kaldi, tweeted today that he is likely joining Chinese smartphone giant Xiaomi at its Beijing headquarters to work on a next generation "PyTorch-y Kaldi." I am very close to signing an agreement to work for Xiaomi in Beijing. Would leave before end of 2019, and would hire a small team there to work on next-gen PyTorch-y' Kaldi. Povey is a leader in voice recognition research, known for his contributions to speech recognition and language processing technologies. He and other researchers first created Kaldi as part of a Johns Hopkins University workshop in 2009.