speech recognition

Machine Translation for User-Generated Content


A specific use case worth exploring in this regard is MT for User Generated Content (UGC). Because of the speed with which UGC (comments, feedback, reviews) is being created and the corresponding costs of its professional translation, many organizations turn to MT. Popular examples of such companies are Skype (in addition to text translation, Microsoft developed the Automatic Speech Recognition (ASR) for audio speech translation in Skype) and Facebook. The social network is aiming to solve the challenge of fine-tuning each system relating to a specific language pair, using neural machine translation (NMT) and benefiting from various contexts for translations. One solution that tackles this issue is the technology developed by Language I/O. It takes into account the client's glossaries and TMs, selects the best MT engine output and then improves on the results using cultural intelligence and/or human linguists who compare machine translations post-facto to ensure that their MT Optimizer engine learns over time.

New AI System Translates Human Brain Signals Into Text With Up to 97% Accuracy


The world is only just getting used to the power and sophistication of virtual assistants made by companies like Amazon and Google, which can decode our spoken speech with eerie precision compared to what the technology was capable of only a few short years ago. In truth, however, a far more impressive and mind-boggling milestone may be just around the corner, making speech recognition seem almost like child's play: artificial intelligence (AI) systems that can translate our brain activity into fully formed text, without hearing a single word uttered. Brain-machine interfaces have evolved in leaps and bounds over recent decades, proceeding from animal models to human participants, and are, in fact, already attempting this very kind of thing. Just not with much accuracy yet, researchers from the University of California San Francisco explain in a new study. To see if they could improve upon that, a team led by neurosurgeon Edward Chang of UCSF's Chang Lab used a new method to decode the electrocorticogram: the record of electrical impulses that occur during cortical activity, picked up by electrodes implanted in the brain.

Extending the power of Azure AI to Microsoft 365 users


Today, Yusuf Mehdi, Corporate Vice President of Modern Life and Devices, announced the availability of new Microsoft 365 Personal and Family subscriptions. In his blog, he shared a few examples of how Microsoft 365 is innovating to deliver experiences powered by artificial intelligence (AI) to billions of users every day. Whether through familiar products like Outlook and PowerPoint, or through new offerings such as Presenter Coach and Microsoft Editor across Word, Outlook, and the web, Microsoft 365 relies on Azure AI to offer new capabilities that make their users even more productive. Azure AI is a set of AI services built on Microsoft's breakthrough innovation from decades of world-class research in vision, speech, language processing, and custom machine learning. What is particularly exciting is that Azure AI provides our customers with access to the same proven AI capabilities that power Microsoft 365, Xbox, HoloLens, and Bing.

Towards an ImageNet Moment for Speech-to-Text


Speech-to-text (STT), also known as automated-speech-recognition (ASR), has a long history and has made amazing progress over the past decade. Currently, it is often believed that only large corporations like Google, Facebook, or Baidu (or local state-backed monopolies for the Russian language) can provide deployable "in-the-wild" solutions. Following the success and the democratization (the so-called "ImageNet moment", i.e. the reduction of hardware requirements, time-to-market and minimal dataset sizes to produce deployable products) of computer vision, it is logical to hope that other branches of Machine Learning (ML) will follow suit. The only questions are, when will it happen and what are the necessary conditions for it to happen? If the above conditions are satisfied, one can develop new useful applications with reasonable costs. Also democratization occurs - one no longer has to rely on giant companies such as Google as the only source of truth in the industry.

Outcomes Rocket Healthcare Using AI and Machine Learning


When you hear the words artificial intelligence, what's the first thing that comes to mind? Driverless cars, Amazon shopping, Netflix movie recommendations and trading software to help bankers. Many think of artificial intelligence in healthcare as a buzz word or just a concept that will fully develop in the near future, but has no impact in your life right now. Some other household examples of current-day technology that use AI include Siri, Alexa, Google Now – these popular speech recognition software assistants all use artificial intelligence! Recently, Alexa was cleared to handle patient information.

Beyond Lipreading: Visual Speech Recognition Looks You in the Eye


Like the lipreading spies of yesteryear peering through their binoculars, almost all visual speech recognition VSR research these days focuses on mouth and lip motion. But a new study suggests that VSR models could perform even better if they used additional available visual information. The VSR field typically looks at the mouth region since it is believed that lip shape and motion contain almost all the information correlated with speech. This has made the information in other facial regions considered as weak by default. But a new paper from the Key Laboratory of Intelligent Information Processing of the Chinese Academy of Sciences and the University of Chinese Academy of Sciences proposes that information from extraoral facial regions can consistently benefit SOTA VSR model performance.

There Is a Racial Divide in Speech-Recognition Systems, Researchers Say


The study tested five publicly available tools from Apple, Amazon, Google, IBM and Microsoft that anyone can use to build speech recognition services. These tools are not necessarily what Apple uses to build Siri or Amazon uses to build Alexa. But they may share underlying technology and practices with services like Siri and Alexa. Each tool was tested last year, in late May and early June, and they may operate differently now. The study also points out that when the tools were tested, Apple's tool was set up differently from the others and required some additional engineering before it could be tested.

Employees urged to turn off smart speakers while working from home during the coronavirus

Daily Mail - Science & tech

Tech companies are known to listen in on private conversation via its smart speakers in order to'improve voice-recognition features.' Now that millions of people are currently working home due to the coronavirus outbreak, employers are urging their stuff to power down the technology in order to keep it from listening to confidential phone calls. Mishcon de Reya LLP, the UK law firm that advised Princess Diana on her divorce, advised staff to mute or shut off listening devices like Amazon's Alexa or Google's voice assistant when they talk about client matters at home, according to a partner at the firm. Video products such as Ring and baby monitors are also on the list of devices to be away of while working from home, as first reported on by Bloomberg. Mishcon de Reya LLP, the UK law firm that advised Princess Diana on her divorce, advised staff to mute or shut off listening devices like Amazon's Alexa or Google's voice assistant when they talk about client matters at home Mishcon de Reya partner Joe Hancock, who also heads the firm's cybersecurity efforts, told Bloombger: 'Perhaps we're being slightly paranoid but we need to have a lot of trust in these organizations and these devices.' 'We'd rather not take those risks.'

Google's New AI Transcribe Feature - A Translating Wiz


Is the improved transcription feature the new replacement of the earlier Google Live Transcribe? The latest audio-to-text translation service is out and about, but only for Android users for the time being. Record the audio in one language and have it rendered in another language altogether! Lengthy discussions can be easily transcribed into text now, without any trouble. January marked the launch of the AI-Powered transcription feature of Google Translate on Android, and now it supports transcribed translations between any of the eight languages, including French, German, Portuguese, English, Thai, Hindi, Spanish, Russian.

10 Artificial Intelligence(AI) Apps That are being Used in Android


Artificial Intelligence is one of the science and technology advancements most common in the world. It explains the application of artificial intelligence & accuracy at the human level. In other words, computers are learning to think and behave like humans. Machines and applications for artificial intelligence have been of great value to humans, helping to provide simpler and more reliable ways to get things completed. Such machines are not subjected to emotional challenges that promote human errors & are not biased in their judgment, nor are they tired.