Speech Recognition

The Ultimate Guide To Speech Recognition With Python – Real Python


Have you ever wondered how to add speech recognition to your Python project? If so, then keep reading! It's easier than you might think. Far from a being a fad, the overwhelming success of speech-enabled products like Amazon Alexa has proven that some degree of speech support will be an essential aspect of household tech for the foreseeable future. If you think about it, the reasons why are pretty obvious. Incorporating speech recognition into your Python application offers a level of interactivity and accessibility that few technologies can match. The accessibility improvements alone are worth considering. Speech recognition allows the elderly and the physically and visually impaired to interact with state-of-the-art products and services quickly and naturally--no GUI needed! Best of all, including speech recognition in a Python project is really simple. In this guide, you'll find out how.

A.I. and speech advances bring virtual assistants to work


ORLANDO, Fla. – Speech recognition technologies have improved so much in recent years – thanks to cloud computing and advances in machine learning – that the virtual assistants created by Amazon, Google and Apple have quickly become popular with consumers. So it should come as little surprise that the underlying natural language technology is making inroads at work, too. "I would say that it [enterprise adoption] is in early stages now, but there are certainly basic capabilities here today," Jon Arnold, of J Arnold & Associates, said at the Enterprise Connect conference last week. The main uses for speech recognition in the office will, at least at first, revolve around improving employee productivity and automating workflows. Thanks to advances in artificial intelligence (A.I.) techniques, the accuracy of speech recognition systems has improved significantly, with Google and others passing the 95% accuracy mark.

IBM turns Watson into a voice assistant for enterprises


IBM is turning its cognitive computing platform Watson into a new voice assistant for enterprises. The Watson Assistant combines two Watson services -- Watson Conversation and Watson Virtual Agent -- to offer businesses a more advanced conversational experience that can be pre-trained on a range of intents. IBM said the aim was to make Watson Assistant an out-of-the-box, white-label type service that's easier for organizations to embed into their consumer-facing offerings than the Watson API. "Everyone was trying to build an assistant using the API, but building a conversational intelligent assistant from the API alone was really hard," said Bret Greenstein, IBM's global vice president for IoT products. "We realized we needed to build a hosted offering for the types of brands that engage with consumers."

IBM's Watson-based voice assistant is coming to cars and smart homes


One of IBM's first partners Harman will demonstrate Watson Assistant at the event through a digital cockpit aboard a Maserati GranCabrio, though the companies didn't elaborate on what it can do. In fact, IBM already released a Watson-powered voice assistant for cybersecurity early last year. You'll be able to access Watson Assistant via text or voice, depending on the device and how IBM's partner decides to incorporate it. So, you'll definitely be using voice if it's a smart speaker, but you might be able to text commands to a home device. Speaking of commands, it wasn't just designed to follow them -- it was designed to learn from your actions and remember your preferences.

Microsoft's Azure gets all emotional with machine learning


Imagine if the things around your house could respond to your voice even when you were shouting over a smoke alarm, keep track of each individual wandering through the house, unlock your front door just by identifying your voice, and even identify your emotions. Those are all capabilities that Microsoft is preparing to add to its Project Oxford, a set of cloud-based machine learning services introduced last May at Microsoft's Build conference. Ars took a deep dive on Project Oxford's first wave of machine learning-based services last year. Those services performed a number of image processing and recognition tasks, offered text-to-speech and speech recognition services, and even converted natural language into intent-based commands for applications. The services are the same technology used in Microsoft's Cortana personal assistant and the Skype Translator service, which translates voice calls in six languages (and text messages in 50 languages) in real-time.

Spotify tests voice assistant sparking rumours of a smart speaker

Daily Mail

Spotify may be about to take on the smart speaker market. The music streaming site is testing an in-app assistant, dubbed'Spotify Voice', that allows users to control their music with their voice. The trial has sparked rumours that the firm is about to release a smart speaker to take on the likes of Apple's HomePod and Amazon's Echo. If the rumours are true, it would allow Spotify to put a microphone and potentially camera in every user's home. Spotify may be about to take on the smart speaker market.

Spotify is testing its own voice assistant to control your music

The Guardian

Spotify is experimenting with a voice-control interface, looking to free itself from reliance on Siri and Alexa and pave the way for the company's forthcoming smart speaker. Users of the service have spotted the new feature hiding in the search bar of Spotify's iOS app. After tapping the magnifying glass to search for a track or playlist, testers see a microphone icon inside a white bubble, according to the Verge. After users tap on the icon, Spotify suggests a number of typical requests for a voice-controlled music system: "Show Calvin Harris", "Play my Discover Weekly" and "Play some upbeat pop", for instance. The move comes as Spotify ramps up its efforts to build a smart speaker to challenge Apple, Amazon and Google in the hardware field, all of which have their own music services.

Voice Assistants: This is How They Will Revolutionize Commerce


The Digital Transformation Institute of Capgemini published in the report titled "Conversational Commerce: Why Consumers Are Embracing Voice Assistants in Their Lives", which highlights how consumers use voice assistants (Google Assistant, Siri, and Alexa above all) and what opportunities these offer systems to companies to connect with their customers. From the report, which was attended by over 5,000 customers in the United States, United Kingdom, France and Germany, it emerges that in the next three years voice assistants will become the predominant mode of interaction between consumers, and that those who use technology to do purchases will be willing to spend 500% more through this new form of interaction than current levels. In fact, consumers are developing a strong preference for interacting with companies through voice assistants. Research has shown that today about a quarter of respondents (24%) would prefer to use a voice assistant instead of a website. However, it is estimated that in the next three years this percentage will grow to 40% and almost a third (31%) will interact with a voice assistant instead of going to a physical store or a bank branch, compared to 20% recorded today.

Microsoft starts testing voice dictation in latest Office apps


Microsoft officials touted back in January a new voice dictation capability for Microsoft Office. On March 12, Microsoft began testing this feature with its Office Insider testers.The @OfficeInsider account tweeted yesterday: Microsoft officials touted the coming Office dictation technology in January, saying it would be available in February 2018. To test dictating using voice, customers must be running the latest version of Office for Windows (Office 2016) and be an Office 365 subscriber. The voice dictation feature, which uses speech recognition technology to convert speech to text, is available for Word 2016, PowerPoint 2016, Outlook 2016 and OneNote 2016 and in U.S. English only for now. To test this, users must be in the Windows desktop Office Insider program.

An Introduction to Snips NLU, the Open Source Library behind Snips Embedded Voice Platform


The Snips Embedded Voice Platform allows any device manufacturer to build a Private by Design voice interface to their product. It handles Wakeword Detection, Speech Recognition, and Natural Language Understanding, fully on-device, so that none of your private voice data goes to the cloud. The Snips platform is also open for non-commercial use, for anyone who'd like to hack a voice assistant at home. Today, the Snips team is making an extra step to promote the use of privacy-preserving Artificial Intelligence: we are fully open sourcing Snips NLU, our Natural Language Understanding library.