Google's Mobile Vision now gains the ability to read text. Google has introduced a new Text API for its Mobile Vision framework that allows Android developers to integrate optical-character recognition (OCR) into their apps. The new Text API appears in the recently-updated Google Play Services version 9.2, which restores Mobile Vision, Google's system to make it easy for developers to add facial detection and barcode-reading functionality to Android apps. The Text OCR technology currently can recognize text in any Latin-based language, covering most European languages, including English, German, and French, as well as Turkish. Google has added Word Lens, a technology acquired last year, to its Google Translate app.
Who knows in future you may be able to emote to your friend's FB posts in your own wide mouthed haha, open mouthed wow or a puckered brow frown? Going by the indications, things are inching towards such a reality. Facebook's latest acquisition – of a face recognition company FacioMetrics -- has become the talk of the tech town regarding the possibilities of inclusion of facial gesture controls on the app front. Close on the heels of this merger, this startup's apps have been withdrawn from the App Store and Play Store. 'Intraface', the facial image analysis app from Faciometrics could enable detection of seven facial emotions.
Facebook could one day build facial gesture controls for its app thanks to the acquisition of a Carnegie Mellon University spinoff company called FacioMetrics. The startup made an app called Intraface that could detect seven different emotions in people's faces, but it's been removed from the app stores. The acquisition aligns with a surprising nugget of information Facebook slipped into a 32-bullet point briefing sent to TechCrunch this month. "Future applications of deep learning platform on mobile: Gesture-based controls, recognize facial expressions and perform related actions" It's not hard to imagine Facebook one day employing FacioMetrics' tech and its own AI to let you add a Like or one of its Wow/Haha/Angry/Sad emoji reactions by showing that emotion with your face. "How people share and communicate is changing and things like masks and other effects allow people to express themselves in fun and creative ways.
Supervised learning needs labels, or annotations, that tell the algorithm what the right answers are in the training phases of your project. In fact, many of the examples of using MXNet, TensorFlow, and PyTorch start with annotated data sets you can use to explore the various features of those frameworks. Unfortunately, when you move from the examples to application, it's much less common to have a fully annotated set of data at your fingertips. This tutorial will show you how you can use Amazon Mechanical Turk (MTurk) from within your Amazon SageMaker notebook to get annotations for your data set and use them for training. TensorFlow provides an example of using an Estimator to classify irises using a neural network classifier.
Apple's annual Worldwide Developers Conference is always full of surprises. USA TODAY's Jefferson Graham previews what we can expect in Talking Tech. The new iPhone X is seen on display at the Apple Union Square store on Nov. 3, 2017, in San Francisco. The iPhone X's lush screen, facial-recognition skills and $1,000 price tag are breaking new ground in Apple's marquee product line. Now, the much-anticipated device is testing the patience of consumers and investors as demand outstrips suppliers' capacity.