Lots of tech companies including Apple, Google, Microsoft, NVIDIA and Intel itself have created chips for image recognition and other deep-learning chores. However, Intel is taking another tack as well with an experimental chip called "Loihi." Rather than relying on raw computing horsepower, it uses an old-school, as-yet-unproven type of "nueromorphic" tech that's modeled after the human brain. Intel has been exploring neuromorphic tech for awhile, and even designed a chip in 2012. Instead of logic gates, it uses "spiking neurons" as a fundamental computing unit.
Toyota is teaming up with Intel, and an assortment of tech and automotive firms, to develop an ecosystem for connected cars. By sharing self-driving vehicle data, the companies aim to develop maps and improved driver assistance systems based on cloud computing. Rounding out the alliance (dubbed the "Automotive Edge Computing Consortium") will be Ericsson, Japanese auto parts-maker Denso Corp, and telecoms firm NTT DoCoMo. Practically everyone is wading into the autonomous car space. And, collaboration between firms is just as common.
Tesla may have promised that all its newly-made vehicles from October 2016 onward would have the groundwork for self-driving capabilities, but that doesn't mean its technology is set in stone. Electrek has learned that Tesla is quietly equipping new Model 3, S and X production units with upgraded Autopilot hardware (HW 2.5). Don't put your barely-used P100D up for sale, though, as this isn't a night-and-day upgrade. Although Electrek says the new gear includes a secondary node to enable more computing power, a spokesperson says 2.5 is really about adding "computing and wiring redundancy" that "very slightly" boosts reliability. Every HW 2.0 or later car should still have the foundations for self-driving functionality, in other words.
Microsoft just announced a new initiative called AI for Earth. Headed by Microsoft Research's computational ecologist Lucas Joppa, the program will help researchers and organizations use AI to solve the major environmental issues we face today. Leaders of projects focusing on water, agriculture, biodiversity and climate change can apply for access to Microsoft's cloud and AI computing resources and it's putting down $2 million towards the initiative this year. In a statement, Microsoft President Brad Smith said, "Our goal is to empower others in new and more impactful ways to help create a more sustainable future. This program expands our commitments to democratizing AI and advancing sustainability around the globe."
Lens, similarly, uses computer vision and AI to make sense of your photos, videos and the real world. Now, with the advantage of improved computer vision algorithms, better cameras and more capable devices and networks, Google is in a much better place to make AR an essential computing tool. With Lens's more advanced capabilities, we're seeing just how far Google's AR technology has come. We're already seeing Amazon invest in that with Alexa and its Echo devices; you can think of its upcoming camera-equipped Echo Look as a big step towards refining its computer vision capabilities.
"That was anticlimactic," Jason Les said with a smirk, getting up from his seat. Unlike nearly everyone else in Pittsburgh's Rivers Casino, Les had just played his last few hands against an artificially intelligent opponent on a computer screen. After his fellow players -- Daniel McAulay next to him and Jimmy Chou and Dong Kim in an office upstairs -- eventually did the same, they started to commiserate. The consensus: That AI was one hell of a player. The four of them had spent the last 20 days playing 120,000 hands of heads-up, no-limit Texas Hold'em against an artificial intelligence called Libratus created by researchers at Carnegie Mellon University.
The Internet of Things still mostly feels like the Internet of Crap, but there's one ray of hope in the connected home on which plenty of companies are jumping aboard: Amazon's Alexa. In particular, I'm talking about the voice-powered ecosystem that's quickly grown around the company's virtual assistant. At CES, we saw Alexa integration in Dish's upcoming Hopper, washing machines and, much to our delight, a dancing robot. It's even heading outside the home: Both Ford and Volkswagen are bringing Alexa to their cars for hands-free commands. In a little over two years, Alexa has gone from being a baffling product (a connected speaker from Amazon, really?!) to an essential feature for any connected device.
Our self-driving future will initially be extremely expensive. That's why GM and Ford are working on autonomous systems for ride-hailing ahead of selling cars to individuals. Meanwhile, Korean automaker Hyundai is researching another approach: a system that uses less computing power and therefore is cheaper. Of course this vehicle, like all autonomous cars, won't be available for a very long time, but what Hyundai showed off in Las Vegas looks promising. The two test Ioniqs (one hybrid and one pure electric) were fitted with cameras in the windshield, radar behind the automaker's logo and lidar sensors in the front and sides of the bumper.
What do you envision being the next major breakthrough for PC form factor? We're working on lots of things that are mind-blowing. To me, we have to figure out how to get to J.A.R.V.I.S. [Iron Man's trusty AI, not Intel's vaporware earpiece]. The ability to manipulate things wherever you are, look at things wherever you are, talk to things in a more natural way. That's the next big breakthrough in computing. And it will be in so many domains, it won't just be PCs.