Engineers have taught an AI the basics of driving in '15 to 20 minutes' – a process that can take some humans dozens of hours behind the wheel. Wayve, which was founded by researchers from Cambridge University's engineering department, used a technique known as'reinforcement learning' to achieve the feat. This teaches the algorithm using trial and error, with correct decisions rewarded with uninterrupted driving, and mistakes being corrected by a safety driver in the car. As the test progressed, the algorithm behind the wheel learnt not to replicate any mistakes that had been corrected by the human safety driver in the car. According to the Wayve team, the AI learnt to drive and corner while staying inside its own lane within '15 to 20 minutes' after it first took to the roads.
The co-founder of Google's DeepMind has slammed self-driving cars for not being safe enough, saying current early tests on public roads are irresponsible. Demis Hassabis has urged developers to be cautious with the new technology, saying it is difficult to prove systems are safe before putting them on public roads. The issue of AI in self-driving cars has flared up this year following the death of a women hit but a self-driving Uber in March. The accident was the first time a pedestrian was killed on a public road by an autonomous car, which had previously been praised as the safer alternative to a traditional car. Speaking at the Royal Society in London, Dr Hassabis said current driverless car programmes could be putting people's lives in danger.
Forget finding the perfect lighting to snap a selfie, Adobe is using AI and machine learning to create the ultimate portrait. Powered by Adobe Sensei, the app combines perspective effect editing, automatic software-only photo masking and photo style transfer technology. The system transforms a bad image into a flattering selfie by altering the perspective of the person's face, adding depth of field and replicating styles of other photos online. Forget finding the perfect lighting to snap a selfie, Adobe is using AI and machine learning to create the ultimate portrait.Powered by Adobe Sensei, the app combines perspective effect editing, automatic software-only photo masking and photo style transfer technology Adobe's new selfie app uses AI and machine learning to edit photos. The 3D-facial mapping lets users adjust their pose and appearance.
It works rather like the BB-8 robot from Star Wars, and could change the way we drive. Goodyear has revealed a radical new spherical tire powered by AI and linked to the car by magnetic force so it can rotate on any axis in any direction. The firm says it will be able to sense road conditions and adapt accordingly, turning itself into either a wet or dry weather configuration instantly. Working like human muscles, the smart tire can re-shape the individual sections of the tire's tread design, adding'dimples' for wet conditions (left) or smoothing the tread for dry conditions (right) Made of super-elastic polymer, the tire's bionic skin has a flexibility similar to that of human skin, allowing it to expand and contract. This outer layer covers a foam-like material that is strong enough to remain flexible despite the weight of a vehicle.