computing


Artificial Neural Networks: How To Understand Them And Why They're Important

#artificialintelligence

If you dip even a toe into the realm of artificial intelligence, you'll come across artificial neural networks. Artificial neural networks are the systems that power artificial intelligence. It's a type of computer that doesn't just read code that it already understands. Neural networks process vast amounts of information to help create an understanding of what's already right in front of you. People think the key to understanding neural networks is calculus, but this system of computing has roots in biology.


How the Turing Test inspired AI

#artificialintelligence

Computer pioneer and artificial intelligence (AI) theorist Alan Turing would have been 100 years old this Saturday. To mark the anniversary the BBC has commissioned a series of essays. In this, the fourth article, his influence on AI research and the resulting controversy are explored. Alan Turing was clearly a man ahead of his time. In 1950, at the dawn of computing, he was already grappling with the question: "Can machines think?"


Teaching the ghost in the machine: How cloud tech experts are managing the rise of AI

#artificialintelligence

When it comes to the breakthroughs that brilliant scientists and engineers are working on in 2018, artificial intelligence technology somehow manages to be both the most promising and most polarizing development of these times. As a collective, Big Tech is throwing billions of dollars at artificial intelligence, which those involved would rather we all call machine learning. The notion that we can teach computers to learn -- to absorb data, recognize patterns, and take action -- could have an enormous impact on nearly everything we do with a computer, and pave the way for computers to move into new and game-changing places, such as the self-driving car. This technology still has a long way to go, despite the fact we've been talking about it for decades. But it's starting to become real, and alongside that progress has come perhaps one of the biggest backlashes against an aspect of the evolution of information technology.


Fleets using AI to accelerate safety, efficiency

#artificialintelligence

"Artificial intelligence" (AI) may evoke fears of robots writing their own software code and not taking orders from humans. The real AI, at least in present form, is delivering results in the business world. Technology companies are using powerful computers and advanced statistical models to accelerate their product development. Most are not calling these efforts AI but rather machine learning. As a form of AI, machine learning is making it possible to quickly find relevant patterns in data captured by Internet of Things (IoT) devices and sensors, explains Adam Kahn, vice president of fleets for Netradyne, which has a vision-based fleet safety system called Driveri ("driver eye").


Me, Myself and AI: Is That My Privacy in the Rearview Mirror?

#artificialintelligence

I had the pleasure of meeting Sophia in London a few weeks ago. Sophia is a popular, outgoing personality that looks a little bit like Audrey Hepburn. As it happens, Sophia is also a machine. What makes her interesting is that she can carry a conversation. She listens to what you say, shows facial expressions as she speaks, answers your questions, and even asks follow-up questions of her own.


Pushing AI to the Edge: Be Part of the Revolution - DZone AI

#artificialintelligence

This article is featured in the new DZone Guide to Artificial Intelligence: Automating Decision-Making. Get your free copy for more insightful articles, industry statistics, and more! It unlocks our phones, creates our shopping list, navigates our commute, and cleans spam from our email. Once you've experienced AI in action, it's difficult to go back. With edge computing becoming a thing, AI on-the-edge is following suit.


Ultimate guide to handle Big Datasets for Machine Learning using Dask (in Python)

#artificialintelligence

We will now have a look at some simple cases for creating arrays using Dask. As you can see here, I had 11 values in the array and I used the chunk size as 5. This distributed my array into three chunks, where the first and second blocks have 5 values each and the third one has 1 value. Dask arrays support most of the numpy functions. For instance, you can use .sum()


How Smarter Phones Will Transform Tech, Media, and Telecom

#artificialintelligence

Your smartphone knows your wedding anniversary is coming up, and you have a chat with it about how you might celebrate the event. Based on its deep understanding of you and your spouse, the device suggests a romantic weekend in Paris. It knows from your photos and calendar that you got engaged at a small bistro in the 9th arrondissement, and from your travel history it knows you favor a boutique hotel near Parc Monceau and your preferred airline is Air France. It creates an itinerary and presents it to you for your approval. After you've used your smartphone a few times to book trips, you trust the artificial intelligence (AI) that powers it so much that you authorize it to book everything -- and to negotiate on your behalf -- without even checking with you.


Management AI: Natural Language Processing (NLP) and Natural Language Generation (NLG)

#artificialintelligence

One of my biggest complaints about terminology in the industry is the claim that data from conversations is "unstructured data". After all, how do people communicate, either in voice or in a written language, if there was no structure that aids meaning? Syntax is the structure of language, and it clearly aids in defining semantics, or the meaning of the communications. To understand how computers are rapidly improving, it's important to look at how natural language is different from what computers have historically processed. From flat file sequential data storage models to relational databases (RDBMS), there is a decade's long history of rigidly structured data.


Artificial intelligence and the impact on our data centers

#artificialintelligence

It never ceases to amaze how filmmakers are able to introduce concepts that at the time seem so far from reality, but in time those concepts make it into our daily lives. In 1990, the Arnold Schwarzenegger movie Total Recall showed us the "Johnny Cab," a driverless vehicle that took them anywhere they wanted to go. Now, most major car companies are investing millions into bringing this technology to the masses. And thanks to Back to the Future II, where Marty McFly evaded the thugs on a hover board, our kids are now crashing into the furniture (and each other) on something similar to what we saw back in 1989. It was way back in 1968 (which some of us can still remember) when we were introduced to Artificial Intelligence (AI) with HAL 9000, a sentient computer on board the Discovery One Spaceship in 2001: A Space Odyssey.