Goto

Collaborating Authors

Results


ProtoTree: Addressing the black-box nature of deep learning models

#artificialintelligence

One of the biggest obstacles in the adoption of Artificial Intelligence is that it cannot explain what a prediction is based on. These machine-learning systems are so-called black boxes when the reasoning for a decision is not self-evident to a user. Meike Nauta, Ph.D. candidate at the Data Science group within the EEMCS faculty of the University of Twente, created a model to address the black-box nature of deep learning models. Algorithms can already make accurate predictions, such as medical diagnoses, but they cannot explain how they arrived at such a prediction. In recent years, a lot of attention has been paid to the explainable AI field.


Validate computer vision deep learning models

#artificialintelligence

This code pattern is part of the Getting started with IBM Maximo Visual Inspection learning path. After a deep learning computer vision model is trained and deployed, it is often necessary to periodically (or continuously) evaluate the model with new test data. This developer code pattern provides a Jupyter Notebook that will take test images with known "ground-truth" categories and evaluate the inference results versus the truth. We will use a Jupyter Notebook to evaluate an IBM Maximo Visual Inspection image classification model. You can train a model using the provided example or test your own deployed model.


Beijing AI academy unveils world's largest pre-trained deep learning model

#artificialintelligence

The Beijing Academy of Artificial Intelligence (BAAI) unveiled a newer version of its hyper-scale pre-trained deep learning model, the country's first and the world's largest, at an ongoing AI-themed forum in Beijing, in the latest signal of China's ambition to become a global leader in AI. The latest version of the model, known as Wudao, literally meaning an understanding of natural laws, sports 1.75 trillion parameters, breaking the record of 1.6 trillion previously set by Google's Switch Transformer AI language model, the academy announced Tuesday at the three-day forum that runs through Thursday. Wudao was only initially released in March. Wudao is intended to create cognitive intelligence dually driven by data and knowledge, making machines think like humans and enabling machine cognitive abilities to pass the Turing test, Tang Jie, BAAI's vice director of academics, said during the forum. The newer version of Wudao is both gigantic and smart, featuring its hyper scale, high precision and efficiency.


Finding Best Hyper Parameters For Deep Learning Model

#artificialintelligence

Creating a deep learning model has become an easy task nowadays because of the advent of new efficient and fast working libraries like Keras. One can easily create the model by using different functionalities of Keras but the difficult part is to optimize the model to get higher accuracy. We can tune the hyperparameters to make the model more efficient but sometimes it can be a never-ending process. Storm tuner is a hyperparameter tuner that is used to search for the best hyperparameters for a deep learning neural network. It helps in finding out the most optimized hyperparameters for the model we create in less than 25 trials.


The Five Ways To Build Machine Learning Models

#artificialintelligence

Machine learning is powering most of the recent advancements in AI, including computer vision, natural language processing, predictive analytics, autonomous systems, and a wide range of applications. Machine learning systems are core to enabling each of these seven patterns of AI. In order to move up the data value chain from the information level to the knowledge level, we need to apply machine learning that will enable systems to identify patterns in data and learn from those patterns to apply to new, never before seen data. Machine learning is not all of AI, but it is a big part of it. While building machine learning models is fundamental to today's narrow applications of AI, there are a variety of different ways to go about realizing the same ends.


Deep learning model compression

#artificialintelligence

This post covers model inference optimization or compression in breadth and hopefully depth as of March 2021. This includes engineering topics like model quantization and binarization, more research-oriented topics like knowledge distillation, as well as well-known-hacks. Each year, larger and larger models are able to find methods for extracting signal from the noise in machine learning. In particular, language models get larger every day. These models are computationally expensive (in both runtime and memory), which can be both costly when served out to customers or too slow or large to function in edge environments like a phone. Researchers and practitioners have come up with many methods for optimizing neural networks to run faster or with less memory usage.


The Five Ways To Build Machine Learning Models

#artificialintelligence

Machine learning is powering most of the recent advancements in AI, including computer vision, natural language processing, predictive analytics, autonomous systems, and a wide range of applications. Machine learning systems are core to enabling each of these seven patterns of AI. In order to move up the data value chain from the information level to the knowledge level, we need to apply machine learning that will enable systems to identify patterns in data and learn from those patterns to apply to new, never before seen data. Machine learning is not all of AI, but it is a big part of it. While building machine learning models is fundamental to today's narrow applications of AI, there are a variety of different ways to go about realizing the same ends.


New AI-powered deep learning model to support medical diagnostics

#artificialintelligence

A new deep-learning model can learn to identify diseases from medical scans faster and more accurately, according to new research by a team of University of Alberta computing scientists and the U of A spinoff company MEDO. The breakthrough model is the work of a team of researchers in the Faculty of Science--including the contributions of Pouneh Gorji, a graduate student lost in Flight PS752. Deep learning is a type of machine learning--a subfield of artificial intelligence; deep learning techniques are computer algorithms that find patterns in large sets of data, producing models that can then be used to make predictions.These models work best when they learn from hundreds of thousands or even millions of examples. But the field of medical diagnostics presents a unique challenge, where researchers typically only have access to a few hundred medical scan images for reasons of privacy. "When a deep-learning model is trained with so few instances, its performance tends to be poor," said Roberto Vega, lead author of the study and graduate student in the Department of Computing Science.


How to train your deep learning models in a distributed fashion.

#artificialintelligence

Deep learning algorithms are well suited for large data sets and also training deep learning networks needs large computation power. With GPUs / TPUs easily available on pay per use basis or for free (like Google collab), it is possible today to train a large neural network on cloud-like say Resnet 152 (152 layers) on ImageNet database which has around 14 million images. But is a multi-core GPU-enabled machine just enough to train huge models. Technically yes, but it might take weeks to train the model. So how do we reduce the training time?


Understanding Transformers, the machine learning model behind GPT-3

#artificialintelligence

You know that expression When you have a hammer, everything looks like a nail? Well, in machine learning, it seems like we really have discovered a magical hammer for which everything is, in fact, a nail, and they're called Transformers. Transformers are models that can be designed to translate text, write poems and op eds, and even generate computer code. In fact, lots of the amazing research I write about on daleonai.com is built on Transformers, like AlphaFold 2, the model that predicts the structures of proteins from their genetic sequences, as well as powerful natural language processing (NLP) models like GPT-3, BERT, T5, Switch, Meena, and others. You might say they're more than meets the… ugh, forget it.