instructional theory




How time can ruin your most precious machine learning model

#artificialintelligence

Once, potential time dependent effects have been identified you should check if these effects are present in the data and if your data actually covers the necessary time spans to detect it. If the data depends on time there are three basic options to handle it. What experience regarding time have you made in your machine learning projects? Would you like to read a story about the various modelling & feature engineering techniques and validation schemes to include these time effects?


Build Machine Learning Model in Python

#artificialintelligence

Build Machine Learning Model in Python In this video, I will show you how to build a simple machine learning model in Python. Particularly, we will be using the scikit-learn package in Python to build a simple classification model (for classifying Iris flowers) using the random forest algorithm. Build Machine Learning Model in Python In this video, I will show you how to build a simple machine learning model in Python. Particularly, we will be using the scikit-learn package in Python to build a simple classification model (for classifying Iris flowers) using the random forest algorithm.


6 Python Libraries to Interpret Machine Learning Models and Build Trust - Analytics Vidhya

#artificialintelligence

The'SHapley Additive exPlanations' Python library, better knows as the SHAP library, is one of the most popular libraries for machine learning interpretability. The SHAP library uses Shapley values at its core and is aimed at explaining individual predictions. But wait – what are Shapley values? Simply put, Shapley values are derived from Game Theory, where each feature in our data is a player, and the final reward is the prediction. Depending on the reward, Shapley values tell us how to distribute this reward among the players fairly. We won't cover this technique in detail here, but you can refer to this excellent article explaining how Shapley values work: A Unique Method for Machine Learning Interpretability: Game Theory & Shapley Values! The best part about SHAP is that it offers a special module for tree-based models. Considering how popular tree-based models are in hackathons and in the industry, this module makes fast computations, even considering dependent features.


PyTorch Tutorial: How to Develop Deep Learning Models with Python

#artificialintelligence

Predictive modeling with deep learning is a skill that modern developers need to know. PyTorch is the premier open-source deep learning framework developed and maintained by Facebook. At its core, PyTorch is a mathematical library that allows you to perform efficient computation and automatic differentiation on graph-based models. Achieving this directly is challenging, although thankfully, the modern PyTorch API provides classes and idioms that allow you to easily develop a suite of deep learning models. In this tutorial, you will discover a step-by-step guide to developing deep learning models in PyTorch. PyTorch Tutorial – How to Develop Deep Learning Models Photo by Dimitry B., some rights reserved. The focus of this tutorial is on using the PyTorch API for common deep learning model development tasks; we will not be diving into the math and theory of deep learning. For that, I recommend starting with this excellent book.


Explaining machine learning models to the business

#artificialintelligence

Explainable machine learning is a sub-discipline of artificial intelligence (AI) and machine learning that attempts to summarize how machine learning systems make decisions. Summarizing how machine learning systems make decisions can be helpful for a lot of reasons, like finding data-driven insights, uncovering problems in machine learning systems, facilitating regulatory compliance, and enabling users to appeal -- or operators to override -- inevitable wrong decisions. Of course all that sounds great, but explainable machine learning is not yet a perfect science. Figure 1: Explanations created by H2O Driverless AI. These explanations are probably better suited for data scientists than for business users.


Introducing GIG: A Practical Method for Explaining Diverse Ensemble Machine Learning Models

#artificialintelligence

Machine learning is proven to yield better underwriting results and mitigate bias in lending. But not all machine learning techniques, including the wide swath at work in unregulated uses, is built to be transparent. Many of the algorithms that get deployed generate results that are difficult to explain. Recently, researchers have proposed novel and powerful methods for explaining machine learning models, notably Shapley Additive Explanations (SHAP Explainers) and Integrated Gradients (IG). These methods provide mechanisms for assigning credit to the data variables used by a model to generate a score.


Machine Learning: Model Selection and Hyperparameter Tuning

#artificialintelligence

In my day-to-day research, a problem I would face quite often is selecting a proper statistical model that fits my data. Mostly, I would be using statistical models for smoothing out erroneous signals from DNA data and I believe it is a common concern among Data Science enthusiasts to pick a model to explain the behavior of data. I thought it would be useful to share how one might tackle such situations using the libraries that are available off-the-shelf. Furthermore, selecting a model is not enough unless you know the most suitable hyperparameters to progress. Let's look at a simple scenario to understand how it can be done.


Leader Stochastic Gradient Descent for Distributed Training of Deep Learning Models

Neural Information Processing Systems

We consider distributed optimization under communication constraints for training deep learning models. We propose a new algorithm, whose parameter updates rely on two forces: a regular gradient step, and a corrective direction dictated by the currently best-performing worker (leader). Our method differs from the parameter-averaging scheme EASGD in a number of ways: (i) our objective formulation does not change the location of stationary points compared to the original optimization problem; (ii) we avoid convergence decelerations caused by pulling local workers descending to different local minima to each other (i.e. to the average of their parameters); (iii) our update by design breaks the curse of symmetry (the phenomenon of being trapped in poorly generalizing sub-optimal solutions in symmetric non-convex landscapes); and (iv) our approach is more communication efficient since it broadcasts only parameters of the leader rather than all workers. We provide theoretical analysis of the batch version of the proposed algorithm, which we call Leader Gradient Descent (LGD), and its stochastic variant (LSGD). Finally, we implement an asynchronous version of our algorithm and extend it to the multi-leader setting, where we form groups of workers, each represented by its own local leader (the best performer in a group), and update each worker with a corrective direction comprised of two attractive forces: one to the local, and one to the global leader (the best performer among all workers).