black-box machine
IIT Jodhpur scientist analyses the explainability of black-box machine learning models
An IIT Jodhpur researcher who works in the cutting-edge field of Artificial Engineering, analysed the problem of explainability of black-box machine learning models. The research shows how the current philosophy of explainable machine learning suffers from certain limitations that have led to a proliferation of black-box models. Machine learning is one of the most sought-after areas of study today. In that context, the explainability of machine learning models represents a fundamental problem. The main aim of this research was to develop more transparent (explainable) machine learning models that can be deployed in several practical applications where intelligent prediction and analysis are required.
LIME -- Explaining Any Machine Learning Prediction
The main goal of the LIME package is to explain any black-box machine learning models. It is used for both classification and regression problems. Let's try to understand why we need to explain machine learning predictions. Consider you are working for a housing finance or bank client. You are tasked with building a machine learning model to predict loan defaults.
The dangers of trusting black-box machine learning
This article is part of our reviews of AI research papers, a series of posts that explore the latest findings in artificial intelligence. Last November, Apple ran into trouble after customers pointed out on Twitter that its credit card service was discriminating against women. David Heinemeir Hansson, the creator of Ruby on Rails, called Apple Card a sexist program. "Apple's black box algorithm thinks I deserve 20x the credit limit [my wife] does," he tweeted. The @AppleCard is such a fucking sexist program.
- Transportation > Air (0.74)
- Banking & Finance > Credit (0.69)
- Health & Medicine (0.47)
- Law > Business Law (1.00)
- Law > Statutes (0.84)
- Government > Regional Government > Europe Government > Poland Government (0.39)