title: 'iml: An R package for Interpretable Machine Learning' authors: - affiliation: 1 name: Christoph Molnar orcid: 0000-0003-2331-868X - affiliation: 1 name: Giuseppe Casalicchio orcid: 0000-0001-5324-5966 - affiliation: 1 name: Bernd Bischl orcid: 0000-0001-6002-6980 date: "19 June 2018" output: pdf_document bibliography: paper.bib tags: - R - machine learning - interpretability affiliations: - index: 1 name: Department of Statistics, LMU Munich
Complex, non-parametric models, which are typically used in machine learning, have proven to be successful in many prediction tasks. But these models usually operate as black boxes: While they are good at predicting, they are often not interpretable. Many inherently interpretable models have been suggested, which come at the cost of losing predictive power. Another option is to apply interpretability methods to a black box model after model training. Given the velocity of research on new machine learning models, it is preferable to have model-agnostic tools which can be applied to a random forest as well as to a neural network. Tools for model-agnostic interpretability methods should improve the adoption of machine learning.
iml
is an R package [@R] that offers a general toolbox for making machine learning models interpretable. It
implements many model-agnostic methods which work for any type of machine learning model. The package covers following methods:
iml
was designed to provide a class-based and user-friendly way to
make black box machine learning models interpretable.
Internally, the implemented methods inherit from the same parent class and share a common framework for the computation.
Many of the methods are already implemented in other packages (e.g. [@pdp1], [@ice], [@lime]), but the iml
package implements all of the methods in one place, uses the same syntax and offers consistent functionality and outputs.
iml
can be used with models from the R machine learning libraries mlr
and caret
, but the package is flexible enough to work with models from other packages as well.
Similar projects are the R package DALEX
[@dalex] and the Python package Skater
[@pramit_choudhary_2018_1198885].
The difference to iml
is that the other two projects do not implement the methods themselves, but depend on other packages.
DALEX
focuses more on model comparison, and Skater
additionally includes interpretable models and has less model-agnostic interpretability methods compared to iml
.
The unified interface provided by the iml
package simplifies the analysis and interpretation of black box machine learning learning models.
This work is funded by the Bavarian State Ministry of Science and the Arts in the framework of the Centre Digitisation.Bavaria (ZD.B)
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.