paper/paper.md

title: 'iml: An R package for Interpretable Machine Learning' authors: - affiliation: 1 name: Christoph Molnar orcid: 0000-0003-2331-868X - affiliation: 1 name: Giuseppe Casalicchio orcid: 0000-0001-5324-5966 - affiliation: 1 name: Bernd Bischl orcid: 0000-0001-6002-6980 date: "19 June 2018" output: pdf_document bibliography: paper.bib tags: - R - machine learning - interpretability affiliations: - index: 1 name: Department of Statistics, LMU Munich

Summary

Complex, non-parametric models, which are typically used in machine learning, have proven to be successful in many prediction tasks. But these models usually operate as black boxes: While they are good at predicting, they are often not interpretable. Many inherently interpretable models have been suggested, which come at the cost of losing predictive power. Another option is to apply interpretability methods to a black box model after model training. Given the velocity of research on new machine learning models, it is preferable to have model-agnostic tools which can be applied to a random forest as well as to a neural network. Tools for model-agnostic interpretability methods should improve the adoption of machine learning.

iml is an R package [@R] that offers a general toolbox for making machine learning models interpretable. It implements many model-agnostic methods which work for any type of machine learning model. The package covers following methods:

iml was designed to provide a class-based and user-friendly way to make black box machine learning models interpretable. Internally, the implemented methods inherit from the same parent class and share a common framework for the computation. Many of the methods are already implemented in other packages (e.g. [@pdp1], [@ice], [@lime]), but the iml package implements all of the methods in one place, uses the same syntax and offers consistent functionality and outputs. iml can be used with models from the R machine learning libraries mlr and caret, but the package is flexible enough to work with models from other packages as well. Similar projects are the R package DALEX [@dalex] and the Python package Skater [@pramit_choudhary_2018_1198885]. The difference to iml is that the other two projects do not implement the methods themselves, but depend on other packages. DALEX focuses more on model comparison, and Skater additionally includes interpretable models and has less model-agnostic interpretability methods compared to iml.

The unified interface provided by the iml package simplifies the analysis and interpretation of black box machine learning learning models.

Acknowledgements

This work is funded by the Bavarian State Ministry of Science and the Arts in the framework of the Centre Digitisation.Bavaria (ZD.B)

References



christophM/iml documentation built on April 23, 2024, 1:25 a.m.