| permshap | R Documentation |
Permutation SHAP algorithm with respect to a background dataset, see Strumbelj and Kononenko (2014) for the basic idea.
By default, for up to p=8 features, exact SHAP values are returned (exact with respect to the selected background data). Otherwise, the sampling process iterates until the resulting values are sufficiently precise, and standard errors are provided.
permshap(object, ...)
## Default S3 method:
permshap(
object,
X,
bg_X = NULL,
pred_fun = stats::predict,
feature_names = colnames(X),
bg_w = NULL,
bg_n = 200L,
exact = length(feature_names) <= 8L,
low_memory = length(feature_names) > 15L,
tol = 0.01,
max_iter = 10L * length(feature_names),
parallel = FALSE,
parallel_args = NULL,
verbose = TRUE,
seed = NULL,
...
)
## S3 method for class 'ranger'
permshap(
object,
X,
bg_X = NULL,
pred_fun = NULL,
feature_names = colnames(X),
bg_w = NULL,
bg_n = 200L,
exact = length(feature_names) <= 8L,
low_memory = length(feature_names) > 15L,
tol = 0.01,
max_iter = 10L * length(feature_names),
parallel = FALSE,
parallel_args = NULL,
verbose = TRUE,
seed = NULL,
survival = c("chf", "prob"),
...
)
object |
Fitted model object. |
... |
Additional arguments passed to |
X |
|
bg_X |
Background data used to integrate out "switched off" features,
often a subset of the training data (typically 50 to 500 rows).
In cases with a natural "off" value (like MNIST digits),
this can also be a single row with all values set to the off value.
If no |
pred_fun |
Prediction function of the form |
feature_names |
Optional vector of column names in |
bg_w |
Optional vector of case weights for each row of |
bg_n |
If |
exact |
If |
low_memory |
If |
tol |
Tolerance determining when to stop. As in CL21, the algorithm keeps
iterating until |
max_iter |
If the stopping criterion (see |
parallel |
If |
parallel_args |
Named list of arguments passed to
|
verbose |
Set to |
seed |
Optional integer random seed. Note that it changes the global seed. |
survival |
Should cumulative hazards ("chf", default) or survival
probabilities ("prob") per time be predicted? Only in |
During each iteration, the algorithm cycles twice through a random permutation: It starts with all feature components "turned on" (i.e., taking them from the observation to be explained), then gradually turning off components according to the permutation. When all components are turned off, the algorithm - one by one - turns the components back on, until all components are turned on again. This antithetic scheme allows to evaluate Shapley's formula twice per feature using a single permutation and a total of 2p disjoint evaluations of the contribution function.
For models with interactions up to order two, one can show that even a single iteration provides exact SHAP values for all features (with respect to the given background dataset).
The Python implementation "shap" uses a similar approach, but without providing standard errors, and without early stopping.
For faster convergence, we use balanced permutations in the sense that
p subsequent permutations each start with a different feature.
Furthermore, the 2p on-off vectors with sum <=1 or >=p-1 are evaluated only once,
similar to the degree 1 hybrid in kernelshap().
An object of class "kernelshap" with the following components:
S: (n \times p) matrix with SHAP values or, if the model output has
dimension K > 1, a list of K such matrices.
X: Same as input argument X.
baseline: Vector of length K representing the average prediction on the
background data.
bg_X: The background data.
bg_w: The background case weights.
m_exact: Number of on-off vectors evaluated once per row of X.
exact: Logical flag indicating whether calculations are exact or not.
txt: Summary text.
predictions: (n \times K) matrix with predictions of X.
algorithm: "permshap".
m: Number of sampled on-off vectors evaluated per iteration (if not exact).
SE: Standard errors corresponding to S (if not exact).
n_iter: Integer vector of length n providing the number of iterations
per row of X (if not exact).
converged: Logical vector of length n indicating convergence per row of X
(if not exact).
permshap(default): Default permutation SHAP method.
permshap(ranger): Permutation SHAP method for "ranger" models, see Readme for an example.
Erik Strumbelj and Igor Kononenko. Explaining prediction models and individual predictions with feature contributions. Knowledge and Information Systems 41, 2014.
# MODEL ONE: Linear regression
fit <- lm(Sepal.Length ~ ., data = iris)
# Select rows to explain (only feature columns)
X_explain <- iris[-1]
# Calculate SHAP values
s <- permshap(fit, X_explain)
s
# MODEL TWO: Multi-response linear regression
fit <- lm(as.matrix(iris[, 1:2]) ~ Petal.Length + Petal.Width + Species, data = iris)
s <- permshap(fit, iris[3:5])
s
# Note 1: Feature columns can also be selected 'feature_names'
# Note 2: Especially when X is small, pass a sufficiently large background data bg_X
s <- permshap(
fit,
iris[1:4, ],
bg_X = iris,
feature_names = c("Petal.Length", "Petal.Width", "Species")
)
s
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.