# Model residuals audit" In auditor: Model Audit - Verification, Validation, and Error Analysis

knitr::opts_chunk$set(echo = TRUE) knitr::opts_chunk$set(message = FALSE)
knitr::opts_chunk$set(warning = FALSE) This vignette demonstrates how to use the auditor package for auditing residuals of machine learning models. The residual is the difference between the observed value and the value predicted by model. The auditor provides methods for model verification and validation by error analysis. It helps in finding answers to questions that may be crucial in deeper analyses. • Does the model fit the data? Is it not missing any information? • Which model has better performance? • How similar are models? Many algorithms, such as random forests and neutral networks are sometimes treated as black boxes and there is a lack of techniques that help in the analysis of errors in those models. Most methods provided in auditor package are model-agnostic, what means that they can be used regardless of expected distribution of residuals. ## Data Use case - predicting a length of life To illustrate application of auditor we will use dataset "dragons" available in the DALEX package. The dataset contains characteristics of fictional creatures (dragons), like year of birth, height, weight, etc (see below). The goal is to predict the length of life of dragons (a regression problem). library(DALEX) data(dragons) head(dragons) ## Models First, we need models to compare. We selected linear regression and random forest because of their different structures. Linear regression model linear relationships between target response and independent variables, while random forest should be able to capture also non-linear relationships between variables. # Linear regression lm_model <- lm(life_length ~ ., data = dragons) # Random forest library(randomForest) set.seed(59) rf_model <- randomForest(life_length ~ ., data = dragons) ## Preparation for residual (error) analysis Analysis begins with creation of an explainer object with explain function from DALEX package. Explainer wraps a model with its meta-data, such as dataset that was used for training or observed response. lm_exp <- DALEX::explain(lm_model, label = "lm", data = dragons, y = dragons$life_length)
rf_exp <- DALEX::explain(rf_model, label = "rf", data = dragons, y = dragons$life_length) Next step requires creation of model_residual objects of each explainer. From this step on, only auditor functions will be used. library(auditor) lm_mr <- model_residual(lm_exp) rf_mr <- model_residual(rf_exp) ## Plots In the following section, we show individual plotting functions which demonstrate different aspects of residual analysis. We devote more attention to selected functions, but usage of each function is more or less similar. ### Observed vs predicted First plot is a basic plot comparising predicted versus observed values. The red line corresponds to the y = x function. The patterns for both models are non-random around the diagonal line. The points corresponding to a random forest (darker dots) show the tendency to underprediction for large values of observed response. Points for linear model (lighter dots) are located more or less around diagonal line which means that this model predicts quite well. plot(rf_mr, lm_mr, type = "prediction", abline = TRUE) # alternatives: # plot_prediction(rf_mr, lm_mr, abline = TRUE) # plot_prediction(rf_mr, lm_mr, variable = "life_length") Function plot_prediction presents observed values on the x-axis. However, on the x-axis there may be values of any model variable or observations ordered by index (variable = NULL). plot(rf_mr, lm_mr, variable = "scars", type = "prediction") plot(rf_mr, lm_mr, variable = "height", type = "prediction") As you can notice, on above plots, there is no relationship for variable height and predicted values while for increasing number of scars model predictions also increase of life length. This means that that model captured monotonic relationship between number of scars and length of life of dragon. ### Residuals vs observed, fitted or variable values Next function (plot_residual()) shows residuals versus observed values. This plot is used to detect dependence of errors, unequal error variances, and outliers. For appropriate model, residuals should not show any functional dependency. Expected mean value should be equal to 0, regardless of$\hat{y}$values, so any structured arrangement of points suggests a problem with the model. It is worth looking at the observations that clearly differ from the others. If points on the plot are not randomly dispersed around the horizontal axis, it may be presumed that model is not appropriate for the data. plot(lm_mr, rf_mr, type = "residual") # alternative: # plot_residual(lm_mr, rf_mr) Values (residuals) may also be ordered by target variable, fitted values, any other variable or may be presented unordered. plot(rf_mr, lm_mr, type = "residual", variable = "_y_hat_") plot(rf_mr, lm_mr, type = "residual", variable = "scars") # alternative: # plot_residual(rf_mr, lm_mr, variable = "_y_hat_") # plot_residual(rf_mr, lm_mr, variable = "scars") In all examples above, we can see that linear model is better fitted for the data than random forest, because for the latter one greater values of selected variables residuals are also greater. Additionally, we can identify most outlying observations:  {r echo = FALSE} plot_residual(rf_mr, variable = "y_hat", nlabel = 10)  {r} # plot_residual(rf_mr, variable = "_y_hat_", nlabel = 10) ### Density of residuals Residual density plot (plot_residual_density()) detects the incorrect behavior of residuals. The function returns plot with estimated densities of residuals. Their values are displayed as marks along the x axis. For some models, the expected shape of density could be derived from the model assumptions. For example, simple linear model residuals should be normally distributed. However, even if the model does not have an assumption about the distribution of residuals residual density plot may be informative. If most of the residuals are not concentrated around zero, it is likely that the model predictions are biased. plot(rf_mr, lm_mr, type = "residual_density") # alternative # plot_residual_density(rf_mr, lm_mr) Residuals may be also divided by values of a chosen variable (median of a numeric variable or levels of a factor). plot_residual_density(rf_mr, lm_mr, variable = "colour") ### Boxplot of residuals Residual boxplot (plotResidualBoxplot()) shows the distribution of the absolute values of residuals. Boxplot usually presents following values: • box width which corresponds to the second and third quartile, • vertical line which reflects median, • the whiskers which extend to the smallest and largest values, no further than 1.5 of interquartile. auditor adds another component to the boxplot which is the root mean square error (RMSE) measure, shown as X. For the appropriate model, box should lay near zero. A large spread of values indicates problems with a model. Comparing our two models we can see that random forest model is much more spread (worse) than linear one. plot(lm_mr, rf_mr, type = "residual_boxplot") # alternative # plot_residual_boxplot(lm_mr, rf_mr) ### Autocorrelation function of residuals Function plot_acf() creates plot which is helpful in checking randomness of errors. If errors are random, autocorrelations should be near zero for lag separations. If they are non random, then autocorrelations will be significantly higher than zero. The vertical dashed lines indicate boundaries, where autocorrelations are significantly different than zero. plot(lm_mr, type = "acf", variable = "year_of_discovery") # alternative: # plot_acf(lm_mr, variable = "year_of_discovery") ### Autocorrelation of residuals Autocorrelation plot (plot_autocorrelation()) shows correlation of i-th vs i+1-th residuals. This plot may be useful for checking autocorrelation of residuals. plot(rf_mr, type = "autocorrelation") # alternative: # plot_autocorrelation(rf_mr) ..... ..... ..... .... ..... Sometimes it might be difficult to compare two models relying only on visualization. Therefore, two following additional scores may be useful for choosing a better model; scores are based on Durbin-Watson and Runs test statistics. Scores can be calculated with score_dw() and score_runs() functions (or the score() function with score = "dw" or score = "runs"). ..... ..... ..... .... ..... score_dw(rf_exp)$score
score_runs(rf_exp)\$score

### Correlation of models

Function plot_correlation() returns grid of plots with correlation of dependent variable and fitted values of model(s). In the diagonal, there are estimated densities; in the upper triangle two kind of correlations: (1) between models and (2) between predicted and observed values. In the lower triangle, there are scatter plots of these values.

plot(rf_mr, lm_mr, type = "correlation")

# alternative:
# plot_correlation(rf_mr, lm_mr)

### Principal component analysis (PCA) of models

Function plot_pca() can be used to assess the similarity of the models in terms of residuals. The idea of PCA is reducing the dimension of the data set matrix by creating a set of linearly uncorrelated variables called principal components. At the same time, keeping as much variance as possible. Model PCA plot is a biplot. On axis of the plot, there are first two principal components. Grey dots represent observations. Arrows are pointing in the direction of the models projected into a two-dimensional space. The interposition of arrows provides information about the similarity of models in terms of residuals. If they are close to each other, it indicates similar residuals structures.

plot(rf_mr, lm_mr, type = "pca")

# alternative:
# plot_pca(rf_mr, lm_mr)

### Regression error characteristic curve (REC)

Error characteristic curves are a generalization of ROC curves. On the x-axis of the plot there is an error tolerance and on the y axis there is a percentage of observations predicted within the given tolerance. REC curve estimates the cumulative distribution function (CDF) of the error. Area over the REC curve is a biased estimate of the expected error.

plot(rf_mr, lm_mr, type = "rec")

# alternative:
# plot_rec(rf_mr, lm_mr)

### Regression receiver operating characteristic (RROC)

The basic idea of the ROC curve for regression is to show model asymmetry. The RROC is a plot where on x-axis total over-estimation is depicted and on the y-axis total under-estimation.

For RROC curves we use a shift which is an equivalent to the threshold for ROC curves. For each observation we calculate new prediction: \eqn{\hat{y}'=\hat{y}+s} where s is the shift. Therefore, there are different error values for each shift: \eqn{e_i = \hat{y_i}' - y_i}

................... - over-estimation is calculated as: \eqn{OVER= \sum(e_i|e_i>0)} - under-estimation is calculated as: \eqn{UNDER = \sum(e_i|e_i<0)} - the shift equals 0 is represented by a dot ...................

The area over the RROC curve (AOC) equals to the variance of the errors multiplied by \eqn{frac{n^2}{2}}.

plot(rf_mr, lm_mr, type = "rroc")

# alternative:
# plot_rroc(rf_mr, lm_mr)

### Scale location

This plot shows if residuals are spread equally along the ranges of predictors. The presence of any trend suggests that the variance depends on fitted values, which is against the assumption of homoscedasticity. Not every model has an explicit assumption of homogeneous variance, however, the heteroscedasticity may indicates potential problems with the goodness-of-fit.

plot(rf_mr, lm_mr, type = "scalelocation")

# alternative:
# plot_scalelocation(rf_mr, lm_mr)

### Two-sided empirical cumulative distribution function (TSECDF)

Cumulative distribution function for positive and negative residuals. The plot shows the distribution of residuals divided into groups with positive and negative values. It helps to identify the asymmetry of the residuals. Points represent residuals.

plot(rf_mr, lm_mr, type = "tsecdf")

# alternative
# plot_tsecdf(rf_mr, lm_mr)

## Other methods

Other methods and plots are described in following vignettes:

## Try the auditor package in your browser

Any scripts or data that you put into this service are public.

auditor documentation built on July 27, 2021, 1:08 a.m.