These functions determine the performance of fitted model based on its
predictions. They are used both for evaluating whole modeling procedures and
to tune model parameters, i.e. find the parameter values with the best
performance.
The parameter tuning routine is designed to minimize its error function (or
optimization criteria), which is why functions that are to be maximized must
have their sign changed, like `neg_auc`

.

1 2 3 4 5 6 7 8 9 10 | ```
error_rate(truth, prediction, allow_rejection = !missing(rejection_cost),
rejection_cost)
neg_auc(truth, prediction)
rmse(truth, prediction, na.rm = FALSE)
mse(truth, prediction, na.rm = FALSE)
neg_harrell_c(truth, prediction, na.rm = FALSE)
``` |

`truth` |
The true response values, be it class labels, numeric values or survival outcomes. |

`prediction` |
A prediction object. |

`allow_rejection` |
If |

`rejection_cost` |
See the argument |

`na.rm` |
Whether to remove missing values or not. |

Custom performance estimation functions should be implemented as follows:

`function(truth, prediction)`

`truth`

A vector of true responses.

`prediction`

Prediction returned from the prediction function.

In most cases the true response and the predictions are of the same type,
e.g. true and fitted values in a regression or class labels in a
classification problem, but it is not a requirement. An example of different
types could be if the prediction function produce class probabilities for
all classes rather than one label, or the risks that the observations will
experience the event of interest, to be compared to the actual outcome that
it did occur or has not yet occurred at a specific time point.
See `neg_harrell_c`

for an example of the latter.

Christofer Bäcklin

`emil`

, `neg_gmpa`

,
`modeling_procedure`

, `extension`

Questions? Problems? Suggestions? Tweet to @rdrrHQ or email at ian@mutexlabs.com.

Please suggest features or report bugs with the GitHub issue tracker.

All documentation is copyright its authors; we didn't write any of that.