Description Usage Arguments Value Examples
Function perf_eval
directly compares a desired output
with the scores from diffusion. It handles the possible
shapes of the scores (vector, matrix, list of matrices)
and gives the desired metrics.
1 2 3 4 5 | perf_eval(
prediction,
validation,
metric = list(auc = metric_fun(curve = "ROC"))
)
|
prediction |
smoothed scores; either a named numeric vector, a column-wise matrix whose rownames are nodes and colnames are different scores, or a named list of such matrices. |
validation |
target scores to which the smoothed scores will be compared to. Must have the same format as the input scores, although the number of rows may vary and only the matching rows will give a performance measure. |
metric |
named list of metrics to apply. Each metric should accept
the form |
A data frame containing the metrics for each comparable pair of output-validation.
1 2 3 4 5 6 7 8 9 10 11 |
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.