| class_check | R Documentation |
The function computes the confusion matrix between the logical output of an outlier detection algorithm and a reference (ground-truth) logical vector. The function also calculates the overall accuracy of the results from the confusion matrix, including recall, precision, and F1-scores for the two classes (regular, versus outlier).
class_check(pred, truth)
pred |
A logical vector with the classification output from an anomaly detection algorithm. |
truth |
A logical vector with the observed classification as a reference (or ground truth). |
The function computes the confusion matrix using the function table. True positive and false negative are successively evaluated to compute overall accuracy, recall, precision, and F1-scores.
An S3 class named checkwise with the confusion matrix, and other accuracy metrices appended as attribues.
attr(, "overall") A numeric value between zero and one with the overall accuracy.
attr(, "recall") A numeric vector of values between zero and one with the recall index for regular and outlier cells.
attr(, "precision") A numeric vector of values between zero and one with the precision index for regular and outlier cells.
attr(, "f1-score") A numeric vector of values between zero and one with the F1-scores for regular and outlier cells.
Luca Sartore drwolf85@gmail.com
## Not run:
# Load the package
library(HRTnomaly)
set.seed(2025L)
# Load the 'toy' data
data(toy)
# Detect cellwise outliers using Cellwise Analysis
res <- cellwise(toy[sample.int(100), ], 0.05, 2L)
class_check(res$outlier, res$anomaly_flag != "")
## End(Not run)
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.