knitr::opts_chunk$set( collapse = TRUE, comment = "#>", warning = FALSE, message = FALSE )
Agreement and reliability are related to correlation, but they are not the same problem. Correlation describes co-movement. Agreement describes similarity on the measurement scale itself. Reliability describes the proportion of variation attributable to stable differences among subjects rather than to measurement error or method disagreement.
This vignette covers the wide-data functions:
ccc()ba()icc()Lin's concordance correlation coefficient combines precision and accuracy in a single number. Bland-Altman analysis separates that question into estimated bias and limits of agreement.
library(matrixCorr) set.seed(40) ref <- rnorm(50, mean = 100, sd = 10) m1 <- ref + rnorm(50, sd = 2) m2 <- ref + 1.2 + rnorm(50, sd = 3) fit_ba <- ba(m1, m2) fit_ccc <- ccc(data.frame(m1 = m1, m2 = m2), ci = TRUE) print(fit_ba) summary(fit_ccc)
The two summaries are complementary rather than redundant. ccc() gives a
single concordance coefficient, while ba() makes the scale of disagreement
explicit.
icc() extends the wide-data reliability workflow in two directions. It can
return a pairwise matrix across method pairs, or it can return the overall
classical ICC table for the full set of methods.
wide_methods <- data.frame( J1 = ref + rnorm(50, sd = 1.5), J2 = ref + 4.0 + rnorm(50, sd = 1.8), J3 = ref - 3.0 + rnorm(50, sd = 2.0), J4 = ref + rnorm(50, sd = 1.6) ) fit_icc_pair <- icc( wide_methods, model = "twoway_random", type = "agreement", unit = "single", scope = "pairwise" ) fit_icc_overall <- icc( wide_methods, model = "twoway_random", type = "agreement", unit = "single", scope = "overall", ci = TRUE ) print(fit_icc_pair, digits = 2) summary(fit_icc_pair) print(fit_icc_overall)
This is the most important distinction in the ICC interface.
scope = "pairwise" answers:
"How reliable is each specific pair of methods?"
scope = "overall" answers:
"How reliable is the full set of methods when analysed jointly?"
Those are different quantities. The overall ICC cannot, in general, be recovered by averaging the pairwise matrix.
This simulation also includes systematic method bias, so it is a natural place
to contrast type = "consistency" with type = "agreement".
fit_icc_cons <- icc( wide_methods, model = "twoway_random", type = "consistency", unit = "single", scope = "overall", ci = FALSE ) fit_icc_agr <- icc( wide_methods, model = "twoway_random", type = "agreement", unit = "single", scope = "overall", ci = FALSE ) data.frame( type = c("consistency", "agreement"), selected_coefficient = c( attr(fit_icc_cons, "selected_coefficient"), attr(fit_icc_agr, "selected_coefficient") ), estimate = c( attr(fit_icc_cons, "selected_row")$estimate, attr(fit_icc_agr, "selected_row")$estimate ) )
Consistency discounts additive method shifts, whereas agreement penalises them. When methods differ mainly by a systematic offset, consistency can therefore look substantially better than agreement.
The classical ICC family is controlled by three arguments.
model selects the one-way, two-way random, or two-way mixed formulation.type selects consistency or agreement.unit selects single-measure or average-measure reliability.For pairwise ICC, average-measure output uses k = 2 because each estimate is
based on exactly two methods. For overall ICC, average-measure output uses the
full number of analysed columns.
In practice these methods answer different questions.
ccc() when one concordance coefficient per pair is the main target.ba() when the size and direction of disagreement should be visible on
the original measurement scale.icc() when the target is reliability under a classical variance
components interpretation.There is overlap in interpretation, but these are not interchangeable estimators.
Any scripts or data that you put into this service are public.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.