epi.tests: Sensitivity, specificity and predictive value of a diagnostic...

Description Usage Arguments Details Value Note Author(s) References Examples

Description

Computes true and apparent prevalence, sensitivity, specificity, positive and negative predictive values and positive and negative likelihood ratios from count data provided in a 2 by 2 table.

Usage

1
2
3
4
5
6
7
epi.tests(dat, conf.level = 0.95)

## S3 method for class 'epi.tests'
print(x, ...)

## S3 method for class 'epi.tests'
summary(object, ...)

Arguments

dat

a vector of length four, an object of class table or an object of class grouped_df from package dplyr containing the individual cell frequencies (see below).

conf.level

magnitude of the returned confidence interval. Must be a single number between 0 and 1.

x, object

an object of class epi.tests.

...

Ignored.

Details

Exact binomial confidence limits are calculated for test sensitivity, specificity, and positive and negative predictive value (see Collett 1999 for details).

Confidence intervals for positive and negative likelihood ratios are based on formulae provided by Simel et al. (1991).

Diagnostic accuracy is defined as the proportion of all tests that give a correct result. Diagnostic odds ratio is defined as how much more likely will the test make a correct diagnosis than an incorrect diagnosis in patients with the disease (Scott et al. 2008). The number needed to diagnose is defined as the number of paitents that need to be tested to give one correct positive test. Youden's index is the difference between the true positive rate and the false positive rate. Youden's index ranges from -1 to +1 with values closer to 1 if both sensitivity and specificity are high (i.e. close to 1).

Value

An object of class epi.tests containing the following:

ap

apparent prevalence.

tp

true prevalence.

se

test sensitivity.

sp

test specificity.

diag.ac

diagnostic accuracy.

diag.or

diagnostic odds ratio.

nndx

number needed to diagnose.

youden

Youden's index.

pv.pos

positive predictive value.

pv.neg

negative predictive value.

lr.pos

likelihood ratio of a positive test.

lr.neg

likelihood ratio of a negative test.

p.rout

the proportion of subjects with the outcome ruled out.

p.rin

the proportion of subjects with the outcome ruled in.

p.fpos

of all the subjects that are truly outcome negative, the proportion that are incorrectly classified as positive (the proportion of false positives).

p.fneg

of all the subjects that are truly outcome positive, the proportion that are incorrectly classified as negative (the proportion of false negative).

Note

----------- ---------- ---------- ----------
Disease + Disease - Total
----------- ---------- ---------- ----------
Test + a b a+b
Test - c d c+d
----------- ---------- ---------- ----------
Total a+c b+d a+b+c+d
----------- ---------- ---------- ----------

Author(s)

Mark Stevenson (Faculty of Veterinary and Agricultural Sciences, The University of Melbourne, Australia). Charles Reynard (School of Medical Sciences, The University of Manchester, United Kingdom).

References

Altman DG, Machin D, Bryant TN, and Gardner MJ (2000). Statistics with Confidence, second edition. British Medical Journal, London, pp. 28 - 29.

Bangdiwala SI, Haedo AS, Natal ML (2008). The agreement chart as an alternative to the receiver-operating characteristic curve for diagnostic tests. Journal of Clinical Epidemiology 61: 866 - 874.

Collett D (1999). Modelling Binary Data. Chapman & Hall/CRC, Boca Raton Florida, pp. 24.

Scott IA, Greenburg PB, Poole PJ (2008). Cautionary tales in the clinical interpretation of studies of diagnostic tests. Internal Medicine Journal 38: 120 - 129.

Simel D, Samsa G, Matchar D (1991). Likelihood ratios with confidence: Sample size estimation for diagnostic test studies. Journal of Clinical Epidemiology 44: 763 - 770.

Greg Snow (2008) Need help in calculating confidence intervals for sensitivity, specificity, PPV & NPV. R-sig-Epi Digest 23(1): 3 March 2008.

Examples

 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
## EXAMPLE 1 (from Scott et al. 2008, Table 1):
## A new diagnostic test was trialled on 1586 patients. Of 744 patients that 
## were disease positive, 670 were test positive. Of 842 patients that were 
## disease negative, 640 were test negative. What is the likeliood ratio of 
## a positive test? What is the number needed to diagnose?

dat.v01 <- c(670,202,74,640)
rval.tes01 <- epi.tests(dat.v01, conf.level = 0.95)
print(rval.tes01)

## Test sensitivity is 0.90 (95% CI 0.88 to 0.92). Test specificity is 
## 0.76 (95% CI 0.73 to 0.79). The likelihood ratio of a positive test 
## is 3.75 (95% CI 3.32 to 4.24). 

## What is the number needed to diagnose?
names(rval.tes01$detail)
rval.tes01$detail$nndx

## The number needed to diagnose is 1.51 (95% CI 1.41 to 1.65). Around 15 
## persons need to be tested to return 10 positive tests.


## EXAMPLE 2:
## Same as Example 1 but showing how a 2 by 2 contingency table can be prepared 
## using tidyverse:

## Not run: 
library(tidyverse)

## Generate a data set listing test results and true disease status:
dis <- c(rep(1, times = 744), rep(0, times = 842))
tes <- c(rep(1, times = 670), rep(0, times = 74), 
   rep(1, times = 202), rep(0, times = 640))
dat.df02 <- data.frame(dis, tes)

tmp.df02 <- dat.df02 %>%
  mutate(dis = factor(dis, levels = c(1,0), labels = c("Dis+","Dis-"))) %>%
  mutate(tes = factor(tes, levels = c(1,0), labels = c("Test+","Test-"))) %>%
  group_by(tes, dis) %>%
  summarise(n = n())
tmp.df02

## View the data in conventional 2 by 2 table format:
pivot_wider(tmp.df02, id_cols = c(tes), names_from = dis, values_from = n)

rval.tes02 <- epi.tests(tmp.df02, conf.level = 0.95)
print(rval.tes02)

## End(Not run)

## Test sensitivity is 0.90 (95% CI 0.88 to 0.92). Test specificity is 
## 0.76 (95% CI 0.73 to 0.79). The likelihood ratio of a positive test 
## is 3.75 (95% CI 3.32 to 4.24).


## EXAMPLE 3:
## A biomarker assay has been developed to identify patients that are at 
## high risk of experiencing myocardial infarction. The assay varies on 
## a continuous scale, from 0 to 1. Researchers believe that a biomarker 
## assay result of greater than or equal to 0.60 renders a patient test 
## positive, that is, at elevated risk of experiencing a heart attack 
## over the next 12 months.

## Generate data consistent with the information provided above. Assume the
## prevalence of high risk subjects in your population is 0.35:
set.seed(1234)
dat.df03 <- data.frame(out = rbinom(n = 200, size = 1, prob = 0.35), 
   bm = runif(n = 200, min = 0, max = 1))

## Classify study subjects as either test positive or test negative 
## according to their biomarker test result:
dat.df03$test <- ifelse(dat.df03$bm >= 0.6, 1, 0)

## Generate a two-by-two table:
dat.tab03 <- table(dat.df03$test, dat.df03$out)[2:1,2:1]
rval.tes03 <- epi.tests(dat.tab03, conf.level = 0.95)
print(rval.tes03)

# What proportion of subjects are ruled out as being at high risk of 
## myocardial infarction?
rval.tes03$detail$p.rout
# Answer: 0.61 (95% CI 0.54 to 0.68).

# What proportion of subjects are ruled in as being at high risk of 
## myocardial infarction?
rval.tes03$detail$p.rin
# Answer: 0.38 (95% CI 0.32 to 0.45).

# What is the proportion of false positive results?
rval.tes03$detail$p.fpos
# Answer: 0.37 (95% CI 0.29 to 0.45).

# What is the proportion of false negative results?
rval.tes03$detail$p.fneg
# Answer: 0.58 (95% CI 0.44 to 0.70).

epiR documentation built on Oct. 11, 2021, 9:08 a.m.