library(gt4ireval) knitr::opts_chunk$set(fig.dpi = 96, fig.width = 5, fig.asp = 1 / 1.4)
gt4ireval
is a package to measure the reliability of an Information Retrieval test collection. It allows users to estimate reliability using Generalizability Theory [@Brennan2001] and map those estimates onto well-known indicators such as Kendall $\tau$ correlation or sensitivity. For background information and details, the reader is referred to [@Urbano2013].
Once loaded, gt4ireval
needs initial evaluation data to run a G-study and the corresponding D-study. These data need to be in a standard data frame or matrix, where columns correspond to systems and rows correspond to queries [^1]. For this vignette, let us use data from the TREC-3 Ad hoc track.
dim(adhoc3) adhoc3[1:5, 1:5]
If your data is transposed (i.e. columns correspond to queries and rows correspond to systems), you can get the correct format with the t
function: data <- t(data)
.
To run a G-study with the initial data we have, we simply call function g.study
.
gstudy(adhoc3)
Additionally, we can tell the function to ignore the systems with lowest average effectiveness scores by setting parameter drop
. For instance, we can ignore the bottom 25\% of systems.
adhoc3.g <- gstudy(adhoc3, drop = .25) adhoc3.g
The summary shows the estimated variance components: variance due to the system effect $\hat\sigma_s^2=0.0028$, due to the query effect $\hat\sigma_q^2=0.0281$, and due to the system-query interaction effect $\hat\sigma_e^2=0.0102$. The second row shows the same values but as a fraction of the total variance. The third row shows the estimated Mean Squares for each component, and finally the sample size in each case. In our example, we have 30 systems and 50 queries as initial data.
The results from the G-study above can now be used to run a D-study. First, let us estimate the stability of the current collection (50 queries).
dstudy(adhoc3.g)
The summary first shows how dstudy
was called. In particular, it tells us that the target number of queries is $n_q'=50$ (set by default from the G-study initial data), the target stability is $\pi=0.95$ (set by default), and the confidence level is $\alpha=0.025$ (set by default).
Next are the estimated stability scores; the relative stability with 50 queries is $\text{E}\hat\rho^2=0.93265$ with a 95\% confidence interval of $[0.89311, 0.96287]$, and the absolute stability is $\hat\Phi=0.78613$ with a 95\% confidence interval of $[0.66141, 0.88039]$.
Regarding the required number of queries to reach the target stability, the estimate is $\hat{n}_q'=69$ with a 95\% confidence interval of $[37, 114]$ to reach $\text{E}\rho^2=\pi$, and $\hat{n}_q'=259$ with a 95\% confidence interval of $[130, 487]$ to reach $\Phi=\pi$.
Function dstudy
can be called with multiple values for $n_q'$, $\pi$ and $\alpha$ to study trends. For instance, we can indicate several query set sizes by setting parameter queries
.
dstudy(adhoc3.g, queries = seq(20, 200, 20))
The output above shows the estimated stability scores, with confidence intervals, for various query set sizes. For example, we have $\text{E}\hat\rho^2=0.96515$ with 100 queries, and $\hat\Phi\in[0.88654, 0.96715]$ with 95\% confidence when having 200 queries.
Similarly, we may indicate several target stability scores by setting parameter stability
.
dstudy(adhoc3.g, stability = c(0.8, 0.85, 0.9, 0.95, 0.97, 0.99))
The output above shows that the estimated number of queries to reach $\text{E}\rho^2=0.97$ is 117, while 123 are required to reach $\Phi=0.9$.
Finally, we can also indicate several confidence levels for the computation of confidence intervals by setting parameter alpha
[^2].
dstudy(adhoc3.g, alpha = c(0.005, 0.025, 0.05))
The summary above shows that with 50 queries a 99\% confidence interval for $\text{E}\rho^2$ is $[0.87737, 0.96967]$, and a 90\% confidence interval on the number of queries to reach $\Phi=0.95$ is $[145, 439]$.
Both gstudy
and dstudy
return objects with all results from the analysis so they can be used in subsequent computations. In fact, object adhoc3.g
above contains all the G-study results, and it is provided to function d.study
.
The full list of available data in both objects can be obtained with function names
.
adhoc3.g <- gstudy(adhoc3, drop = 0.25) names(adhoc3.g) adhoc3.g$var.s adhoc3.d <- dstudy(adhoc3.g, queries = seq(10, 100, 10), stability = seq(0.5, 0.99, .05)) names(adhoc3.d) adhoc3.d$Erho2 cbind(lwr = adhoc3.d$n.q_Phi.lwr, upr = adhoc3.d$n.q_Phi.upr)
With all these data we can for instance plot the estimated $\text{E}\hat\rho^2$ score, with a 95\% confidence interval, as a function of the number of queries in the collection.
xx <- seq(10, 200, 5) adhoc3.d <- dstudy(adhoc3.g, queries = xx) plot(xx, adhoc3.d$Erho2, yaxs = "i", ylim = c(0.75, 1), lwd = 2, type = "l", xlab = "Number of queries", ylab = "Relative stability") lines(xx, adhoc3.d$Erho2.lwr) # lower confidence limit lines(xx, adhoc3.d$Erho2.upr) # upper confidence limit grid()
Finally, the following functions can be used to map stability indicators from Generalizability Theory onto well-known data-based indicators (see [@Urbano2013] for details):
gt2tau
and gt2tauAP
map $\text{E}\rho^2$ onto Kendall $\tau$ correlation and $AP$ correlation coefficients.gt2power
, gt2minor
and gt2major
map $\text{E}\rho^2$ onto expected power, minor conflict rate and major conflict rate of 2-tailed t-tests.gt2asens
and gt2rsens
map $\text{E}\rho^2$ and $\Phi$ onto absolute and relative sensitivity, respectively.gt2rmse
maps $\Phi$ onto rooted mean squared error.gt2tau(Erho2 = 0.95) gt2rsens(Phi = 0.8)
The results show that the estimated rank correlation at $\text{E}\rho^2=0.95$ is $\hat\tau=0.86412$, and that the relative sensitivity at $\Phi=0.8$ is estimated as $\hat\delta_r=12.389\%$.
In order to map the stability of a certain D-study, we can simply use the returned dstudy
object.
These functions can be used for instance to plot the estimated $\hat\tau$ correlation as a function of the query set size.
xx <- seq(10, 200, 5) adhoc3.d <- dstudy(adhoc3.g, queries = xx) plot(xx, gt2tau(adhoc3.d$Erho2), yaxs = "i", ylim = c(0.5, 1), lwd = 2, type = "l", xlab = "Number of queries", ylab = "Kendall rank correlation") lines(xx, gt2tau(adhoc3.d$Erho2.lwr)) # lower confidence limit lines(xx, gt2tau(adhoc3.d$Erho2.upr)) # upper confidence limit grid()
In any case, the user is strongly advised to take these mappings with a grain of salt (see Fig. 3 in [@Urbano2013]).
This work was supported by an A4U postdoctoral grant and a Juan de la Cierva postdoctoral fellowship.
[^1]: For general information on how to read data in R
, the reader is referred to the R Data Import/Export manual.
[^2]: Recall that $100(1-2\alpha)\%$ intervals are computed, so for an 80\% confidence interval we set $\alpha=0.1$.
Any scripts or data that you put into this service are public.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.