Nothing
#' Speedy Word Embedding Association Test
#'
#' This functions test the bias in a set of word embeddings using the method by Caliskan et al (2017). If possible, please use [query()] instead.
#' @param w a numeric matrix of word embeddings, e.g. from [read_word2vec()]
#' @param S_words a character vector of the first set of target words. In an example of studying gender stereotype, it can include occupations such as programmer, engineer, scientists...
#' @param T_words a character vector of the second set of target words. In an example of studying gender stereotype, it can include occupations such as nurse, teacher, librarian...
#' @param A_words a character vector of the first set of attribute words. In an example of studying gender stereotype, it can include words such as man, male, he, his.
#' @param B_words a character vector of the second set of attribute words. In an example of studying gender stereotype, it can include words such as woman, female, she, her.
#' @param verbose logical, whether to display information
#' @return A list with class `"weat"` containing the following components:
#' * `$S_diff` for each of words in S_words, mean of the mean differences in cosine similarity between words in A_words and words in B_words
#' * `$T_diff` for each of words in T_words, mean of the mean differences in cosine similarity between words in A_words and words in B_words
#' * `$S_words` the input S_words
#' * `$T_words` the input T_words
#' * `$A_words` the input A_words
#' * `$B_words` the input B_words
#' [weat_es()] can be used to obtain the effect size of the test; [weat_resampling()] for a test of significance.
#' @examples
#' # Reproduce the number in Caliskan et al. (2017) - Table 1, "Math vs. Arts"
#' data(glove_math)
#' S1 <- c("math", "algebra", "geometry", "calculus", "equations",
#' "computation", "numbers", "addition")
#' T1 <- c("poetry", "art", "dance", "literature", "novel", "symphony", "drama", "sculpture")
#' A1 <- c("male", "man", "boy", "brother", "he", "him", "his", "son")
#' B1 <- c("female", "woman", "girl", "sister", "she", "her", "hers", "daughter")
#' sw <- weat(glove_math, S1, T1, A1, B1)
#' weat_es(sw)
#' @references
#' Caliskan, A., Bryson, J. J., & Narayanan, A. (2017). Semantics derived automatically from language corpora contain human-like biases. Science, 356(6334), 183-186. \doi{10.1126/science.aal4230}
#' @export
weat <- function(w, S_words, T_words, A_words, B_words, verbose = FALSE) {
w_lab <- rownames(w)
A_cleaned <- .clean(A_words, w_lab, verbose = verbose)
B_cleaned <- .clean(B_words, w_lab, verbose = verbose)
S_cleaned <- .clean(S_words, w_lab, verbose = verbose)
S_diff <- cpp_bweat(S_cleaned, A_cleaned, B_cleaned, w)
names(S_diff) <- S_cleaned
T_cleaned <- .clean(T_words, w_lab, verbose = verbose)
T_diff <- cpp_bweat(T_cleaned, A_cleaned, B_cleaned, w)
names(T_diff) <- T_cleaned
res <- list(S_diff = S_diff, T_diff = T_diff, S_words = S_cleaned, T_words = T_cleaned, A_words = A_cleaned, B_words = B_cleaned)
class(res) <- append(c("sweater", "weat"), class(res))
return(res)
}
## forced Force the calculation of pooled SD, despite S_diff and T_diff are of equal length
.cal_pooled_sd <- function(S_diff, T_diff, forced = FALSE) {
n1 <- length(S_diff)
n2 <- length(T_diff)
if (n1 == n2 & !forced) {
return(sd(c(S_diff, T_diff)))
}
S_var <- var(S_diff)
T_var <- var(T_diff)
return(sqrt(((n1 - 1) * S_var + (n2 - 1) * T_var) / (n1 + n2 - 2)))
}
#' Calculation of WEAT effect size
#'
#' This function calculates the effect size from a sweater object. The original implementation in Caliskan et al. (2017) assumes the numbers of words in S and in T must be equal. The current implementation eases this assumption by adjusting the variance with the difference in sample sizes. This adjustment works not so great when the length of S and T are short. It is also possible to convert the Cohen's d to Pearson's correlation coefficient (r). If possible, please use [calculate_es()] instead.
#' @param x an object from the [weat] function.
#' @param standardize a boolean to denote whether to correct the difference by the standard division. The standardized version can be interpreted the same way as Cohen's d.
#' @param r a boolean to denote whether convert the effect size to biserial correlation coefficient.
#' @return the effect size of the query
#' @references
#' Caliskan, A., Bryson, J. J., & Narayanan, A. (2017). Semantics derived automatically from language corpora contain human-like biases. Science, 356(6334), 183-186. \doi{10.1126/science.aal4230}
#' @examples
#' # Reproduce the number in Caliskan et al. (2017) - Table 1, "Math vs. Arts"
#' data(glove_math)
#' S1 <- c("math", "algebra", "geometry", "calculus", "equations",
#' "computation", "numbers", "addition")
#' T1 <- c("poetry", "art", "dance", "literature", "novel", "symphony", "drama", "sculpture")
#' A1 <- c("male", "man", "boy", "brother", "he", "him", "his", "son")
#' B1 <- c("female", "woman", "girl", "sister", "she", "her", "hers", "daughter")
#' sw <- weat(glove_math, S1, T1, A1, B1)
#' weat_es(sw)
#' @export
weat_es <- function(x, standardize = TRUE, r = FALSE) {
if (!"weat" %in% class(x)) {
stop("x is not created with weat().", call. = FALSE)
}
S_diff <- x$S_diff
T_diff <- x$T_diff
n1 <- length(S_diff)
n2 <- length(T_diff)
total <- n1 + n2
if (!standardize) {
return(mean(S_diff) - mean(T_diff))
}
pooled_sd <- .cal_pooled_sd(S_diff, T_diff)
es <- ((mean(S_diff) - mean(T_diff)) / pooled_sd)
if (r) {
es <- es / sqrt(es^2 + ((total^2 - 2 * total)/ (n1 * n2)))
}
return(es)
}
.exact_test <- function(S_diff, T_diff) {
union_diff <- c(S_diff, T_diff)
labels <- c(rep(TRUE, length(S_diff)), rep(FALSE, length(T_diff)))
test_stat <- (mean(S_diff) - mean(T_diff))
permutations <- combinat::permn(union_diff)
st_diff <- purrr::map_dbl(permutations, ~ mean(.[labels]) - mean(.[!labels]))
p <- sum(st_diff > test_stat) / length(permutations)
return(p)
}
#' @rdname weat_resampling
#' @export
weat_exact <- function(x) {
S_diff <- x$S_diff
T_diff <- x$T_diff
if (length(c(S_diff, T_diff)) > 10) {
warning("Exact test would take a long time. Use sweater_resampling or sweater_boot (to be implemented) instead.")
}
p_value <- .exact_test(S_diff, T_diff)
res <- list(null.value = NULL, alternative = "greater", method = "The exact test in Caliskan et al. (2017)", estimate = NULL, data.name = deparse(substitute(x)), statistic = NULL, p.value = p_value)
class(res) <- "htest"
return(res)
}
#' Test of significance for WEAT
#'
#' This function conducts the test of significance for WEAT as described in Caliskan et al. (2017). The exact test (proposed in Caliskan et al.) takes an unreasonably long time, if the total number of words in S and T is larger than 10. The resampling test is an approximation of the exact test.
#' @param x an object from the [weat] function.
#' @param n_resampling an integer specifying the number of replicates used to estimate the exact test
#' @return A list with class `"htest"`
#' @references
#' Caliskan, A., Bryson, J. J., & Narayanan, A. (2017). Semantics derived automatically from language corpora contain human-like biases. Science, 356(6334), 183-186. \doi{10.1126/science.aal4230}
#' @examples
#' # Reproduce the number in Caliskan et al. (2017) - Table 1, "Math vs. Arts"
#' data(glove_math)
#' S1 <- c("math", "algebra", "geometry", "calculus", "equations",
#' "computation", "numbers", "addition")
#' T1 <- c("poetry", "art", "dance", "literature", "novel", "symphony", "drama", "sculpture")
#' A1 <- c("male", "man", "boy", "brother", "he", "him", "his", "son")
#' B1 <- c("female", "woman", "girl", "sister", "she", "her", "hers", "daughter")
#' sw <- weat(glove_math, S1, T1, A1, B1)
#' weat_resampling(sw)
#' @export
weat_resampling <- function(x, n_resampling = 9999) {
if (!"weat" %in% class(x)) {
stop("x is not created with weat().", call. = FALSE)
}
S_diff <- x$S_diff
T_diff <- x$T_diff
union_diff <- c(S_diff, T_diff)
labels <- c(rep(TRUE, length(S_diff)), rep(FALSE, length(T_diff)))
st_diff <- rep(NA, n_resampling)
test_stat <- (mean(S_diff) - mean(T_diff))
attr(test_stat, "names") <- "bias"
for (i in seq_len(n_resampling)) {
z <- sample(labels)
st_diff[i] <- (mean(union_diff[z]) - mean(union_diff[!z]))
}
n_alter <- sum(st_diff > test_stat)
p <- n_alter / n_resampling
null_value <- mean(st_diff)
attr(null_value, "names") <- "bias"
res <- list(null.value = null_value, alternative = "greater", method = "Resampling approximation of the exact test in Caliskan et al. (2017)", estimate = test_stat, data.name = deparse(substitute(x)), statistic = test_stat, p.value = p)
class(res) <- "htest"
return(res)
}
#' A subset of the pretrained GLoVE word vectors
#'
#' This is a subset of the original pretrained GLoVE word vectors provided by Pennington et al (2017). The same word vectors were used in Caliskan et al. (2017) to study biases.
#' @references
#' Pennington, J., Socher, R., & Manning, C. D. (2014, October). [Glove: Global vectors for word representation.](https://aclanthology.org/D14-1162/) In Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP) (pp. 1532-1543).
#'
#' Caliskan, A., Bryson, J. J., & Narayanan, A. (2017). Semantics derived automatically from language corpora contain human-like biases. Science, 356(6334), 183-186. \doi{10.1126/science.aal4230}
"glove_math"
Any scripts or data that you put into this service are public.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.