Nothing
#' Extract noun phrases from texts using spaCy
#'
#' This function extracts noun phrases from documents, based on the
#' `noun_chunks` attributes of documents objects parsed by spaCy (see
#' <https://spacy.io/usage/linguistic-features#noun-chunks>).
#'
#' @param x a character object or a TIF-compliant corpus data.frame (see
#' <https://github.com/ropenscilabs/tif>)
#' @inheritParams spacy_parse
#' @param output type of returned object, either `"data.frame"` or
#' `"list"`
#' @param ... unused
#' @details When the option `output = "data.frame"` is selected, the
#' function returns a `data.frame` with the following fields.
#' \describe{\item{`text`}{contents of noun-phrase}
#' \item{`root_text`}{contents of root token}
#' \item{`start_id`}{serial number ID of starting token. This number
#' corresponds with the number of `data.frame` returned from
#' `spacy_tokenize(x)` with default options.}
#' \item{`root_id`}{serial number ID of root token}
#' \item{`length`}{number of words (tokens) included in a noun-phrase (e.g.
#' for a noun-phrase, "individual car owners", `length = 3`)}}
#'
#' @return either a `list` or `data.frame` of tokens
#' @export
#' @examples
#' \dontrun{
#' spacy_initialize()
#'
#' txt <- c(doc1 = "Natural language processing is a branch of computer science.",
#' doc2 = "Paul earned a postgraduate degree from MIT.")
#' spacy_extract_nounphrases(txt)
#' spacy_extract_nounphrases(txt, output = "list")
#' }
spacy_extract_nounphrases <- function(x, output = c("data.frame", "list"),
multithread = TRUE, ...) {
UseMethod("spacy_extract_nounphrases")
}
#' @importFrom data.table data.table
#' @export
spacy_extract_nounphrases.character <- function(x,
output = c("data.frame", "list"),
multithread = TRUE, ...) {
`root_id` <- `start_id` <- `:=` <- NULL
output <- match.arg(output)
if (!is.null(names(x))) {
docnames <- names(x)
} else {
docnames <- paste0("text", 1:length(x))
}
if (length(x) == 1) {
multithread <- FALSE
}
if (all(!duplicated(docnames)) == FALSE) {
stop("Docnames are duplicated.")
} else if (all(nchar(docnames) > 0L) == FALSE) {
stop("Some docnames are missing.")
}
if (is.null(options()$spacy_initialized)) spacy_initialize()
spacyr_pyexec("try:\n del spobj\nexcept NameError:\n 1")
spacyr_pyexec("texts = []")
if (spacyr_pyget("py_version") != 3) {
message("multithreading for python 2 is not supported by spacy_tokenize()")
multithread <- FALSE
}
x <- gsub("\\\\n", "\\\n", x) # replace two quotes \\n with \n
x <- gsub("\\\\t", "\\\t", x) # replace two quotes \\t with \t
x <- gsub("\\\\", "", x) # delete unnecessary backslashes
x <- unname(x)
## send documents to python
spacyr_pyassign("texts", x)
spacyr_pyassign("docnames", docnames)
spacyr_pyassign("multithread", multithread)
## run noun phrase extraction
spacyr_pyexec("spobj = spacyr()")
if (identical(output, "list")) {
command_str <- paste("noun_phrases = spobj.extract_nounphrases_list(texts = texts,",
"docnames = docnames,",
"multithread = multithread)")
spacyr_pyexec(command_str)
return(spacyr_pyget("noun_phrases"))
} else {
command_str <- paste("noun_phrases = spobj.extract_nounphrases_dataframe(texts = texts,",
"docnames = docnames,",
"multithread = multithread)")
spacyr_pyexec(command_str)
noun_phrases <- spacyr_pyget("noun_phrases")
doc_id <- names(noun_phrases)
data_out <-
data.table::rbindlist(lapply(doc_id, function(x) {
df <- as.data.frame(noun_phrases[[x]], stringsAsFactors = FALSE)
if (nrow(df) == 0) return(NULL)
df$doc_id <- x
return(df)
}))
if (nrow(data_out) == 0) {
message("No noun phrase found in documents")
return(NULL)
}
data_out[, start_id := start_id + 1][, root_id := root_id + 1]
data.table::setDF(data_out)
data_out <- data_out[, c(6, 1:5)]
return(data_out)
}
}
#' @export
spacy_extract_nounphrases.data.frame <- function(x, ...) {
# insert compliance check here - replace with tif package
if (!all(c("doc_id", "text") %in% names(x)))
stop("input data.frame does not conform to the TIF standard")
txt <- x$text
names(txt) <- x$doc_id
spacy_extract_nounphrases(txt, ...)
}
Any scripts or data that you put into this service are public.
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.