tokenize_text: Tokenize Text

Description Usage Arguments Value Examples

View source: R/cleaning.R

Description

performs cleaning, word stemming, and tokenization on a single line of text using clean_text and the stemming and boosting utilities maintained in the tm package

Usage

1
tokenize_text(text, stopwords = TRUE)

Arguments

text

a line of text

stopwords

logical - should stopwords be removed

Value

a cleaned, stemmed and tokenized character vector

Examples

1
tokenize_text('This is a basic line. It contains UNCLEANED, un-stemmed, un-tokenized text.')

duncankmckinnon/ezRnlp documentation built on Aug. 6, 2020, 1:28 a.m.