Description Usage Arguments Value Author(s) See Also Examples
An n-gram is used for tokenization. This function can also be used to limit the total number of tokens.
1 | text_token(text, ngram_min = 1, ngram_max = 1, num_tokens)
|
text |
text data |
ngram_min |
minimum size of an n-gram (default: 1) |
ngram_max |
maximum size of an n-gram (default: 1) |
num_tokens |
maximum number of tokens |
token |
result of tokenizing text |
ngram_min |
minimum size of an n-gram |
ngram_max |
maximum size of an n-gram |
Dongmin Jung
tm::removeWords, stopwords::stopwords, textstem::lemmatize_strings, text2vec::create_vocabulary, text2vec::prune_vocabulary
1 2 3 4 5 6 7 8 9 |
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.