SentencePieceTokenizer | R Documentation |
SentencePiece tokenizer for 'lang'
SentencePieceTokenizer(
lang = "en",
special_toks = NULL,
sp_model = NULL,
vocab_sz = NULL,
max_vocab_sz = 30000,
model_type = "unigram",
char_coverage = NULL,
cache_dir = "tmp"
)
lang |
lang |
special_toks |
special_toks |
sp_model |
sp_model |
vocab_sz |
vocab_sz |
max_vocab_sz |
max_vocab_sz |
model_type |
model_type |
char_coverage |
char_coverage |
cache_dir |
cache_dir |
None
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.