| SentencePieceTokenizer | R Documentation | 
SentencePiece tokenizer for 'lang'
SentencePieceTokenizer(
  lang = "en",
  special_toks = NULL,
  sp_model = NULL,
  vocab_sz = NULL,
  max_vocab_sz = 30000,
  model_type = "unigram",
  char_coverage = NULL,
  cache_dir = "tmp"
)
| lang | lang | 
| special_toks | special_toks | 
| sp_model | sp_model | 
| vocab_sz | vocab_sz | 
| max_vocab_sz | max_vocab_sz | 
| model_type | model_type | 
| char_coverage | char_coverage | 
| cache_dir | cache_dir | 
None
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.