mirror of
https://github.com/mii443/tokenizers.git
synced 2025-08-22 16:25:30 +00:00
Python - Replace last BPETokenizer occurences
This commit is contained in:
@ -42,7 +42,7 @@ Start using in a matter of seconds:
|
||||
```python
|
||||
# Tokenizers provides ultra-fast implementations of most current tokenizers:
|
||||
>>> from tokenizers import (ByteLevelBPETokenizer,
|
||||
BPETokenizer,
|
||||
CharBPETokenizer,
|
||||
SentencePieceBPETokenizer,
|
||||
BertWordPieceTokenizer)
|
||||
# Ultra-fast => they can encode 1GB of text in ~20sec on a standard server's CPU
|
||||
|
Reference in New Issue
Block a user