mirror of
https://github.com/mii443/tokenizers.git
synced 2025-08-23 16:49:27 +00:00
Update README.md
This commit is contained in:
committed by
Anthony MOI
parent
d788a950ac
commit
b35d514785
@ -54,7 +54,7 @@ from tokenizers.pre_tokenizers import Whitespace
|
|||||||
tokenizer.pre_tokenizer = Whitespace()
|
tokenizer.pre_tokenizer = Whitespace()
|
||||||
```
|
```
|
||||||
|
|
||||||
Then training your tokenizer on a set of files jsut takes two lines of codes:
|
Then training your tokenizer on a set of files just takes two lines of codes:
|
||||||
|
|
||||||
```python
|
```python
|
||||||
from tokenizers.trainers import BpeTrainer
|
from tokenizers.trainers import BpeTrainer
|
||||||
|
Reference in New Issue
Block a user