mirror of
https://github.com/mii443/tokenizers.git
synced 2025-12-08 13:48:19 +00:00
Python - Fix bug in bert wordpiece example script
This commit is contained in:
@@ -36,7 +36,7 @@ tokenizer = BertWordPieceTokenizer(
|
|||||||
)
|
)
|
||||||
|
|
||||||
# And then train
|
# And then train
|
||||||
trainer = tokenizer.train(
|
tokenizer.train(
|
||||||
files,
|
files,
|
||||||
vocab_size=10000,
|
vocab_size=10000,
|
||||||
min_frequency=2,
|
min_frequency=2,
|
||||||
|
|||||||
Reference in New Issue
Block a user