mirror of
https://github.com/mii443/tokenizers.git
synced 2025-12-04 19:58:21 +00:00
Upgrading to black 20.8b1
This commit is contained in:
committed by
Anthony MOI
parent
dc1d0711cf
commit
a410903051
@@ -32,7 +32,10 @@ if not files:
|
||||
|
||||
# Initialize an empty tokenizer
|
||||
tokenizer = BertWordPieceTokenizer(
|
||||
clean_text=True, handle_chinese_chars=True, strip_accents=True, lowercase=True,
|
||||
clean_text=True,
|
||||
handle_chinese_chars=True,
|
||||
strip_accents=True,
|
||||
lowercase=True,
|
||||
)
|
||||
|
||||
# And then train
|
||||
|
||||
Reference in New Issue
Block a user