mirror of
https://github.com/mii443/tokenizers.git
synced 2025-09-01 14:59:20 +00:00
We use 19.10b0 not 20 here...
This commit is contained in:
@ -67,10 +67,7 @@ class TestByteLevelBPE:
|
||||
|
||||
def test_lowerspace(self, roberta_files):
|
||||
tokenizer = ByteLevelBPETokenizer.from_files(
|
||||
roberta_files["vocab"],
|
||||
roberta_files["merges"],
|
||||
add_prefix_space=True,
|
||||
lowercase=True,
|
||||
roberta_files["vocab"], roberta_files["merges"], add_prefix_space=True, lowercase=True,
|
||||
)
|
||||
output = tokenizer.encode("The Quick Brown Fox Jumps Over The Lazy Dog")
|
||||
|
||||
|
Reference in New Issue
Block a user