mirror of
https://github.com/mii443/tokenizers.git
synced 2025-12-09 06:08:22 +00:00
Update python example
This commit is contained in:
@@ -69,4 +69,6 @@ print(f"Transformer tokenizer took: {end - start} sec")
|
|||||||
ids_r = [ [ token.id for token in sentence ] for sentence in encoded_r ]
|
ids_r = [ [ token.id for token in sentence ] for sentence in encoded_r ]
|
||||||
assert(ids_r == encoded_p)
|
assert(ids_r == encoded_p)
|
||||||
|
|
||||||
print(f"Decoded sentences: {tok_r.decode_batch(ids_r)}")
|
decoded_r = tok_r.decode_batch(ids_r)
|
||||||
|
print(f"Decoded sentences: {decoded_r}")
|
||||||
|
assert(decoded_r == text)
|
||||||
|
|||||||
Reference in New Issue
Block a user