mirror of
https://github.com/mii443/tokenizers.git
synced 2025-12-05 20:28:22 +00:00
Implement suggestions by @sebpuetz
Co-authored-by: Sebastian Pütz <sebastian.puetz@uni-tuebingen.de>
This commit is contained in:
@@ -61,7 +61,7 @@ impl PreTokenizer {
|
||||
.into_py()?;
|
||||
|
||||
Ok(pretokenized
|
||||
.get_normalized(true)
|
||||
.get_normalized(tk::OffsetReferential::Original)
|
||||
.into_iter()
|
||||
.map(|(s, o)| (s.to_owned(), o))
|
||||
.collect())
|
||||
|
||||
Reference in New Issue
Block a user