mirror of
https://github.com/mii443/tokenizers.git
synced 2025-08-22 16:25:30 +00:00
Remove some useless modifications.
This commit is contained in:
@ -58,4 +58,3 @@ spm_precompiled = "0.1"
|
||||
criterion = "0.3"
|
||||
tempfile = "3.1"
|
||||
assert_approx_eq = "1.1"
|
||||
simple_logger = "1.9"
|
||||
|
@ -489,15 +489,6 @@ impl NormalizedString {
|
||||
}
|
||||
})
|
||||
.collect::<Vec<_>>();
|
||||
debug!(
|
||||
"Filter transform {:?}",
|
||||
filtered
|
||||
.iter()
|
||||
.cloned()
|
||||
.rev()
|
||||
.filter_map(|o| o)
|
||||
.collect::<Vec<_>>()
|
||||
);
|
||||
self.transform(
|
||||
filtered.into_iter().rev().filter_map(|o| o),
|
||||
removed as usize,
|
||||
|
Reference in New Issue
Block a user