mirror of
https://github.com/mii443/tokenizers.git
synced 2025-08-23 00:35:35 +00:00
Remove some useless modifications.
This commit is contained in:
@ -58,4 +58,3 @@ spm_precompiled = "0.1"
|
|||||||
criterion = "0.3"
|
criterion = "0.3"
|
||||||
tempfile = "3.1"
|
tempfile = "3.1"
|
||||||
assert_approx_eq = "1.1"
|
assert_approx_eq = "1.1"
|
||||||
simple_logger = "1.9"
|
|
||||||
|
@ -489,15 +489,6 @@ impl NormalizedString {
|
|||||||
}
|
}
|
||||||
})
|
})
|
||||||
.collect::<Vec<_>>();
|
.collect::<Vec<_>>();
|
||||||
debug!(
|
|
||||||
"Filter transform {:?}",
|
|
||||||
filtered
|
|
||||||
.iter()
|
|
||||||
.cloned()
|
|
||||||
.rev()
|
|
||||||
.filter_map(|o| o)
|
|
||||||
.collect::<Vec<_>>()
|
|
||||||
);
|
|
||||||
self.transform(
|
self.transform(
|
||||||
filtered.into_iter().rev().filter_map(|o| o),
|
filtered.into_iter().rev().filter_map(|o| o),
|
||||||
removed as usize,
|
removed as usize,
|
||||||
|
Reference in New Issue
Block a user