mirror of
https://github.com/mii443/tokenizers.git
synced 2025-08-22 16:25:30 +00:00
Merge pull request #333 from huggingface/fix-added-tokens
Python - Fix Added token deserialization
This commit is contained in:
@ -107,6 +107,7 @@ impl AddedToken {
|
||||
for (key, value) in state {
|
||||
let key: &str = key.extract()?;
|
||||
match key {
|
||||
"content" => self.content = value.extract()?,
|
||||
"single_word" => self.single_word = Some(value.extract()?),
|
||||
"lstrip" => self.lstrip = Some(value.extract()?),
|
||||
"rstrip" => self.rstrip = Some(value.extract()?),
|
||||
|
Reference in New Issue
Block a user