mirror of
https://github.com/mii443/tokenizers.git
synced 2025-08-23 00:35:35 +00:00
Merge pull request #333 from huggingface/fix-added-tokens
Python - Fix Added token deserialization
This commit is contained in:
@ -107,6 +107,7 @@ impl AddedToken {
|
|||||||
for (key, value) in state {
|
for (key, value) in state {
|
||||||
let key: &str = key.extract()?;
|
let key: &str = key.extract()?;
|
||||||
match key {
|
match key {
|
||||||
|
"content" => self.content = value.extract()?,
|
||||||
"single_word" => self.single_word = Some(value.extract()?),
|
"single_word" => self.single_word = Some(value.extract()?),
|
||||||
"lstrip" => self.lstrip = Some(value.extract()?),
|
"lstrip" => self.lstrip = Some(value.extract()?),
|
||||||
"rstrip" => self.rstrip = Some(value.extract()?),
|
"rstrip" => self.rstrip = Some(value.extract()?),
|
||||||
|
Reference in New Issue
Block a user