mirror of
https://github.com/mii443/tokenizers.git
synced 2025-08-22 16:25:30 +00:00
Merge pull request #93 from Tomarchelone/patch-1
Fix indexing bug in add_tokens()
This commit is contained in:
@ -587,7 +587,7 @@ impl Tokenizer {
|
||||
continue;
|
||||
}
|
||||
|
||||
let new_id = (self.model.get_vocab_size() - 1 + self.added_tokens.len()) as u32;
|
||||
let new_id = (self.model.get_vocab_size() + self.added_tokens.len()) as u32;
|
||||
let id = self
|
||||
.added_tokens
|
||||
.entry(token.clone())
|
||||
|
Reference in New Issue
Block a user