Handle vocab size with added tokens

This commit is contained in:
Anthony MOI
2019-12-19 20:19:56 -05:00
parent b7040e0412
commit f2b9c30ad9
2 changed files with 8 additions and 4 deletions

View File

@ -36,9 +36,8 @@ impl Tokenizer {
}
}
#[getter]
fn get_vocab_size(&self) -> usize {
self.tokenizer.get_vocab_size()
fn get_vocab_size(&self, with_added_tokens: bool) -> usize {
self.tokenizer.get_vocab_size(with_added_tokens)
}
fn with_model(&mut self, model: &mut Model) -> PyResult<()> {