mirror of
https://github.com/mii443/tokenizers.git
synced 2025-08-23 00:35:35 +00:00
Word indices are None for special tokens
This commit is contained in:
@ -62,7 +62,7 @@ impl Encoding {
|
||||
}
|
||||
|
||||
#[getter]
|
||||
fn get_words(&self) -> Vec<u32> {
|
||||
fn get_words(&self) -> Vec<Option<u32>> {
|
||||
self.encoding.get_words().to_vec()
|
||||
}
|
||||
|
||||
|
Reference in New Issue
Block a user