mirror of
https://github.com/mii443/tokenizers.git
synced 2025-08-31 04:29:21 +00:00
fix typos
This commit is contained in:
@ -109,7 +109,7 @@ impl Encoding {
|
||||
impl PyObjectProtocol for Encoding {
|
||||
fn __repr__(&self) -> PyResult<String> {
|
||||
Ok(format!(
|
||||
"Encoding(num_tokens={}, attributs=[ids, type_ids, tokens, offsets, \
|
||||
"Encoding(num_tokens={}, attributes=[ids, type_ids, tokens, offsets, \
|
||||
attention_mask, special_tokens_mask, overflowing, original_str, normalized_str])",
|
||||
self.encoding.get_ids().len()
|
||||
))
|
||||
|
@ -12,7 +12,7 @@ class BaseTokenizer:
|
||||
def __repr__(self):
|
||||
return "Tokenizer(vocabulary_size={}, {})".format(
|
||||
self._tokenizer.get_vocab_size(),
|
||||
', '.join(k + ': ' + str(v) for k, v in self._parameters.items()))
|
||||
', '.join(k + '=' + str(v) for k, v in self._parameters.items()))
|
||||
|
||||
def enable_padding(self,
|
||||
direction: Optional[str] = "right",
|
||||
|
Reference in New Issue
Block a user