mirror of
https://github.com/mii443/tokenizers.git
synced 2025-08-23 00:35:35 +00:00
Python - last fixes on Encoding bindings/typings
This commit is contained in:
@ -39,6 +39,7 @@ impl PySequenceProtocol for Encoding {
|
||||
#[pymethods]
|
||||
impl Encoding {
|
||||
#[staticmethod]
|
||||
#[args(growing_offsets = true)]
|
||||
fn merge(encodings: Vec<&Encoding>, growing_offsets: bool) -> Encoding {
|
||||
Encoding::new(tk::tokenizer::Encoding::merge(
|
||||
encodings
|
||||
|
Reference in New Issue
Block a user