mirror of
https://github.com/mii443/tokenizers.git
synced 2025-12-05 20:28:22 +00:00
Python - Also allow creating Tokenizer from_buffer
This commit is contained in:
@@ -253,6 +253,18 @@ impl Tokenizer {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[staticmethod]
|
||||||
|
fn from_buffer(buffer: &PyBytes) -> PyResult<Self> {
|
||||||
|
let tokenizer: tk::tokenizer::Tokenizer = serde_json::from_slice(buffer.as_bytes())
|
||||||
|
.map_err(|e| {
|
||||||
|
exceptions::Exception::py_err(format!(
|
||||||
|
"Cannot instantiate Tokenizer from buffer: {}",
|
||||||
|
e.to_string()
|
||||||
|
))
|
||||||
|
})?;
|
||||||
|
Ok(Self { tokenizer })
|
||||||
|
}
|
||||||
|
|
||||||
#[args(pretty = false)]
|
#[args(pretty = false)]
|
||||||
fn to_str(&self, pretty: bool) -> PyResult<String> {
|
fn to_str(&self, pretty: bool) -> PyResult<String> {
|
||||||
ToPyResult(self.tokenizer.to_string(pretty)).into()
|
ToPyResult(self.tokenizer.to_string(pretty)).into()
|
||||||
|
|||||||
@@ -272,6 +272,18 @@ class Tokenizer:
|
|||||||
path: str:
|
path: str:
|
||||||
Path to a file containing a Tokenizer
|
Path to a file containing a Tokenizer
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Tokenizer
|
||||||
|
"""
|
||||||
|
pass
|
||||||
|
@staticmethod
|
||||||
|
def from_buffer(buffer: bytes) -> Tokenizer:
|
||||||
|
""" Instantiate a new Tokenizer from the given buffer
|
||||||
|
|
||||||
|
Args:
|
||||||
|
buffer: bytes:
|
||||||
|
A buffer used to instantiate a new Tokenizer
|
||||||
|
|
||||||
Returns:
|
Returns:
|
||||||
Tokenizer
|
Tokenizer
|
||||||
"""
|
"""
|
||||||
|
|||||||
Reference in New Issue
Block a user