mirror of
https://github.com/mii443/tokenizers.git
synced 2025-08-23 00:35:35 +00:00
Expand documentation of UnigramTrainer (#770)
* Expand documentation of UnigramTrainer * Put doc at the source * Add signature * make style Co-authored-by: Anthony Moi <m.anthony.moi@gmail.com>
This commit is contained in:
@ -669,8 +669,22 @@ impl PyWordLevelTrainer {
|
||||
/// if not seen in the training dataset.
|
||||
/// If the strings contain more than one character, only the first one
|
||||
/// is kept.
|
||||
///
|
||||
/// shrinking_factor (:obj:`float`):
|
||||
/// The shrinking factor used at each step of the training to prune the
|
||||
/// vocabulary.
|
||||
///
|
||||
/// unk_token (:obj:`str`):
|
||||
/// The token used for out-of-vocabulary tokens.
|
||||
///
|
||||
/// max_piece_length (:obj:`int`):
|
||||
/// The maximum length of a given token.
|
||||
///
|
||||
/// n_sub_iterations (:obj:`int`):
|
||||
/// The number of iterations of the EM algorithm to perform before
|
||||
/// pruning the vocabulary.
|
||||
#[pyclass(extends=PyTrainer, module = "tokenizers.trainers", name=UnigramTrainer)]
|
||||
#[text_signature = "(self, vocab_size=8000, show_progress=True, special_tokens= [])"]
|
||||
#[text_signature = "(self, vocab_size=8000, show_progress=True, special_tokens=[], shrinking_factor=0.75, unk_token=None, max_piece_length=16, n_sub_iterations=2)"]
|
||||
pub struct PyUnigramTrainer {}
|
||||
#[pymethods]
|
||||
impl PyUnigramTrainer {
|
||||
|
Reference in New Issue
Block a user