mirror of
https://github.com/mii443/tokenizers.git
synced 2025-08-22 16:25:30 +00:00
Update python README
This commit is contained in:
@ -14,7 +14,10 @@ maturin develop --release
|
|||||||
# Then test:
|
# Then test:
|
||||||
pip install transformers
|
pip install transformers
|
||||||
|
|
||||||
python example.py --file <FILE_PATH>
|
# Download vocab/merges from GPT-2
|
||||||
# or
|
wget https://s3.amazonaws.com/models.huggingface.co/bert/gpt2-vocab.json
|
||||||
python example.py
|
wget https://s3.amazonaws.com/models.huggingface.co/bert/gpt2-merges.txt
|
||||||
|
|
||||||
|
python examples/example.py --file <FILE_PATH> --merges gpt2-merges.txt --vocab gpt2-vocab.json
|
||||||
|
python custom_pre_tokenizer.py --merges gpt2-merges.txt --vocab gpt2-vocab.json
|
||||||
```
|
```
|
||||||
|
Reference in New Issue
Block a user