mirror of
https://github.com/mii443/tokenizers.git
synced 2025-08-22 16:25:30 +00:00
* current updates * simplify * set_item works, but `tokenizer._tokenizer.post_processor[1].single = ["$0", "</s>"]` does not ! * fix: `normalizers` deserialization and other refactoring * fix: `pre_tokenizer` deserialization * feat: add `__len__` implementation for `normalizer::PySequence` * feat: add `__setitem__` impl for `normalizers::PySequence` * feat: add `__setitem__` impl to `pre_tokenizer::PySequence` * feat: add `__setitem__` impl to `post_processor::PySequence` * test: add normalizer sequence setter check * refactor: allow unused `processors::setter` macro * test: add `__setitem__` test for processors & pretok * refactor: `unwrap` -> `PyException::new_err()?` * refactor: fmt * refactor: remove unnecessary `pub` * feat(bindings): add missing getters & setters for pretoks * feat(bindings): add missing getters & setters for processors * refactor(bindings): rewrite RwLock poison error msg * refactor: remove debug print * feat(bindings): add description as to why custom deser is needed * feat: make post proc sequence elements mutable * fix(binding): serialization --------- Co-authored-by: Luc Georges <luc.sydney.georges@gmail.com>
69 lines
1.9 KiB
TOML
69 lines
1.9 KiB
TOML
[project]
|
|
name = 'tokenizers'
|
|
requires-python = '>=3.9'
|
|
authors = [
|
|
{ name = 'Nicolas Patry', email = 'patry.nicolas@protonmail.com' },
|
|
{ name = 'Anthony Moi', email = 'anthony@huggingface.co' },
|
|
]
|
|
classifiers = [
|
|
"Development Status :: 5 - Production/Stable",
|
|
"Intended Audience :: Developers",
|
|
"Intended Audience :: Education",
|
|
"Intended Audience :: Science/Research",
|
|
"License :: OSI Approved :: Apache Software License",
|
|
"Operating System :: OS Independent",
|
|
"Programming Language :: Python :: 3",
|
|
"Programming Language :: Python :: 3.7",
|
|
"Programming Language :: Python :: 3.8",
|
|
"Programming Language :: Python :: 3.9",
|
|
"Programming Language :: Python :: 3.10",
|
|
"Programming Language :: Python :: 3.11",
|
|
"Topic :: Scientific/Engineering :: Artificial Intelligence",
|
|
]
|
|
keywords = ["NLP", "tokenizer", "BPE", "transformer", "deep learning"]
|
|
dynamic = ['description', 'license', 'readme', 'version']
|
|
dependencies = ["huggingface_hub>=0.16.4,<1.0"]
|
|
|
|
[project.urls]
|
|
Homepage = 'https://github.com/huggingface/tokenizers'
|
|
Source = 'https://github.com/huggingface/tokenizers'
|
|
|
|
|
|
[project.optional-dependencies]
|
|
testing = ["pytest", "requests", "numpy", "datasets", "black==22.3", "ruff"]
|
|
docs = ["sphinx", "sphinx_rtd_theme", "setuptools_rust"]
|
|
dev = ["tokenizers[testing]"]
|
|
|
|
|
|
[build-system]
|
|
requires = ["maturin>=1.0,<2.0"]
|
|
build-backend = "maturin"
|
|
|
|
[tool.maturin]
|
|
python-source = "py_src"
|
|
module-name = "tokenizers.tokenizers"
|
|
bindings = 'pyo3'
|
|
features = ["pyo3/extension-module"]
|
|
|
|
[tool.black]
|
|
line-length = 119
|
|
target-version = ['py35']
|
|
|
|
[tool.ruff]
|
|
line-length = 119
|
|
target-version = "py311"
|
|
lint.ignore = [
|
|
# a == None in tests vs is None.
|
|
"E711",
|
|
# a == False in tests vs is False.
|
|
"E712",
|
|
# try.. import except.. pattern without using the lib.
|
|
"F401",
|
|
# Raw type equality is required in asserts
|
|
"E721",
|
|
# Import order
|
|
"E402",
|
|
# Fixtures unused import
|
|
"F811",
|
|
]
|