mirror of
https://github.com/mii443/tokenizers.git
synced 2025-08-22 16:25:30 +00:00
Reduce number of different revisions by 1 (#1329)
This commit is contained in:
@ -7,9 +7,12 @@ extras["testing"] = ["pytest", "requests", "numpy", "datasets", "black==22.3"]
|
||||
extras["docs"] = ["sphinx", "sphinx_rtd_theme", "setuptools_rust"]
|
||||
extras["dev"] = extras["testing"]
|
||||
|
||||
with open("py_src/tokenizers/__init__.py", "r") as f:
|
||||
version = f.readline().split("=")[-1].strip().strip('"')
|
||||
|
||||
setup(
|
||||
name="tokenizers",
|
||||
version="0.13.4.rc2",
|
||||
version=version,
|
||||
description="Fast and Customizable Tokenizers",
|
||||
long_description=open("README.md", "r", encoding="utf-8").read(),
|
||||
long_description_content_type="text/markdown",
|
||||
|
Reference in New Issue
Block a user