From 91f602f744f7fee72f2f5fa8d6c7c48bb1d72d3b Mon Sep 17 00:00:00 2001 From: Anthony MOI Date: Thu, 15 Oct 2020 10:14:58 -0400 Subject: [PATCH] Python - Update CHANGELOG and bump to 0.9.2 for release --- bindings/python/CHANGELOG.md | 6 ++++++ bindings/python/Cargo.lock | 2 +- bindings/python/Cargo.toml | 2 +- bindings/python/py_src/tokenizers/__init__.py | 2 +- bindings/python/setup.py | 2 +- 5 files changed, 10 insertions(+), 4 deletions(-) diff --git a/bindings/python/CHANGELOG.md b/bindings/python/CHANGELOG.md index 13d1e837..d22dab62 100644 --- a/bindings/python/CHANGELOG.md +++ b/bindings/python/CHANGELOG.md @@ -4,6 +4,11 @@ All notable changes to this project will be documented in this file. The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.0.0/), and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.html). +## [0.9.2] + +### Fixed +- [#464] Fix a problem with RobertaProcessing being deserialized as BertProcessing + ## [0.9.1] ### Fixed @@ -243,6 +248,7 @@ delimiter (Works like `.split(delimiter)`) - Fix a bug with the IDs associated with added tokens. - Fix a bug that was causing crashes in Python 3.5 +[#464]: https://github.com/huggingface/tokenizers/pull/464 [#459]: https://github.com/huggingface/tokenizers/pull/459 [#420]: https://github.com/huggingface/tokenizers/pull/420 [#417]: https://github.com/huggingface/tokenizers/pull/417 diff --git a/bindings/python/Cargo.lock b/bindings/python/Cargo.lock index da1ea496..e09b8d97 100644 --- a/bindings/python/Cargo.lock +++ b/bindings/python/Cargo.lock @@ -893,7 +893,7 @@ dependencies = [ [[package]] name = "tokenizers-python" -version = "0.9.1" +version = "0.9.2" dependencies = [ "env_logger 0.7.1 (registry+https://github.com/rust-lang/crates.io-index)", "libc 0.2.79 (registry+https://github.com/rust-lang/crates.io-index)", diff --git a/bindings/python/Cargo.toml b/bindings/python/Cargo.toml index 4a38e421..5acae788 100644 --- a/bindings/python/Cargo.toml +++ b/bindings/python/Cargo.toml @@ -1,6 +1,6 @@ [package] name = "tokenizers-python" -version = "0.9.1" +version = "0.9.2" authors = ["Anthony MOI "] edition = "2018" diff --git a/bindings/python/py_src/tokenizers/__init__.py b/bindings/python/py_src/tokenizers/__init__.py index b4e36b8c..6f972a6b 100644 --- a/bindings/python/py_src/tokenizers/__init__.py +++ b/bindings/python/py_src/tokenizers/__init__.py @@ -1,4 +1,4 @@ -__version__ = "0.9.1" +__version__ = "0.9.2" from typing import Tuple, Union, Tuple, List from enum import Enum diff --git a/bindings/python/setup.py b/bindings/python/setup.py index 91ae55c3..d424412e 100644 --- a/bindings/python/setup.py +++ b/bindings/python/setup.py @@ -6,7 +6,7 @@ extras["testing"] = ["pytest"] setup( name="tokenizers", - version="0.9.1", + version="0.9.2", description="Fast and Customizable Tokenizers", long_description=open("README.md", "r", encoding="utf-8").read(), long_description_content_type="text/markdown",