mirror of
https://github.com/mii443/tokenizers.git
synced 2025-08-23 00:35:35 +00:00
Python - Adding tokenizers classes - WIP
This commit is contained in:
@ -20,7 +20,8 @@ setup(
|
||||
"tokenizers.normalizers",
|
||||
"tokenizers.pre_tokenizers",
|
||||
"tokenizers.processors",
|
||||
"tokenizers.trainers"
|
||||
"tokenizers.trainers",
|
||||
"tokenizers.implementations",
|
||||
],
|
||||
package_data = {
|
||||
'tokenizers': [ 'py.typed', '__init__.pyi' ],
|
||||
@ -30,6 +31,7 @@ setup(
|
||||
'tokenizers.pre_tokenizers': [ 'py.typed', '__init__.pyi' ],
|
||||
'tokenizers.processors': [ 'py.typed', '__init__.pyi' ],
|
||||
'tokenizers.trainers': [ 'py.typed', '__init__.pyi' ],
|
||||
'tokenizers.implementations': [ 'py.typed' ],
|
||||
},
|
||||
zip_safe=False,
|
||||
)
|
||||
|
Reference in New Issue
Block a user