mirror of
https://github.com/mii443/tokenizers.git
synced 2025-08-22 16:25:30 +00:00
Legacy test.
This commit is contained in:
@ -252,7 +252,11 @@ mod tests {
|
||||
r#"{"type":"BPE","dropout":null,"unk_token":"<unk>","continuing_subword_prefix":null,"end_of_word_suffix":null,"fuse_unk":false,"byte_fallback":false,"ignore_merges":true,"vocab":{"<unk>":0,"a":1,"b":2,"ab":3},"merges":["a b"]}"#
|
||||
);
|
||||
let reconstructed = serde_json::from_str(&data).unwrap();
|
||||
assert_eq!(model, reconstructed);
|
||||
|
||||
// Legacy check, type is not necessary.
|
||||
let legacy = r#"{"dropout":null,"unk_token":"<unk>","continuing_subword_prefix":null,"end_of_word_suffix":null,"fuse_unk":false,"byte_fallback":false,"ignore_merges":true,"vocab":{"<unk>":0,"a":1,"b":2,"ab":3},"merges":["a b"]}"#;
|
||||
let reconstructed = serde_json::from_str(&legacy).unwrap();
|
||||
assert_eq!(model, reconstructed);
|
||||
|
||||
let invalid = r#"{"type":"BPE","dropout":null,"unk_token":"<unk>","continuing_subword_prefix":null,"end_of_word_suffix":null,"fuse_unk":false,"byte_fallback":false,"ignore_merges":true,"vocab":{"<unk>":0,"a":1,"b":2,"ab":3},"merges":["a b c"]}"#;
|
||||
|
Reference in New Issue
Block a user