|
import unittest |
|
import warnings |
|
from dataclasses import dataclass |
|
|
|
from transformers.convert_slow_tokenizer import SpmConverter |
|
from transformers.testing_utils import get_tests_dir |
|
|
|
|
|
@dataclass |
|
class FakeOriginalTokenizer: |
|
vocab_file: str |
|
|
|
|
|
class ConvertSlowTokenizerTest(unittest.TestCase): |
|
def test_spm_converter_bytefallback_warning(self): |
|
spm_model_file_without_bytefallback = get_tests_dir("fixtures/test_sentencepiece.model") |
|
spm_model_file_with_bytefallback = get_tests_dir("fixtures/test_sentencepiece_with_bytefallback.model") |
|
|
|
original_tokenizer_without_bytefallback = FakeOriginalTokenizer(vocab_file=spm_model_file_without_bytefallback) |
|
|
|
with warnings.catch_warnings(record=True) as w: |
|
_ = SpmConverter(original_tokenizer_without_bytefallback) |
|
self.assertEqual(len(w), 0) |
|
|
|
original_tokenizer_with_bytefallback = FakeOriginalTokenizer(vocab_file=spm_model_file_with_bytefallback) |
|
|
|
with warnings.catch_warnings(record=True) as w: |
|
_ = SpmConverter(original_tokenizer_with_bytefallback) |
|
self.assertEqual(len(w), 1) |
|
self.assertIn( |
|
"The sentencepiece tokenizer that you are converting to a fast tokenizer uses the byte fallback option" |
|
" which is not implemented in the fast tokenizers.", |
|
str(w[0].message), |
|
) |
|
|