|
15 | 15 | import tempfile |
16 | 16 | import unittest |
17 | 17 |
|
18 | | -from transformers import AutoTokenizer |
| 18 | +from transformers import AutoProcessor, AutoTokenizer |
19 | 19 | from transformers.testing_utils import require_jmespath |
20 | 20 | from transformers.utils.chat_parsing_utils import recursive_parse |
21 | 21 |
|
|
150 | 150 | @require_jmespath |
151 | 151 | class ChatSchemaParserTest(unittest.TestCase): |
152 | 152 | def test_schema_save_load(self): |
| 153 | + # Has no schema by default |
153 | 154 | tokenizer = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2") |
154 | 155 | tokenizer.response_schema = ernie_schema |
155 | 156 | with tempfile.TemporaryDirectory() as tmpdir: |
156 | 157 | tokenizer.save_pretrained(tmpdir) |
157 | 158 | reloaded_tokenizer = AutoTokenizer.from_pretrained(tmpdir) |
158 | 159 | self.assertEqual(reloaded_tokenizer.response_schema, ernie_schema) |
159 | 160 |
|
| 161 | + # Has no schema by default |
| 162 | + processor = AutoProcessor.from_pretrained("hf-internal-testing/tiny-random-Qwen2VLForConditionalGeneration") |
| 163 | + processor.response_schema = ernie_schema |
| 164 | + with tempfile.TemporaryDirectory() as tmpdir: |
| 165 | + processor.save_pretrained(tmpdir) |
| 166 | + reloaded_processor = AutoProcessor.from_pretrained(tmpdir) |
| 167 | + self.assertEqual(reloaded_processor.response_schema, ernie_schema) |
| 168 | + |
160 | 169 | def test_tokenizer_method(self): |
161 | 170 | tokenizer = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2") |
162 | 171 | model_out = '<|START_THINKING|>I should call a tool.<|END_THINKING|><|START_ACTION|>[\n {"tool_call_id": "0", "tool_name": "simple_tool", "parameters": {"temperature_format": "Celsius"}}\n]<|END_ACTION|><|END_OF_TURN_TOKEN|>' |
|
0 commit comments