|
15 | 15 | import tempfile
|
16 | 16 | import unittest
|
17 | 17 |
|
18 |
| -from transformers import AutoTokenizer |
| 18 | +from transformers import AutoProcessor, AutoTokenizer |
19 | 19 | from transformers.testing_utils import require_jmespath
|
20 | 20 | from transformers.utils.chat_parsing_utils import recursive_parse
|
21 | 21 |
|
|
150 | 150 | @require_jmespath
|
151 | 151 | class ChatSchemaParserTest(unittest.TestCase):
|
152 | 152 | def test_schema_save_load(self):
|
| 153 | + # Has no schema by default |
153 | 154 | tokenizer = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2")
|
154 | 155 | tokenizer.response_schema = ernie_schema
|
155 | 156 | with tempfile.TemporaryDirectory() as tmpdir:
|
156 | 157 | tokenizer.save_pretrained(tmpdir)
|
157 | 158 | reloaded_tokenizer = AutoTokenizer.from_pretrained(tmpdir)
|
158 | 159 | self.assertEqual(reloaded_tokenizer.response_schema, ernie_schema)
|
159 | 160 |
|
| 161 | + # Has no schema by default |
| 162 | + processor = AutoProcessor.from_pretrained("hf-internal-testing/tiny-random-Qwen2VLForConditionalGeneration") |
| 163 | + processor.response_schema = ernie_schema |
| 164 | + with tempfile.TemporaryDirectory() as tmpdir: |
| 165 | + processor.save_pretrained(tmpdir) |
| 166 | + reloaded_processor = AutoProcessor.from_pretrained(tmpdir) |
| 167 | + self.assertEqual(reloaded_processor.response_schema, ernie_schema) |
| 168 | + |
160 | 169 | def test_tokenizer_method(self):
|
161 | 170 | tokenizer = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2")
|
162 | 171 | model_out = '<|START_THINKING|>I should call a tool.<|END_THINKING|><|START_ACTION|>[\n {"tool_call_id": "0", "tool_name": "simple_tool", "parameters": {"temperature_format": "Celsius"}}\n]<|END_ACTION|><|END_OF_TURN_TOKEN|>'
|
|
0 commit comments