Help with workaround for issue 581, ChatML chat format causing assertion error #585
Unanswered
chris-cortner
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
My project is currently blocked on this. I haven't found a way to use this prompt format. Can someone help me with a workaround in the short term?
Thanks!
Issue #581
Copying the text here:
I'm trying to apply dolphin mistral's prompt template format:
<|im_start|>system
{system_prompt}<|im_end|>
<|im_start|>user
{user_prompt}<|im_end|>
<|im_start|>assistant
I've tried this a couple of different ways:
quant_path = "TheBloke/dolphin-2.6-mistral-7B-AWQ"
lm = models.Transformers(quant_path, device_map="auto")
stop_char = '"'
prompt_template = '<|im_start|>system\n{system_prompt}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant\n'
lm2 = lm + (prompt_template.format(system_prompt="You are a helpful AI", prompt="What is the distance to mars?")
And by using TransformersChat:
quant_path = "TheBloke/dolphin-2.6-mistral-7B-AWQ"
lm = models.TransformersChat(quant_path, device_map="auto")
stop_char = '"'
with system():
lm2 = lm + "You are a helpful AI"
with user():
lm2 += "What is the distance to mars?"
with assistant():
lm2 += 'The distance to mars is "' + gen("answer", max_tokens=500, stop=stop_char, temperature=0.8)
Both method produce the same error:
An assertion error is thrown in _cleanup_tokens in _model.py
Traceback (most recent call last):
File "/home/user/.cache/pypoetry/virtualenvs/llm-proficiency-testing-hKJXaDzo-py3.11/lib64/python3.11/site-packages/guidance/models/_model.py", line 309, in add
out = lm + partial_grammar
Beta Was this translation helpful? Give feedback.
All reactions