Is the prompt format correct?

#1
by Olafangensan - opened

From my testing, this model seems to work best with ChatML, while the format in this model card is just... bad.
SillyTavern even automatically detects and uses ChatML.

It's definitely the format that's in the original model template, so that's weird!

Not sure why it would work better with a format other than the one they provide πŸ€”

My bad, llama-server did it for me:

srv    load_model: load_model: Chat template parsing error: this custom template is not supported, try using --jinja
srv    load_model: load_model: The chat template that comes with this model is not yet supported, falling back to chatml. This may cause the model to output suboptimal responses
srv          init: initializing slots, n_slots = 1
slot         init: id  0 | task -1 | new slot n_ctx_slot = 32768
main: model loaded
main: chat template, chat_template: {%- for message in messages -%}
  {{- '<|im_start|>' + message.role + '
' + message.content + '<|im_end|>
' -}}
{%- endfor -%}
{%- if add_generation_prompt -%}
  {{- '<|im_start|>assistant
' -}}
{%- endif -%}, example_format: '<|im_start|>system
You are a helpful assistant<|im_end|>
<|im_start|>user
Hello<|im_end|>
<|im_start|>assistant
Hi there<|im_end|>
<|im_start|>user
How are you?<|im_end|>
<|im_start|>assistant
'
main: server is listening on http://127.0.0.1:8080 - starting the main loop
srv  update_slots: all slots are idle

Just out of curiosity, how could I run this with the correct template?

Edit: Right, all I had to do is add --jinja to the llama-server command. Closing!

Olafangensan changed discussion status to closed

Sign up or log in to comment