Great! one thing.. FIX for: going on after responses with Q&A pairs ie: 'USER:" ...
add --> stop=["USER:"]
in your params
GREAT WORK! THANKS!
add --> stop=["USER:"]
in your params
GREAT WORK! THANKS!
Does it have trouble keeping it's turns seperated
add --> stop=["USER:"]
in your params
GREAT WORK! THANKS!Does it have trouble keeping it's turns seperated
I'm 90% sure that stuff is handled by the frontend one uses when running the AI, not by the model
most likely cause because I am running this llm with the 'hammer down'
max_tokens=32768
no / max_new_tokens
only seems to occur on short responses to to simple questions.. ie "hello", "who are you" ..
ask something specific and its very good.
cant wait to see the the dataset on this thing - Just fantastic!
most likely cause because I am running this llm with the 'hammer down'
max_tokens=32768
no / max_new_tokens
only seems to occur on short responses to to simple questions.. ie "hello", "who are you" ..
ask something specific and its very good.
cant wait to see the the dataset on this thing - Just fantastic!
fyi the dataset is released :) https://huggingface.co/datasets/teknium/trismegistus-project