cedpsam commited on
Commit
b4905cb
·
1 Parent(s): bc4f800

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -4
app.py CHANGED
@@ -27,9 +27,9 @@ llm = LlamaCpp(
27
  def format_prompt(message, history):
28
  prompt = "<s>"
29
  for user_prompt, bot_response in history:
30
- prompt += f"[INST] {user_prompt} [/INST]"
31
- prompt += f" {bot_response}</s> "
32
- prompt += f"[INST] {message} [/INST]"
33
  return prompt
34
 
35
  def generate(
@@ -51,7 +51,9 @@ def generate(
51
  temperature=temperature,
52
  max_tokens=max_new_tokens,
53
  repeat_penalty=repetition_penalty,
54
- top_p=top_p,)
 
 
55
  # output=formatted_prompt+"ans:"+output
56
  # for response in stream:
57
  # output += response.token.text
 
27
  def format_prompt(message, history):
28
  prompt = "<s>"
29
  for user_prompt, bot_response in history:
30
+ prompt += f"<|im_start|>user\n {user_prompt} <|im_end|>\n"
31
+ prompt += f"<|im_start|>assistant\n {bot_response}<|im_end|>\n"
32
+ prompt += f"<|im_start|>user\n {message} <|im_end|>\n<|im_start|>assistant\n"
33
  return prompt
34
 
35
  def generate(
 
51
  temperature=temperature,
52
  max_tokens=max_new_tokens,
53
  repeat_penalty=repetition_penalty,
54
+ top_p=top_p,
55
+ stop=["<|im_end|>","<|im_start|>user"]
56
+ )
57
  # output=formatted_prompt+"ans:"+output
58
  # for response in stream:
59
  # output += response.token.text