sablab commited on
Commit
b9c3147
·
verified ·
1 Parent(s): 7c54c70

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -8
app.py CHANGED
@@ -7,12 +7,12 @@ client = InferenceClient(
7
 
8
 
9
  def format_prompt(message, history):
10
- prompt = "<s>"
11
- for user_prompt, bot_response in history:
12
- prompt += f"[INST] {user_prompt} [/INST]"
13
- prompt += f" {bot_response}</s> "
14
- prompt += f"[INST] {message} [/INST]"
15
- return prompt
16
 
17
  def generate(
18
  prompt, history, temperature=0.9, max_new_tokens=16000, top_p=0.95, repetition_penalty=1.0,):
@@ -24,7 +24,6 @@ def generate(
24
  do_sample=True,
25
  seed=42,)
26
  formatted_prompt = format_prompt(prompt, history)
27
-
28
  stream = client.text_generation(formatted_prompt, **generate_kwargs, stream=True, details=True, return_full_text=False)
29
  output = ""
30
 
@@ -34,7 +33,7 @@ def generate(
34
  return output
35
 
36
  css = """
37
- #mkd {
38
  height: 500px;
39
  overflow: auto;
40
  border: 1px solid #ccc;
 
7
 
8
 
9
  def format_prompt(message, history):
10
+ prompt = "<s>"
11
+ for user_prompt, bot_response in history:
12
+ prompt += f"[INST] {user_prompt} [/INST]"
13
+ prompt += f" {bot_response}</s> "
14
+ prompt += f"[INST] {message} [/INST]"
15
+ return prompt
16
 
17
  def generate(
18
  prompt, history, temperature=0.9, max_new_tokens=16000, top_p=0.95, repetition_penalty=1.0,):
 
24
  do_sample=True,
25
  seed=42,)
26
  formatted_prompt = format_prompt(prompt, history)
 
27
  stream = client.text_generation(formatted_prompt, **generate_kwargs, stream=True, details=True, return_full_text=False)
28
  output = ""
29
 
 
33
  return output
34
 
35
  css = """
36
+ #mkd {
37
  height: 500px;
38
  overflow: auto;
39
  border: 1px solid #ccc;