jartine commited on
Commit
d600bc1
1 Parent(s): 0cc0c66

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -2
README.md CHANGED
@@ -39,6 +39,8 @@ chmod +x Mistral-Nemo-Instruct-2407.Q6_K.llamafile
39
  ./Mistral-Nemo-Instruct-2407.Q6_K.llamafile
40
  ```
41
 
 
 
42
  This model has a max context window size of 128k tokens. By default, a
43
  context window size of 8192 tokens is used. You may increase this to the
44
  maximum by passing the `-c 0` flag.
@@ -58,8 +60,23 @@ of the README.
58
 
59
  ## Prompting
60
 
61
- Mistral models work well with the default settings of the llamafile
62
- server GUI. You shouldn't need to specify a custom prompt template.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
63
 
64
  Here's an example of how to prompt Mistral on the command line:
65
 
 
39
  ./Mistral-Nemo-Instruct-2407.Q6_K.llamafile
40
  ```
41
 
42
+ You then need to fill out the prompt / history template (see below).
43
+
44
  This model has a max context window size of 128k tokens. By default, a
45
  context window size of 8192 tokens is used. You may increase this to the
46
  maximum by passing the `-c 0` flag.
 
60
 
61
  ## Prompting
62
 
63
+ To have a good working chat experience when using the web GUI, you need
64
+ to fill out the text fields with the following values.
65
+
66
+ Prompt template:
67
+
68
+ ```
69
+ {{prompt}}
70
+
71
+ {{history}}
72
+ {{char}}:
73
+ ```
74
+
75
+ History template:
76
+
77
+ ```
78
+ {{name}}: {{message}}
79
+ ```
80
 
81
  Here's an example of how to prompt Mistral on the command line:
82