apepkuss79 commited on
Commit
279a80e
·
verified ·
1 Parent(s): 02f0044

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -6
README.md CHANGED
@@ -26,14 +26,18 @@ library_name: transformers
26
 
27
  - LlamaEdge version: [v0.16.8](https://github.com/LlamaEdge/LlamaEdge/releases/tag/0.16.8)
28
 
29
- <!-- - Prompt template
30
 
31
- - Prompt type: `deepseek-chat-25`
32
 
33
  - Prompt string
34
 
35
  ```text
36
- <|begin_of_sentence|>{system_message}<|User|>{user_message_1}<|Assistant|>{assistant_message_1}<|end_of_sentence|><|User|>{user_message_2}<|Assistant|>
 
 
 
 
37
  ```
38
 
39
  - Context size: `128000`
@@ -44,7 +48,7 @@ library_name: transformers
44
  wasmedge --dir .:. --nn-preload default:GGML:AUTO:DeepSeek-R1-Distill-Qwen-7B-Q5_K_M.gguf \
45
  llama-api-server.wasm \
46
  --model-name DeepSeek-R1-Distill-Qwen-7B \
47
- --prompt-template deepseek-chat-25 \
48
  --ctx-size 128000
49
  ```
50
 
@@ -53,9 +57,9 @@ library_name: transformers
53
  ```bash
54
  wasmedge --dir .:. --nn-preload default:GGML:AUTO:DeepSeek-R1-Distill-Qwen-7B-Q5_K_M.gguf \
55
  llama-chat.wasm \
56
- --prompt-template deepseek-chat-25 \
57
  --ctx-size 128000
58
- ``` -->
59
 
60
  ## Quantized GGUF Models
61
 
 
26
 
27
  - LlamaEdge version: [v0.16.8](https://github.com/LlamaEdge/LlamaEdge/releases/tag/0.16.8)
28
 
29
+ - Prompt template
30
 
31
+ - Prompt type: `chatml`
32
 
33
  - Prompt string
34
 
35
  ```text
36
+ <|im_start|>system
37
+ {system_message}<|im_end|>
38
+ <|im_start|>user
39
+ {prompt}<|im_end|>
40
+ <|im_start|>assistant
41
  ```
42
 
43
  - Context size: `128000`
 
48
  wasmedge --dir .:. --nn-preload default:GGML:AUTO:DeepSeek-R1-Distill-Qwen-7B-Q5_K_M.gguf \
49
  llama-api-server.wasm \
50
  --model-name DeepSeek-R1-Distill-Qwen-7B \
51
+ --prompt-template chatml \
52
  --ctx-size 128000
53
  ```
54
 
 
57
  ```bash
58
  wasmedge --dir .:. --nn-preload default:GGML:AUTO:DeepSeek-R1-Distill-Qwen-7B-Q5_K_M.gguf \
59
  llama-chat.wasm \
60
+ --prompt-template chatml \
61
  --ctx-size 128000
62
+ ```
63
 
64
  ## Quantized GGUF Models
65