Update chain_setup.py
Browse files- chain_setup.py +17 -3
chain_setup.py
CHANGED
@@ -21,9 +21,9 @@ def load_llm():
|
|
21 |
n_ctx=2048,
|
22 |
n_batch=512,
|
23 |
chat_format="chatml",
|
24 |
-
grammar=
|
25 |
streaming=True,
|
26 |
-
grammar_path=
|
27 |
use_jinja=False,
|
28 |
rope_freq_base=10000.0,
|
29 |
rope_freq_scale=1.0,
|
@@ -33,7 +33,21 @@ def load_llm():
|
|
33 |
repeat_penalty=1.1,
|
34 |
temperature=0.8,
|
35 |
top_k=40,
|
36 |
-
top_p=0.95
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
37 |
)
|
38 |
|
39 |
return llm
|
|
|
21 |
n_ctx=2048,
|
22 |
n_batch=512,
|
23 |
chat_format="chatml",
|
24 |
+
grammar=None,
|
25 |
streaming=True,
|
26 |
+
grammar_path=None, # ensure this file exists and is empty
|
27 |
use_jinja=False,
|
28 |
rope_freq_base=10000.0,
|
29 |
rope_freq_scale=1.0,
|
|
|
33 |
repeat_penalty=1.1,
|
34 |
temperature=0.8,
|
35 |
top_k=40,
|
36 |
+
top_p=0.95,
|
37 |
+
logprobs=None,
|
38 |
+
callback_manager=None,
|
39 |
+
custom_get_token_ids = None,
|
40 |
+
logprobs= None,
|
41 |
+
lora_base = None,
|
42 |
+
lora_path = None,
|
43 |
+
max_tokens = 256,
|
44 |
+
metadata= None,
|
45 |
+
n_gpu_layers= None,
|
46 |
+
n_threads= None,
|
47 |
+
stop=[],
|
48 |
+
suffix= None,
|
49 |
+
tags = None
|
50 |
+
|
51 |
)
|
52 |
|
53 |
return llm
|