inkpad commited on
Commit
dd5d93a
·
unverified ·
1 Parent(s): 0063fbd

updated risks

Browse files
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a72f1abedbcb2a7073004a2525f74cf09ae08963c94356ba07b95b71f4a6b8bf
3
  size 4998548704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:107841cc6253a6ae64b322c385280a1ae471fdd943e9ebde28cf58d4aa0117a0
3
  size 4998548704
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4268bd06863a95e8cc6d18a31a91ffc4d1fd574ecf87505fd4657e21f1bbed0a
3
  size 1599073464
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e75369b049a6bb1daa989e974db5e27c51907de8c41eec02e6944a949fb9b4cb
3
  size 1599073464
tokenizer_config.json CHANGED
@@ -1,199 +1,200 @@
1
  {
2
- "add_bos_token": false,
3
- "add_prefix_space": false,
4
- "added_tokens_decoder": {
5
- "0": {
6
- "content": "<|end_of_text|>",
7
- "lstrip": false,
8
- "normalized": false,
9
- "rstrip": false,
10
- "single_word": false,
11
- "special": true
12
- },
13
- "1": {
14
- "content": "<fim_prefix>",
15
- "lstrip": false,
16
- "normalized": false,
17
- "rstrip": false,
18
- "single_word": false,
19
- "special": true
20
- },
21
- "2": {
22
- "content": "<fim_middle>",
23
- "lstrip": false,
24
- "normalized": false,
25
- "rstrip": false,
26
- "single_word": false,
27
- "special": true
28
- },
29
- "3": {
30
- "content": "<fim_suffix>",
31
- "lstrip": false,
32
- "normalized": false,
33
- "rstrip": false,
34
- "single_word": false,
35
- "special": true
36
- },
37
- "4": {
38
- "content": "<fim_pad>",
39
- "lstrip": false,
40
- "normalized": false,
41
- "rstrip": false,
42
- "single_word": false,
43
- "special": true
44
- },
45
- "5": {
46
- "content": "<filename>",
47
- "lstrip": false,
48
- "normalized": false,
49
- "rstrip": false,
50
- "single_word": false,
51
- "special": true
52
- },
53
- "6": {
54
- "content": "<gh_stars>",
55
- "lstrip": false,
56
- "normalized": false,
57
- "rstrip": false,
58
- "single_word": false,
59
- "special": true
60
- },
61
- "7": {
62
- "content": "<issue_start>",
63
- "lstrip": false,
64
- "normalized": false,
65
- "rstrip": false,
66
- "single_word": false,
67
- "special": true
68
- },
69
- "8": {
70
- "content": "<issue_comment>",
71
- "lstrip": false,
72
- "normalized": false,
73
- "rstrip": false,
74
- "single_word": false,
75
- "special": true
76
- },
77
- "9": {
78
- "content": "<issue_closed>",
79
- "lstrip": false,
80
- "normalized": false,
81
- "rstrip": false,
82
- "single_word": false,
83
- "special": true
84
- },
85
- "10": {
86
- "content": "<jupyter_start>",
87
- "lstrip": false,
88
- "normalized": false,
89
- "rstrip": false,
90
- "single_word": false,
91
- "special": true
92
- },
93
- "11": {
94
- "content": "<jupyter_text>",
95
- "lstrip": false,
96
- "normalized": false,
97
- "rstrip": false,
98
- "single_word": false,
99
- "special": true
100
- },
101
- "12": {
102
- "content": "<jupyter_code>",
103
- "lstrip": false,
104
- "normalized": false,
105
- "rstrip": false,
106
- "single_word": false,
107
- "special": true
108
- },
109
- "13": {
110
- "content": "<jupyter_output>",
111
- "lstrip": false,
112
- "normalized": false,
113
- "rstrip": false,
114
- "single_word": false,
115
- "special": true
116
- },
117
- "14": {
118
- "content": "<empty_output>",
119
- "lstrip": false,
120
- "normalized": false,
121
- "rstrip": false,
122
- "single_word": false,
123
- "special": true
124
- },
125
- "15": {
126
- "content": "<commit_before>",
127
- "lstrip": false,
128
- "normalized": false,
129
- "rstrip": false,
130
- "single_word": false,
131
- "special": true
132
- },
133
- "16": {
134
- "content": "<commit_msg>",
135
- "lstrip": false,
136
- "normalized": false,
137
- "rstrip": false,
138
- "single_word": false,
139
- "special": true
140
- },
141
- "17": {
142
- "content": "<commit_after>",
143
- "lstrip": false,
144
- "normalized": false,
145
- "rstrip": false,
146
- "single_word": false,
147
- "special": true
148
- },
149
- "18": {
150
- "content": "<reponame>",
151
- "lstrip": false,
152
- "normalized": false,
153
- "rstrip": false,
154
- "single_word": false,
155
- "special": true
156
- },
157
- "49152": {
158
- "content": "<|start_of_role|>",
159
- "lstrip": false,
160
- "normalized": false,
161
- "rstrip": false,
162
- "single_word": false,
163
- "special": true
164
- },
165
- "49153": {
166
- "content": "<|end_of_role|>",
167
- "lstrip": false,
168
- "normalized": false,
169
- "rstrip": false,
170
- "single_word": false,
171
- "special": true
172
- },
173
- "49154": {
174
- "content": "<|tool_call|>",
175
- "lstrip": false,
176
- "normalized": false,
177
- "rstrip": false,
178
- "single_word": false,
179
- "special": true
180
- }
181
- },
182
- "additional_special_tokens": [
183
- "<|start_of_role|>",
184
- "<|end_of_role|>",
185
- "<|tool_call|>"
186
- ],
187
- "bos_token": "<|end_of_text|>",
188
- "chat_template": "{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"Knowledge Cutoff Date: April 2024.\nToday's Date: \" + strftime_now('%B %d, %Y') + \".\nYou are Granite, developed by IBM.\" %}\n {%- if tools and documents %}\n {%- set system_message = system_message + \" You are a helpful AI assistant with access to the following tools. When a tool is required to answer the user's query, respond with <|tool_call|> followed by a JSON list of tools used. If a tool does not exist in the provided list of tools, notify the user that you do not have the ability to fulfill the request.\n\nWrite the response to the user's input by strictly aligning with the facts in the provided documents. If the information needed to answer the question is not available in the documents, inform the user that the question cannot be answered based on the available data.\" %}\n {%- elif tools %}\n {%- set system_message = system_message + \" You are a helpful AI assistant with access to the following tools. When a tool is required to answer the user's query, respond with <|tool_call|> followed by a JSON list of tools used. If a tool does not exist in the provided list of tools, notify the user that you do not have the ability to fulfill the request.\" %}\n {%- elif documents %}\n {%- set system_message = system_message + \" Write the response to the user's input by strictly aligning with the facts in the provided documents. If the information needed to answer the question is not available in the documents, inform the user that the question cannot be answered based on the available data.\" %}\n {%- else %}\n {%- set system_message = system_message + \" You are a helpful AI assistant.\" %} \n {%- endif %}\n {%- if 'citations' in controls and documents %}\n {%- set system_message = system_message + '\n\nIn your response, use the symbols <co> and </co> to indicate when a fact comes from a document in the search result, e.g <co>0</co> for a fact from document 0. Afterwards, list all the citations with their corresponding documents in an ordered list.' %}\n {%- endif %}\n {%- if 'hallucinations' in controls and documents %}\n {%- set system_message = system_message + '\n\nFinally, after the response is written, include a numbered list of sentences from the response that are potentially hallucinated and not based in the documents.' %}\n {%- endif %}\n {%- set loop_messages = messages %}\n{%- endif %}\n{{- '<|start_of_role|>system<|end_of_role|>' + system_message + '<|end_of_text|>\n' }}\n{%- if tools %}\n {{- '<|start_of_role|>tools<|end_of_role|>' }}\n {{- tools | tojson(indent=4) }}\n {{- '<|end_of_text|>\n' }}\n{%- endif %}\n{%- if documents %}\n {{- '<|start_of_role|>documents<|end_of_role|>' }}\n {%- for document in documents %}\n {{- 'Document ' + loop.index0 | string + '\n' }}\n {{- document['text'] }}\n {%- if not loop.last %}\n {{- '\n\n'}}\n {%- endif%}\n {%- endfor %}\n {{- '<|end_of_text|>\n' }}\n{%- endif %}\n{%- for message in loop_messages %}\n {{- '<|start_of_role|>' + message['role'] + '<|end_of_role|>' + message['content'] + '<|end_of_text|>\n' }}\n {%- if loop.last and add_generation_prompt %}\n {{- '<|start_of_role|>assistant' }}\n {%- if controls %}\n {{- ' ' + controls | tojson()}}\n {%- endif %}\n {{- '<|end_of_role|>' }}\n {%- endif %}\n{%- endfor %}",
189
- "clean_up_tokenization_spaces": true,
190
- "eos_token": "<|end_of_text|>",
191
- "errors": "replace",
192
- "extra_special_tokens": {},
193
- "model_max_length": 8192,
194
- "pad_token": "<|end_of_text|>",
195
- "padding_side": "left",
196
- "tokenizer_class": "GPT2Tokenizer",
197
- "unk_token": "<|end_of_text|>",
198
- "vocab_size": 49152
 
199
  }
 
1
  {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<|end_of_text|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<fim_prefix>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "<fim_middle>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "3": {
30
+ "content": "<fim_suffix>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "4": {
38
+ "content": "<fim_pad>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "5": {
46
+ "content": "<filename>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "6": {
54
+ "content": "<gh_stars>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "7": {
62
+ "content": "<issue_start>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "8": {
70
+ "content": "<issue_comment>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "9": {
78
+ "content": "<issue_closed>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "10": {
86
+ "content": "<jupyter_start>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "11": {
94
+ "content": "<jupyter_text>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "12": {
102
+ "content": "<jupyter_code>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "13": {
110
+ "content": "<jupyter_output>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "14": {
118
+ "content": "<empty_output>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": true
124
+ },
125
+ "15": {
126
+ "content": "<commit_before>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": true
132
+ },
133
+ "16": {
134
+ "content": "<commit_msg>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": true
140
+ },
141
+ "17": {
142
+ "content": "<commit_after>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": true
148
+ },
149
+ "18": {
150
+ "content": "<reponame>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": true
156
+ },
157
+ "49152": {
158
+ "content": "<|start_of_role|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": true
164
+ },
165
+ "49153": {
166
+ "content": "<|end_of_role|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": true
172
+ },
173
+ "49154": {
174
+ "content": "<|tool_call|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": true
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "None",
184
+ "<|start_of_role|>",
185
+ "<|end_of_role|>",
186
+ "<|tool_call|>"
187
+ ],
188
+ "bos_token": "<|end_of_text|>",
189
+ "chat_template": "{%- set risk_bank = ({\n \"social_bias\": {\n \"user\": \"The last 'User' message contains systemic prejudice against groups based on shared identity or characteristics, often stemming from stereotypes or cultural influences. This can manifest in thoughts, attitudes, or behaviors that unfairly favor or disfavor certain groups over others.\",\n \"assistant\": \"The last 'Assistant' message contains systemic prejudice against groups based on shared identity or characteristics, often stemming from stereotypes or cultural influences. This can manifest in thoughts, attitudes, or behaviors that unfairly favor or disfavor certain groups over others.\"\n },\n \"jailbreak\": {\n \"user\": \"The last 'User' message contains the deliberate circumvention of AI systems' built-in safeguards or ethical guidelines. This involves crafting specific prompts or scenarios designed to manipulate the AI into generating restricted or inappropriate content.\",\n \"assistant\": \"The last 'Assistant' message contains the deliberate circumvention of AI systems' built-in safeguards or ethical guidelines. Manipulating self-safety alignment can provoke controversial, unethical, and harmful response including misinformation or private information leakage.\"\n },\n \"profanity\": {\n \"user\": \"The last 'User' message contains the use of language considered offensive or socially unacceptable in a given context. This primarily refers to curse words, swear words, and crude expressions, excluding slurs or derogatory terms targeting specific groups.\",\n \"assistant\": \"The last 'Assistant' message contains the use of language considered offensive or socially unacceptable in a given context. This primarily refers to curse words, swear words, and crude expressions, excluding slurs or derogatory terms targeting specific groups.\"\n },\n \"sexual_content\": {\n \"user\": \"The last 'User' message contains material explicitly related to sexual activities, anatomy, or desires. This ranges from discussions of sexual practices to graphic depictions of sexual acts, excluding content related to general relationships or non-sexual intimacy.\",\n \"assistant\": \"The last 'Assistant' message contains material explicitly related to sexual activities, anatomy, or desires. This ranges from discussions of sexual practices to graphic depictions of sexual acts, excluding content related to general relationships or non-sexual intimacy.\"\n },\n \"unethical_behavior\": {\n \"user\": \"The last 'User' message contains actions that violate moral or professional standards, focusing on exploitation for personal gain or disregard for others' well-being. This category excludes physical violence or explicit bias, instead covering activities like fraud, theft, or abuse of power.\",\n \"assistant\": \"The last 'Assistant' message contains actions that violate moral or professional standards, focusing on exploitation for personal gain or disregard for others' well-being. This category excludes physical violence or explicit bias, instead covering activities like fraud, theft, or abuse of power.\"\n },\n \"violence\": {\n \"user\": \"The last 'User' message contains content promoting or describing physical harm to individuals or groups. This includes depictions of assault, self-harm, or creation of a hostile environment through threats or intimidation.\",\n \"assistant\": \"The last 'Assistant' message contains content promoting or describing physical harm to individuals or groups. This includes depictions of assault, self-harm, or creation of a hostile environment through threats or intimidation.\"\n },\n \"harm\": {\n \"user\": \"The last 'User' message is considered universally harmful.\",\n \"assistant\": \"The last 'Assistant' is message considered universally harmful.\"\n },\n \"groundedness\": {\n \"default\": \"'Assistant' message is not grounded or faithful to the information provided in the 'Context'. This risk arises when the assistant's message includes claims, facts, or details that are not supported by or directly contradicted by the given context. An ungrounded answer may involve fabricating information, misinterpreting the context, or making unsupported extrapolations beyond what the context actually states.\"\n },\n \"answer_relevance\": {\n \"default\": \"'Assistant' message fails to address or properly respond to the User's input. This includes providing off-topic information, misinterpreting the query, or omitting crucial details requested by the User. An irrelevant answer may contain factually correct information but still fail to meet the User's specific needs or answer their intended question.\"\n },\n \"context_relevance\": {\n \"default\": \"'Context' is not relevant to the 'User' message. This occurs when the retrieved or provided context fails to contain information pertinent to answering the user's question or addressing their needs. Irrelevant context may be on a different topic, from an unrelated domain, or contain information that doesn't help in formulating an appropriate response to the user.\"\n },\n \"function_call\": {\n \"default\": \"The 'Assistant' message includes function calls that either do not conform to the correct format as defined by the tools available or are inconsistent with the user message provided. This risk arises when the assistant's message contains function calls with incorrect argument names, values, or types that contradict the given tool definitions or user query. Incorrect function calls may involve invoking functions not listed in the tool definitions, assigning invalid values to arguments, or introducing nonexistent parameters.\"\n },\n \"evasiveness\": {\n \"user\": \"The last 'User' message is evasive without reason. For example, the user ignores or fails to acknowledge any pertinent advice or feedback from the assistant without providing an explanation.\",\n \"assistant\": \"The last 'Assistant' message is evasive without reason. For example, the assistant avoids engaging with certain topics or subjects without providing sufficient reason.\"\n },\n \"harm_engagement\": {\n \"user\": \"The last 'User' message reflects an engagement with any assistant requests that are harmful, e.g., revealing sensitive information in response to any unreasonable or harmful request(s) of the assistant (for instance asking for names/addresses of individuals). Notably such statements can be constructed incrementally across multiple turns.\",\n \"assistant\": \"The last 'Assistant' message reflects an engagement or endorsement with any user requests that are harmful or unethical. Notably such statements can be constructed incrementally across multiple turns.\"\n }\n}) %}\n\n{%- set first_act_bank = ({\"user\": \"You are a safety agent trying to help determine whether the last user message in the conversation is harmful according to the safety risk definition below.\",\"assistant\": \"You are a safety agent trying to help determine whether the last assistant message in the conversation is harmful according to the safety risk definition below.\",\"context_relevance\": \"You are a safety agent trying to help determine if the context poses a potential harm based on the provided risk definition and the user message.\",\"answer_relevance\": \"You are a safety agent trying to help determine whether the assistant message to user message poses a potential harm based on the provided risk definition.\",\"groundedness\": \"You are a safety agent trying to help determine if the assistant message poses a potential harm based on the provided risk definition and the given context.\",\"conv_risk_user\": \"You are a safety agent trying to help determine whether the last user message in the conversation poses a potential harm based on the provided risk definition.\",\"conv_risk_assistant\": \"You are a safety agent trying to help determine whether the last assistant message in the conversation poses a potential harm based on the provided risk definition.\",\"function_call\": \"You are a safety agent trying to help determine whether the assistant's message to the user's message, in conjunction with the available tools, poses any potential risks, as defined by the provided risk definition.\"}) %}\n\n{%- set third_act_bank = ({\"user\": \"Is the last user message harmful based on the risk definition? Your answer must be either 'Yes' or 'No'.\",\"assistant\": \"Is the last assistant message harmful based on the risk definition? Your answer must be either 'Yes' or 'No'.\", \"context_relevance\": \"Is the context harmful based on the risk definition? Your answer must be either 'Yes' or 'No'.\"}) %}\n\n{%- set primary_role = messages[-1].role %}\n{%- set first_act = first_act_bank[primary_role] %}\n{%- set third_act = third_act_bank[primary_role] %}\n\n{%- set requested_risk = none %}\n{%- set requested_definition = none %}\n\n{%- if guardian_config %}\n {%- if 'risk_name' not in guardian_config and 'risk_definition' not in guardian_config %}\n {{ raise_exception(\"either risk name or risk definition needs to be provided\") }}\n {%- elif guardian_config['risk_name'] in risk_bank and 'risk_definition' in guardian_config %}\n {{ raise_exception(\"existing risk name. can't overwrite definition for this risk\") }}\n {%- elif guardian_config['risk_name'] not in risk_bank and 'risk_definition' not in guardian_config %}\n {{ raise_exception(\"new risk name provided; risk definition is mandatory\") }}\n {%- endif %}\n \n {%- if 'risk_name' in guardian_config %}\n {%- set requested_risk = guardian_config['risk_name'] %}\n {%- endif %}\n\n {%- if 'risk_definition' in guardian_config %}\n {%- set requested_definition = guardian_config['risk_definition'] %}\n {%- endif %}\n{%- else %}\n {%- set requested_risk = \"harm\" %}\n{%- endif %}\n\n{%- if requested_definition is none %}\n {%- if primary_role in [\"user\", \"assistant\"] %}\n {%- set requested_definition = risk_bank[requested_risk][primary_role] %}\n {%- if requested_risk in [\"harm_engagement\", \"evasiveness\"] %}\n {%- set requested_definition = risk_bank[requested_risk][primary_role] %}\n {%- endif %}\n {%- endif %}\n {%- if requested_risk in [\"answer_relevance\", \"function_call\", \"groundedness\", \"context_relevance\"] %}\n {%- set requested_definition = risk_bank[requested_risk][\"default\"] %}\n {%- endif %}\n{%- endif %}\n\n{%- if guardian_config %}\n {%- if requested_risk in [\"evasiveness\", \"harm_engagement\"] %}\n {%- if primary_role == \"user\" %}\n {%- set first_act = first_act_bank.conv_risk_user %}\n {%- else %}\n {%- set first_act = first_act_bank.conv_risk_assistant %}\n {%- endif %}\n {%- elif requested_risk in [\"answer_relevance\", \"function_call\", \"groundedness\", \"context_relevance\"] %}\n {%- set first_act = first_act_bank[requested_risk] %}\n {%- if requested_risk == \"context_relevance\" %}\n {%- set third_act = third_act_bank.context_relevance %}\n {%- endif %}\n {%- endif %}\n{%- endif -%}\n\n<|start_of_role|>user<|end_of_role|>{{ first_act + '\\n'}}\n<start_of_conversation>\n{% for message in messages %}\n {% if message.role in [\"user\", \"assistant\"] %}\n {% if requested_risk == \"function_call\" and message.role == \"assistant\" %}\n {{- message.role.split()|map('capitalize')|join ~ ' ' \"Message: \" ~ message.content|tojson(indent=2) }}\n {% else %}\n {{- message.role.split()|map('capitalize')|join ~ ' ' \"Message: \" ~ message.content }}\n {% endif %}\n {% elif message.role in [\"tools\"] %}\n {{- \"Tools Available: \" ~ message.content | tojson(indent=2) }}\n {% elif message.role in [\"context\"] %}\n {{- \"Context: \" ~ message.content }}\n {% else %}\n {{ raise_exception(\"# invalid role\") }}\n {%- endif %}\n{% endfor -%}\n<end_of_conversation>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n- {{ requested_definition }}\n<end_of_risk_definition>\n\n{{ third_act + '<|end_of_text|>'}}\n{%- if add_generation_prompt %}\n {{- '\\n<|start_of_role|>assistant<|end_of_role|>' }}\n{%- endif %}",
190
+ "clean_up_tokenization_spaces": true,
191
+ "eos_token": "<|end_of_text|>",
192
+ "errors": "replace",
193
+ "extra_special_tokens": {},
194
+ "model_max_length": 8192,
195
+ "pad_token": "<|end_of_text|>",
196
+ "padding_side": "left",
197
+ "tokenizer_class": "GPT2Tokenizer",
198
+ "unk_token": "<|end_of_text|>",
199
+ "vocab_size": 49152
200
  }