Spaces:
Running
Running
viktor
commited on
Commit
·
3ed8b70
1
Parent(s):
76914ae
feat. add max_input_tokens config
Browse files- src/lib/constants.ts +5 -0
- src/lib/inference-utils.ts +3 -2
src/lib/constants.ts
CHANGED
@@ -3,6 +3,7 @@ export const REDIRECT_URI =
|
|
3 |
|
4 |
export const MODEL_CONFIG_PROMPT_IMPROVEMENT = {
|
5 |
id: "deepseek-ai/DeepSeek-V3-0324",
|
|
|
6 |
max_tokens: 16_000,
|
7 |
}
|
8 |
|
@@ -11,24 +12,28 @@ const default_system_prompt = `ONLY USE HTML, CSS AND JAVASCRIPT. If you want to
|
|
11 |
export const MODEL_CONFIG_CODE_GENERATION = [
|
12 |
{
|
13 |
id: "deepseek-ai/DeepSeek-V3-0324",
|
|
|
14 |
max_tokens: 16_000,
|
15 |
default_enable_thinking: false,
|
16 |
system_prompt: default_system_prompt,
|
17 |
},
|
18 |
{
|
19 |
id: "Qwen/Qwen3-235B-A22B",
|
|
|
20 |
max_tokens: 16_000,
|
21 |
default_enable_thinking: true,
|
22 |
system_prompt: default_system_prompt,
|
23 |
},
|
24 |
{
|
25 |
id: "Qwen/Qwen3-30B-A3B",
|
|
|
26 |
max_tokens: 16_000,
|
27 |
default_enable_thinking: true,
|
28 |
system_prompt: default_system_prompt,
|
29 |
},
|
30 |
{
|
31 |
id: "Qwen/Qwen3-32B",
|
|
|
32 |
max_tokens: 16_000,
|
33 |
default_enable_thinking: true,
|
34 |
system_prompt: default_system_prompt,
|
|
|
3 |
|
4 |
export const MODEL_CONFIG_PROMPT_IMPROVEMENT = {
|
5 |
id: "deepseek-ai/DeepSeek-V3-0324",
|
6 |
+
max_input_tokens: 48_000,
|
7 |
max_tokens: 16_000,
|
8 |
}
|
9 |
|
|
|
12 |
export const MODEL_CONFIG_CODE_GENERATION = [
|
13 |
{
|
14 |
id: "deepseek-ai/DeepSeek-V3-0324",
|
15 |
+
max_input_tokens: 48_000,
|
16 |
max_tokens: 16_000,
|
17 |
default_enable_thinking: false,
|
18 |
system_prompt: default_system_prompt,
|
19 |
},
|
20 |
{
|
21 |
id: "Qwen/Qwen3-235B-A22B",
|
22 |
+
max_input_tokens: 24_000,
|
23 |
max_tokens: 16_000,
|
24 |
default_enable_thinking: true,
|
25 |
system_prompt: default_system_prompt,
|
26 |
},
|
27 |
{
|
28 |
id: "Qwen/Qwen3-30B-A3B",
|
29 |
+
max_input_tokens: 24_000,
|
30 |
max_tokens: 16_000,
|
31 |
default_enable_thinking: true,
|
32 |
system_prompt: default_system_prompt,
|
33 |
},
|
34 |
{
|
35 |
id: "Qwen/Qwen3-32B",
|
36 |
+
max_input_tokens: 24_000,
|
37 |
max_tokens: 16_000,
|
38 |
default_enable_thinking: true,
|
39 |
system_prompt: default_system_prompt,
|
src/lib/inference-utils.ts
CHANGED
@@ -7,6 +7,7 @@ const sharedEncoder = new TextEncoder();
|
|
7 |
export interface ModelConfig {
|
8 |
id: string;
|
9 |
max_tokens: number;
|
|
|
10 |
system_prompt?: string;
|
11 |
default_enable_thinking?: boolean;
|
12 |
}
|
@@ -31,11 +32,11 @@ export async function getInferenceToken(request: NextRequest): Promise<{
|
|
31 |
}
|
32 |
|
33 |
export function checkTokenLimit(tokensUsed: number, modelConfig: ModelConfig) {
|
34 |
-
if (tokensUsed >= modelConfig.
|
35 |
return {
|
36 |
ok: false,
|
37 |
openSelectProvider: true,
|
38 |
-
message: `Context is too long. ${modelConfig.id} allow ${modelConfig.
|
39 |
};
|
40 |
}
|
41 |
return null;
|
|
|
7 |
export interface ModelConfig {
|
8 |
id: string;
|
9 |
max_tokens: number;
|
10 |
+
max_input_tokens: number;
|
11 |
system_prompt?: string;
|
12 |
default_enable_thinking?: boolean;
|
13 |
}
|
|
|
32 |
}
|
33 |
|
34 |
export function checkTokenLimit(tokensUsed: number, modelConfig: ModelConfig) {
|
35 |
+
if (tokensUsed >= modelConfig.max_input_tokens) {
|
36 |
return {
|
37 |
ok: false,
|
38 |
openSelectProvider: true,
|
39 |
+
message: `Context is too long. ${modelConfig.id} allow ${modelConfig.max_input_tokens} max input tokens.`,
|
40 |
};
|
41 |
}
|
42 |
return null;
|