viktor commited on
Commit
3ed8b70
·
1 Parent(s): 76914ae

feat. add max_input_tokens config

Browse files
src/lib/constants.ts CHANGED
@@ -3,6 +3,7 @@ export const REDIRECT_URI =
3
 
4
  export const MODEL_CONFIG_PROMPT_IMPROVEMENT = {
5
  id: "deepseek-ai/DeepSeek-V3-0324",
 
6
  max_tokens: 16_000,
7
  }
8
 
@@ -11,24 +12,28 @@ const default_system_prompt = `ONLY USE HTML, CSS AND JAVASCRIPT. If you want to
11
  export const MODEL_CONFIG_CODE_GENERATION = [
12
  {
13
  id: "deepseek-ai/DeepSeek-V3-0324",
 
14
  max_tokens: 16_000,
15
  default_enable_thinking: false,
16
  system_prompt: default_system_prompt,
17
  },
18
  {
19
  id: "Qwen/Qwen3-235B-A22B",
 
20
  max_tokens: 16_000,
21
  default_enable_thinking: true,
22
  system_prompt: default_system_prompt,
23
  },
24
  {
25
  id: "Qwen/Qwen3-30B-A3B",
 
26
  max_tokens: 16_000,
27
  default_enable_thinking: true,
28
  system_prompt: default_system_prompt,
29
  },
30
  {
31
  id: "Qwen/Qwen3-32B",
 
32
  max_tokens: 16_000,
33
  default_enable_thinking: true,
34
  system_prompt: default_system_prompt,
 
3
 
4
  export const MODEL_CONFIG_PROMPT_IMPROVEMENT = {
5
  id: "deepseek-ai/DeepSeek-V3-0324",
6
+ max_input_tokens: 48_000,
7
  max_tokens: 16_000,
8
  }
9
 
 
12
  export const MODEL_CONFIG_CODE_GENERATION = [
13
  {
14
  id: "deepseek-ai/DeepSeek-V3-0324",
15
+ max_input_tokens: 48_000,
16
  max_tokens: 16_000,
17
  default_enable_thinking: false,
18
  system_prompt: default_system_prompt,
19
  },
20
  {
21
  id: "Qwen/Qwen3-235B-A22B",
22
+ max_input_tokens: 24_000,
23
  max_tokens: 16_000,
24
  default_enable_thinking: true,
25
  system_prompt: default_system_prompt,
26
  },
27
  {
28
  id: "Qwen/Qwen3-30B-A3B",
29
+ max_input_tokens: 24_000,
30
  max_tokens: 16_000,
31
  default_enable_thinking: true,
32
  system_prompt: default_system_prompt,
33
  },
34
  {
35
  id: "Qwen/Qwen3-32B",
36
+ max_input_tokens: 24_000,
37
  max_tokens: 16_000,
38
  default_enable_thinking: true,
39
  system_prompt: default_system_prompt,
src/lib/inference-utils.ts CHANGED
@@ -7,6 +7,7 @@ const sharedEncoder = new TextEncoder();
7
  export interface ModelConfig {
8
  id: string;
9
  max_tokens: number;
 
10
  system_prompt?: string;
11
  default_enable_thinking?: boolean;
12
  }
@@ -31,11 +32,11 @@ export async function getInferenceToken(request: NextRequest): Promise<{
31
  }
32
 
33
  export function checkTokenLimit(tokensUsed: number, modelConfig: ModelConfig) {
34
- if (tokensUsed >= modelConfig.max_tokens) {
35
  return {
36
  ok: false,
37
  openSelectProvider: true,
38
- message: `Context is too long. ${modelConfig.id} allow ${modelConfig.max_tokens} max tokens.`,
39
  };
40
  }
41
  return null;
 
7
  export interface ModelConfig {
8
  id: string;
9
  max_tokens: number;
10
+ max_input_tokens: number;
11
  system_prompt?: string;
12
  default_enable_thinking?: boolean;
13
  }
 
32
  }
33
 
34
  export function checkTokenLimit(tokensUsed: number, modelConfig: ModelConfig) {
35
+ if (tokensUsed >= modelConfig.max_input_tokens) {
36
  return {
37
  ok: false,
38
  openSelectProvider: true,
39
+ message: `Context is too long. ${modelConfig.id} allow ${modelConfig.max_input_tokens} max input tokens.`,
40
  };
41
  }
42
  return null;