File size: 1,282 Bytes
2abd53f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a95272c
2abd53f
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
base_model: Pinkstack/llama-3.2-superthoughtslite-expert-chat
gate_mode: hidden # Common gating mechanism using hidden states. Alternatives: 'cheap_embed', 'random'
dtype: float16 # Use float16 to save memory/disk space, common for inference

experts:
  - source_model: Pinkstack/llama-3.2-superthoughtslite-expert-chat
    positive_prompts:
      - "General use"
      - "Conversational"
      - "Question answering"
      - "Multilingual"
      - "Translation"
      - "Roleplay"
  - source_model: Pinkstack/llama-3.2-superthoughts-expert-math
    positive_prompts:
      - "Mathematical"
      - "Algebra"
      - "Shape understanding"
      - "counting problem"
      - "Explain math"
      - "placing objects"
  - source_model: Pinkstack/llama-3.2-superthoughtslite-expert-medical
    positive_prompts:
      - "Medical"
      - "Biology"
      - "Science"
      - "Sickness"
      - "Illness"
      - "emotional reasoning" # Note: Might overlap slightly with general chat, use prompts carefully
  - source_model: Pinkstack/llama-3.2-superthoughts-lite-expert-code
    positive_prompts:
      - "Code generation"
      - "Debugging"
      - "Finish code"
      - "Explain code"
      - "Refine code"
      - "Coding assistance"

# MoE Specific Parameters
num_experts_per_tok: 2