File size: 1,282 Bytes
2abd53f a95272c 2abd53f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 |
base_model: Pinkstack/llama-3.2-superthoughtslite-expert-chat
gate_mode: hidden # Common gating mechanism using hidden states. Alternatives: 'cheap_embed', 'random'
dtype: float16 # Use float16 to save memory/disk space, common for inference
experts:
- source_model: Pinkstack/llama-3.2-superthoughtslite-expert-chat
positive_prompts:
- "General use"
- "Conversational"
- "Question answering"
- "Multilingual"
- "Translation"
- "Roleplay"
- source_model: Pinkstack/llama-3.2-superthoughts-expert-math
positive_prompts:
- "Mathematical"
- "Algebra"
- "Shape understanding"
- "counting problem"
- "Explain math"
- "placing objects"
- source_model: Pinkstack/llama-3.2-superthoughtslite-expert-medical
positive_prompts:
- "Medical"
- "Biology"
- "Science"
- "Sickness"
- "Illness"
- "emotional reasoning" # Note: Might overlap slightly with general chat, use prompts carefully
- source_model: Pinkstack/llama-3.2-superthoughts-lite-expert-code
positive_prompts:
- "Code generation"
- "Debugging"
- "Finish code"
- "Explain code"
- "Refine code"
- "Coding assistance"
# MoE Specific Parameters
num_experts_per_tok: 2 |