ModernBERT-ecoRouter

ModernBERT-ecoRouter is a classifier designed to route chatbot prompts to the most appropriate model based on task complexity. It predicts whether a user input should be handled by a small or large language model, helping reduce unnecessary compute for simple tasks.

This model is used in the โ€œDo I really need a huge LLM?โ€ demo, where a chatbot interface reveals which model responded to each user message based on routing decisions from this classifier.

Intended use

The model takes a single user prompt as input and returns a label:

  • small if the input can likely be handled well by a small model (e.g. TinyLlama, Phi)
  • large if it would likely benefit from a larger model (e.g. Mistral, GPT-4)

This enables dynamic model selection in chatbot or API systems where cost and latency matter.

Training data

The model was fine-tuned on a mix of samples from:

A Mistral model was used to classify prompts into small or large categories based on expected complexity of the task for an LLM.

Performance

On a held-out test set, the classifier achieves:

  • ~80% accuracy

Limitations

  • The model is trained on single prompts, not full conversations
  • Some prompts are ambiguous without user history
  • It may overpredict large in borderline cases
Downloads last month
13
Safetensors
Model size
150M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Space using monsimas/ModernBERT-ecoRouter 1