merge

This is a merge of pre-trained language models created using mergekit.

Merge Details

<|im_start|>system<|im_sep|> {system_message}<|im_end|> <|im_start|>user<|im_sep|> {prompt}<|im_end|> <|im_start|>assistant<|im_sep|>

Merge Method

This model was merged using the Model Stock merge method using Pinkstack/SuperThoughts-CoT-14B-16k-o1-QwQ as a base.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

models:
  - model: bunnycore/Phi-4-Model-Stock-v2
  - model: Pinkstack/SuperThoughts-CoT-14B-16k-o1-QwQ
  - model: Pinkstack/SuperThoughts-CoT-14B-16k-o1-QwQ+bunnycore/Phi-4-rp-v1-lora
base_model: Pinkstack/SuperThoughts-CoT-14B-16k-o1-QwQ
merge_method: model_stock
parameters:
  normalize: true
dtype: bfloat16
tokenizer_source: unsloth/phi-4

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 37.32
IFEval (0-Shot) 59.12
BBH (3-Shot) 52.78
MATH Lvl 5 (4-Shot) 46.90
GPQA (0-shot) 5.26
MuSR (0-shot) 11.18
MMLU-PRO (5-shot) 48.68
Downloads last month
42
Safetensors
Model size
14.7B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for bunnycore/Phi-4-Model-Stock-v3

Evaluation results