Model Summary

Neuralphi-2 is an experiment in DPO finetuning. It was made following Max Labonne's excellent article about fine-tuning mistral-7b. Neuralphi-2 is phi-2-sft finetuned using DPO with Intel/orca_dpo_pairs.

Prompt Format

"""### Human: {instruction}

### Assistant:"""
Downloads last month
2
Safetensors
Model size
2.78B params
Tensor type
FP16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train xz56/neuralphi-2