image/png

Mistral-Gutenberg-Doppel-7B-FFT

mistralai/Mistral-7B-Instruct-v0.2 finetuned on jondurbin/gutenberg-dpo-v0.1 and nbeerbower/gutenberg2-dpo.

This is a full finetune rather than my usual QLoRA tunes. Mostly for learning purposes.

Method

ORPO tuned with 4x A100 for 2 epochs.

Downloads last month
72
Safetensors
Model size
7.24B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for nbeerbower/Mistral-Gutenberg-Doppel-7B-FFT

Finetuned
(373)
this model
Quantizations
3 models

Datasets used to train nbeerbower/Mistral-Gutenberg-Doppel-7B-FFT