Merge AIDC-ai-business/Marcoroni-7B-v3 and rwitz/go-bruins-v2 using slerp merge from https://github.com/cg123/mergekit. After that we trained DPO with HF data

config.yaml

slices:
  - sources:
      - model: AIDC-ai-business/Marcoroni-7B-v3
        layer_range: [0, 32]
      - model: rwitz/go-bruins-v2
        layer_range: [0, 32]
merge_method: slerp
base_model: AIDC-ai-business/Marcoroni-7B-v3
parameters:
  t:
    - filter: self_attn
      value: [0, 0.5, 0.3, 0.7, 1]
    - filter: mlp
      value: [1, 0.5, 0.7, 0.3, 0]
    - value: 0.5 
dtype: float16

You can use alpaca template.

template_format = """{system}
### Instruction:
{prompt}

### Response:
"""
Downloads last month
1,145
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for viethq188/Rabbit-7B-DPO-Chat

Quantizations
1 model

Spaces using viethq188/Rabbit-7B-DPO-Chat 6