This project is sponsored by PrimeLine

Model Card

This model is an finetuned version for german instructions and conversations in style of Open Assistant tokens. "<|prompter|>" "<|endoftext|>" "<|assistant|>"

The dataset used is deduplicated and cleaned, with no codes inside. The focus is on instruction following and conversational tasks.

The model archictecture is based on falcon with 7B parameters, trained on 100% renewable energy powered hardware.

This work is contributed by private research of flozi00

Downloads last month
23
Safetensors
Model size
6.92B params
Tensor type
FP16
·
Inference Examples
Inference API (serverless) does not yet support model repos that contain custom code.

Dataset used to train flozi00/falcon-7b-german-assistant-v2