PEFT
Safetensors
German
llama
trl
sft
Generated from Trainer

LLäMmlein 7B Chat

image/png

While the base versions of our LLäMmlein are quite good, our chat versions are research demonstrations and are not ready to be used in settings where close instruction following is necessary. Please check the paper for more details.

This is an early preview of our instruction-tuned 7B model, trained using limited German-language resources. Please note that it is not the final version - we are actively working on improvements!

Find more details on our page and our preprint!

Example Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("LSX-UniWue/LLaMmlein_7B_chat")
tokenizer = AutoTokenizer.from_pretrained("LSX-UniWue/LLaMmlein_7B_chat")
model = model.to("mps")

messages = [
    {
        "role": "user",
        "content": "Was sind die wichtigsten Sehenswürdigkeiten von Berlin?",
    },
]

chat = tokenizer.apply_chat_template(
    messages,
    return_tensors="pt",
    add_generation_prompt=True,
).to("mps")


print(
    tokenizer.decode(
        model.generate(
            chat,
            max_new_tokens=100,
            pad_token_id=tokenizer.pad_token_id,
            eos_token_id=tokenizer.eos_token_id,
            repetition_penalty=1.1,
        )[0],
        skip_special_tokens=False,
    )
)
Downloads last month
-
Safetensors
Model size
6.74B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for LSX-UniWue/LLaMmlein_7B_chat

Adapter
(1)
this model
Finetunes
1 model
Quantizations
2 models

Datasets used to train LSX-UniWue/LLaMmlein_7B_chat

Collection including LSX-UniWue/LLaMmlein_7B_chat