---
language:
- de
pipeline_tag: text-generation
library_name: transformers
tags:
- bloom
- LLM
inference: false
widget:
- text: "TODO"
---
# IGEL: Instruction-tuned German large Language Model for Text
IGEL is a LLM model family developed for German. The first version of IGEL is built on top [BigScience BLOOM](https://bigscience.huggingface.co/blog/bloom) adapted to [German from Malte Ostendorff](https://huggingface.co/malteos/bloom-6b4-clp-german). IGEL designed to provide accurate and reliable language understanding capabilities for a wide range of natural language understanding tasks, including sentiment analysis, language translation, and question answering.
### You can try out the model at [igel-playground]().
The IGEL family includes instruction `instruct-igel-001` and `chat-igel-001` _coming soon_.
## Model Description
LoRA tuned [BLOOM-CLP German (6.4B parameters)](https://huggingface.co/malteos/bloom-6b4-clp-german) with merged weights.
## Training data
`instruct-igel-001` is trained on naive translated instruction datasets, without much post-processing.
### Known limitations
`instruct-igel-001` also exhibits several common deficiencies of language models, including hallucination, toxicity, and stereotypes.
For example, in the following figure, `instruct-igel-001` wrongly says that the cancelor of Germany is Angela Merkel.
![cancelor](./assets/cancelor.png)
### Training procedure
_coming soon_
## How to use
You can test the model in this LLM playground.
_coming soon_