File size: 1,516 Bytes
12ac9a7 d999884 12ac9a7 74ce4a1 12ac9a7 74ce4a1 12ac9a7 2b468fd 12ac9a7 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 |
---
language:
- de
pipeline_tag: text-generation
library_name: transformers
tags:
- bloom
- LLM
inference: false
widget:
- text: "TODO"
---
# IGEL: Instruction-tuned German large Language Model for Text
IGEL is a LLM model family developed for German. The first version of IGEL is built on top [BigScience BLOOM](https://bigscience.huggingface.co/blog/bloom) adapted to [German from Malte Ostendorff](https://huggingface.co/malteos/bloom-6b4-clp-german). IGEL designed to provide accurate and reliable language understanding capabilities for a wide range of natural language understanding tasks, including sentiment analysis, language translation, and question answering.
### You can try out the model at [igel-playground]().
The IGEL family includes `instruct-igel-001` and `chat-igel-001` _coming soon_.
## Model Description
LoRA tuned [BLOOM-CLP German (6.4B parameters)](https://huggingface.co/malteos/bloom-6b4-clp-german) with merged weights.
## Training data
`instruct-igel-001` is trained on naive translated instruction datasets, without much post-processing.
### Known limitations
`instruct-igel-001` also exhibits several common deficiencies of language models, including hallucination, toxicity, and stereotypes.
For example, in the following figure, `instruct-igel-001` wrongly says that the cancelor of Germany is Angela Merkel.
![cancelor](./assets/cancelor.png)
### Training procedure
_coming soon_
## How to use
You can test the model in this LLM playground.
_coming soon_
|