File size: 1,716 Bytes
12ac9a7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
---
language:
- de
pipeline_tag: text-generation
library_name: transformers
tags:
- bloom
- LLM
inference: false
widget:
- text: "TODO"
---

<div style="text-align:center;width:250px;height:250px;">
    <img src="https://huggingface.co/philschmid/instruct-igel-001/resolve/mainigel_removedbg.png" alt="IGEL logo"">
</div>



# IGEL: Instruction-tuned German large Language Model for Text

IGEL is a LLM model family developed for German. The first version of IGEL is built on top [BigScience BLOOM](https://bigscience.huggingface.co/blog/bloom) adapted to the [German language by Malte Ostendorff](https://huggingface.co/malteos/bloom-6b4-clp-german). IGEL designed to provide accurate and reliable language understanding capabilities for a wide range of natural language understanding tasks, including sentiment analysis, language translation, and question answering.

You can try out the model at [igel-playground]().

The IGEL family includes instruction `instruct-igel-001` and `chat-igel-001` _coming soon_.


## Model Description

LoRA tuned [BLOOM-CLP German (6.4B parameters)](https://huggingface.co/malteos/bloom-6b4-clp-german) with merged weights. 

## Training data

`instruct-igel-001` is trained on naive translated instruction datasets, without much post-processing. 

### Known limitations

`instruct-igel-001` also exhibits several common deficiencies of language models, including hallucination, toxicity, and stereotypes. 

For example, in the following figure, `instruct-igel-001` wrongly says that the cancelor of Germany is Angela Merkel. 

![cancelor](./assets/cancelor.png)


### Training procedure

_coming soon_

## How to use

You can test the model in this LLM playground.

_coming soon_