--- title: README emoji: ๐Ÿš€ colorFrom: red colorTo: gray sdk: static pinned: false --- We advance German NLP through transparent, open-source research with three flagship projects: **[๐Ÿ‘ LLรคMmlein](https://www.informatik.uni-wuerzburg.de/datascience/projects/nlp/llammlein/)** - Comprehensive family of German-only transformer models (120M, 1B, 7B parameters) trained transparently from scratch with full training data and code documentation. **[๐Ÿ“Š SuperGLEBer](https://lsx-uniwue.github.io/SuperGLEBer-site/leaderboard_v1)** - First comprehensive German benchmark suite featuring 29 diverse NLP tasks across domains, providing systematic evaluation for German language models. **[๐Ÿค– ModernGBERT](https://arxiv.org/abs/2505.13136)** - Transparent encoder models (138M, 1B parameters) based on modernBERT architecture, specifically optimized for German language understanding. Beyond our German NLP ecosystem, we specialize in **LLM-Knowledge Graph Integration** for text mining applications. Our work combines language models with explicit knowledge representations, developing: - **Character Analysis**: Models like LitBERT for understanding character networks in novels - **Temporal Text Analysis**: Tracking narrative development through relation detection and scene segmentation - **Sentiment & Engagement Analysis**: Measuring emotional dynamics in streaming platforms and social media - **Knowledge Enrichment**: Semantic Web technologies for ontology learning and KG enhancement ๐ŸŒธ We foster reproducible, collaborative research by open-sourcing models, datasets, and evaluation frameworks, establishing German as a first-class language in the global AI ecosystem while advancing the intersection of symbolic knowledge and neural language understanding.