Clinical ModernBERT: An efficient and long context encoder for biomedical text
Abstract
We introduce Clinical ModernBERT, a transformer based encoder pretrained on large scale biomedical literature, clinical notes, and medical ontologies, incorporating PubMed abstracts, MIMIC IV clinical data, and medical codes with their textual descriptions. Building on ModernBERT the current state of the art natural language text encoder featuring architectural upgrades such as rotary positional embeddings (RoPE), Flash Attention, and extended context length up to 8,192 tokens our model adapts these innovations specifically for biomedical and clinical domains. Clinical ModernBERT excels at producing semantically rich representations tailored for long context tasks. We validate this both by analyzing its pretrained weights and through empirical evaluation on a comprehensive suite of clinical NLP benchmarks.
Community
A new long and efficient encoder for biomedical text. Our goal was to follow the experimental setup of BioClinicalBERT (https://arxiv.org/abs/1904.03323), and find a new modern replacement for biomedical text that can benefit from token sequences beyond 512
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- GLiNER-biomed: A Suite of Efficient Models for Open Biomedical Named Entity Recognition (2025)
- Generating Clinically Realistic EHR Data via a Hierarchy- and Semantics-Guided Transformer (2025)
- Applications of Large Models in Medicine (2025)
- Improving Clinical Question Answering with Multi-Task Learning: A Joint Approach for Answer Extraction and Medical Categorization (2025)
- Large Language Models are Powerful EHR Encoders (2025)
- Towards Scalable and Cross-Lingual Specialist Language Models for Oncology (2025)
- Med-gte-hybrid: A contextual embedding transformer model for extracting actionable information from clinical texts (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 1
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper