Update README.md
Browse files
README.md
CHANGED
@@ -5,6 +5,13 @@ base_model:
|
|
5 |
---
|
6 |
|
7 |
# 🤖 Model Card: InfiX-ai/InfiAlign-Qwen-7B-DPO
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
8 |
|
9 |
**InfiAlign** is a scalable and data-efficient post-training framework that combines supervised fine-tuning (SFT) and Direct Preference Optimization (DPO) with a high-quality data selection pipeline to enhance reasoning in large language models.
|
10 |
|
|
|
5 |
---
|
6 |
|
7 |
# 🤖 Model Card: InfiX-ai/InfiAlign-Qwen-7B-DPO
|
8 |
+
<p align="center">
|
9 |
+
<a href="https://arxiv.org/abs/2508.05496"><img src="https://img.shields.io/badge/arXiv-Paper-b31b1b?style=flat&logo=arxiv&logoColor=white" alt="arXiv Paper"></a>
|
10 |
+
<a href="https://huggingface.co/papers/2508.05496"><img src="https://img.shields.io/badge/🤗%20HuggingFace-Daily%20Papers-ff9800?style=flat" alt="Hugging Face Paper"></a>
|
11 |
+
<a href="https://huggingface.co/InfiX-ai/InfiAlign-Qwen-7B-SFT"><img src="https://img.shields.io/badge/🤗%20HuggingFace-SFT%20Model-ff9800?style=flat" alt="Hugging Face SFT Model"></a>
|
12 |
+
<a href="https://huggingface.co/InfiX-ai/InfiAlign-Qwen-7B-DPO"><img src="https://img.shields.io/badge/🤗%20HuggingFace-DPO%20Model-ff9800?style=flat" alt="Hugging Face DPO Model"></a>
|
13 |
+
<a href="https://github.com/InfiXAI/InfiAlign"><img src="https://img.shields.io/badge/GitHub-Repository-181717?style=flat&logo=github&logoColor=white" alt="GitHub Repository"></a>
|
14 |
+
</p>
|
15 |
|
16 |
**InfiAlign** is a scalable and data-efficient post-training framework that combines supervised fine-tuning (SFT) and Direct Preference Optimization (DPO) with a high-quality data selection pipeline to enhance reasoning in large language models.
|
17 |
|