sslu commited on
Commit
bd48c6e
·
verified ·
1 Parent(s): b0ede44

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -0
README.md CHANGED
@@ -5,6 +5,13 @@ base_model:
5
  ---
6
 
7
  # 🤖 Model Card: InfiX-ai/InfiAlign-Qwen-7B-DPO
 
 
 
 
 
 
 
8
 
9
  **InfiAlign** is a scalable and data-efficient post-training framework that combines supervised fine-tuning (SFT) and Direct Preference Optimization (DPO) with a high-quality data selection pipeline to enhance reasoning in large language models.
10
 
 
5
  ---
6
 
7
  # 🤖 Model Card: InfiX-ai/InfiAlign-Qwen-7B-DPO
8
+ <p align="center">
9
+ <a href="https://arxiv.org/abs/2508.05496"><img src="https://img.shields.io/badge/arXiv-Paper-b31b1b?style=flat&logo=arxiv&logoColor=white" alt="arXiv Paper"></a>
10
+ <a href="https://huggingface.co/papers/2508.05496"><img src="https://img.shields.io/badge/🤗%20HuggingFace-Daily%20Papers-ff9800?style=flat" alt="Hugging Face Paper"></a>
11
+ <a href="https://huggingface.co/InfiX-ai/InfiAlign-Qwen-7B-SFT"><img src="https://img.shields.io/badge/🤗%20HuggingFace-SFT%20Model-ff9800?style=flat" alt="Hugging Face SFT Model"></a>
12
+ <a href="https://huggingface.co/InfiX-ai/InfiAlign-Qwen-7B-DPO"><img src="https://img.shields.io/badge/🤗%20HuggingFace-DPO%20Model-ff9800?style=flat" alt="Hugging Face DPO Model"></a>
13
+ <a href="https://github.com/InfiXAI/InfiAlign"><img src="https://img.shields.io/badge/GitHub-Repository-181717?style=flat&logo=github&logoColor=white" alt="GitHub Repository"></a>
14
+ </p>
15
 
16
  **InfiAlign** is a scalable and data-efficient post-training framework that combines supervised fine-tuning (SFT) and Direct Preference Optimization (DPO) with a high-quality data selection pipeline to enhance reasoning in large language models.
17