hp-l33 commited on
Commit
ef784a3
·
verified ·
1 Parent(s): 2e05679

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -3
README.md CHANGED
@@ -11,15 +11,13 @@ library_name: pytorch
11
  <sup>1</sup> Westlake University,
12
  <sup>2</sup> Institute of Automation, Chinese Academy of Sciences
13
 
14
- [![arXiv](https://img.shields.io/badge/arXiv-2503.10568-A42C25?style=flat&logo=arXiv)](https://arxiv.org/abs/2503.10568) [![Project](https://img.shields.io/badge/Project-Page-green?style=flat&logo=Google%20chrome&logoColor=green)](https://hp-l33.github.io/projects/arpg) [![HuggingFace](https://img.shields.io/badge/HuggingFace-Model-blue?style=flat&logo=HuggingFace)](https://huggingface.co/hp-l33/ARPG)
15
-
16
  ## News
17
  * **2025-03-14**: The paper and code are released!
18
 
19
  ## Introduction
20
  We introduce a novel autoregressive image generation framework named **ARPG**. This framework is capable of conducting **BERT-style masked modeling** by employing a **GPT-style causal architecture**. Consequently, it is able to generate images in parallel following a random token order and also provides support for the KV cache.
21
  * 💪 **ARPG** achieves an FID of **1.94**
22
- * 🚀 **ARPG** delivers throughput **26 times faster** than [LlamaGen](https://github.com/FoundationVision/LlamaGen)—nearly matching [VAR](https://github.com/FoundationVision/VAR)
23
  * ♻️ **ARPG** reducing memory consumption by over **75%** compared to [VAR](https://github.com/FoundationVision/VAR).
24
  * 🔍 **ARPG** supports **zero-shot inference** (e.g., inpainting and outpainting).
25
  * 🛠️ **ARPG** can be easily extended to **controllable generation**.
 
11
  <sup>1</sup> Westlake University,
12
  <sup>2</sup> Institute of Automation, Chinese Academy of Sciences
13
 
 
 
14
  ## News
15
  * **2025-03-14**: The paper and code are released!
16
 
17
  ## Introduction
18
  We introduce a novel autoregressive image generation framework named **ARPG**. This framework is capable of conducting **BERT-style masked modeling** by employing a **GPT-style causal architecture**. Consequently, it is able to generate images in parallel following a random token order and also provides support for the KV cache.
19
  * 💪 **ARPG** achieves an FID of **1.94**
20
+ * 🚀 **ARPG** delivers throughput **26 times faster** than [LlamaGen](https://github.com/FoundationVision/LlamaGen).
21
  * ♻️ **ARPG** reducing memory consumption by over **75%** compared to [VAR](https://github.com/FoundationVision/VAR).
22
  * 🔍 **ARPG** supports **zero-shot inference** (e.g., inpainting and outpainting).
23
  * 🛠️ **ARPG** can be easily extended to **controllable generation**.