Improve model card: Add pipeline tag, library name, paper abstract, and update paper link
#2
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,17 +1,23 @@
|
|
| 1 |
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
base_model:
|
| 4 |
- TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
|
| 5 |
- baichuan-inc/Baichuan2-7B-Base
|
| 6 |
- Qwen/Qwen3-8B
|
|
|
|
|
|
|
|
|
|
| 7 |
---
|
| 8 |
|
| 9 |
# Model Card for HLLM
|
| 10 |
|
| 11 |
[](https://arxiv.org/abs/2409.12740)
|
| 12 |
-
[](https://
|
| 13 |
[](https://github.com/bytedance/HLLM)
|
| 14 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 15 |
This repo is used for hosting HLLM and HLLM-Creator checkpoints.
|
| 16 |
|
| 17 |
For more details or tutorials see https://github.com/bytedance/HLLM.
|
|
@@ -56,4 +62,7 @@ author={Junyi Chen and Lu Chi and Siliang Xu and Shiwei Ran and Bingyue Peng and
|
|
| 56 |
journal={arXiv preprint arXiv:2508.18118},
|
| 57 |
year={2025}
|
| 58 |
}
|
| 59 |
-
```
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
|
|
|
| 2 |
base_model:
|
| 3 |
- TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
|
| 4 |
- baichuan-inc/Baichuan2-7B-Base
|
| 5 |
- Qwen/Qwen3-8B
|
| 6 |
+
license: apache-2.0
|
| 7 |
+
pipeline_tag: text-generation
|
| 8 |
+
library_name: transformers
|
| 9 |
---
|
| 10 |
|
| 11 |
# Model Card for HLLM
|
| 12 |
|
| 13 |
[](https://arxiv.org/abs/2409.12740)
|
| 14 |
+
[](https://huggingface.co/papers/2508.18118)
|
| 15 |
[](https://github.com/bytedance/HLLM)
|
| 16 |
|
| 17 |
+
## Paper Abstract
|
| 18 |
+
AI-generated content technologies are widely used in content creation. However, current AIGC systems rely heavily on creators' inspiration, rarely generating truly user-personalized content. In real-world applications such as online advertising, a single product may have multiple selling points, with different users focusing on different features. This underscores the significant value of personalized, user-centric creative generation. Effective personalized content generation faces two main challenges: (1) accurately modeling user interests and integrating them into the content generation process while adhering to factual constraints, and (2) ensuring high efficiency and scalability to handle the massive user base in industrial scenarios. Additionally, the scarcity of personalized creative data in practice complicates model training, making data construction another key hurdle. We propose HLLM-Creator, a hierarchical LLM framework for efficient user interest modeling and personalized content generation. During inference, a combination of user clustering and a user-ad-matching-prediction based pruning strategy is employed to significantly enhance generation efficiency and reduce computational overhead, making the approach suitable for large-scale deployment. Moreover, we design a data construction pipeline based on chain-of-thought reasoning, which generates high-quality, user-specific creative titles and ensures factual consistency despite limited personalized data. This pipeline serves as a critical foundation for the effectiveness of our model. Extensive experiments on personalized title generation for Douyin Search Ads show the effectiveness of HLLM-Creator. Online A/B test shows a 0.476% increase on Adss, paving the way for more effective and efficient personalized generation in industrial scenarios.
|
| 19 |
+
|
| 20 |
+
## Model Overview
|
| 21 |
This repo is used for hosting HLLM and HLLM-Creator checkpoints.
|
| 22 |
|
| 23 |
For more details or tutorials see https://github.com/bytedance/HLLM.
|
|
|
|
| 62 |
journal={arXiv preprint arXiv:2508.18118},
|
| 63 |
year={2025}
|
| 64 |
}
|
| 65 |
+
```
|
| 66 |
+
|
| 67 |
+
> Thanks to the excellent code repository [RecBole](https://github.com/RUCAIBox/RecBole), [VisRec](https://github.com/ialab-puc/VisualRecSys-Tutorial-IUI2021), [PixelRec](https://github.com/westlake-repl/PixelRec) and [HSTU](https://github.com/facebookresearch/generative-recommenders/tree/main) !
|
| 68 |
+
> HLLM is released under the Apache License 2.0, some codes are modified from HSTU and PixelRec, which are released under the Apache License 2.0 and MIT License, respectively.
|