Text Generation
Transformers
Safetensors
English
qwen2
conversational
text-generation-inference
michaelzhiluo commited on
Commit
103033d
·
verified ·
1 Parent(s): 055fec9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -96,6 +96,7 @@ This permissive license ensures that researchers, developers, and enthusiasts wo
96
 
97
  ## Acknowledgement
98
  - Our training experiments are powered by our heavily modified fork of [Verl](https://github.com/agentica-project/verl), an open-source post-training library.
 
99
  - Our model is trained on top of [`DeepSeek-R1-Distill-Qwen-1.5B`](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B).
100
  - Our work is done as part of [Berkeley Sky Computing Lab](https://skycomputing.berkeley.edu/) and [Berkeley AI Research](https://bair.berkeley.edu/).
101
 
@@ -104,7 +105,7 @@ This permissive license ensures that researchers, developers, and enthusiasts wo
104
  ```bibtex
105
  @misc{deepcoder2025,
106
  title={DeepCoder: A Fully Open-Source 14B Coder at O3-mini Level},
107
- author={Michael Luo, Sijun Tan, Roy Huang, Ameen Patel, Alpay Ariyak, Qingyang Wu, Xiaoxiang Shi, Rachel Xin, Colin Cai, Maurice Weber, Ce Zhang, Li Erran Li, Raluca Ada Popa, Ion Stoica, Tianjun Zhang},
108
  howpublished={\url{https://pretty-radio-b75.notion.site/DeepCoder-A-Fully-Open-Source-14B-Coder-at-O3-mini-Level-1cf81902c14680b3bee5eb349a512a51}},
109
  note={Notion Blog},
110
  year={2025}
 
96
 
97
  ## Acknowledgement
98
  - Our training experiments are powered by our heavily modified fork of [Verl](https://github.com/agentica-project/verl), an open-source post-training library.
99
+ - Notably, we train 1.5B with [verl pipeline](https://github.com/agentica-project/verl-pipeline), an extension of the original verl.
100
  - Our model is trained on top of [`DeepSeek-R1-Distill-Qwen-1.5B`](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B).
101
  - Our work is done as part of [Berkeley Sky Computing Lab](https://skycomputing.berkeley.edu/) and [Berkeley AI Research](https://bair.berkeley.edu/).
102
 
 
105
  ```bibtex
106
  @misc{deepcoder2025,
107
  title={DeepCoder: A Fully Open-Source 14B Coder at O3-mini Level},
108
+ author={Michael Luo, Sijun Tan, Roy Huang, Ameen Patel, Alpay Ariyak, Qingyang Wu, Xiaoxiang Shi, Rachel Xin, Colin Cai, Maurice Weber, Ce Zhang, Li Erran Li, Raluca Ada Popa, Ion Stoica},
109
  howpublished={\url{https://pretty-radio-b75.notion.site/DeepCoder-A-Fully-Open-Source-14B-Coder-at-O3-mini-Level-1cf81902c14680b3bee5eb349a512a51}},
110
  note={Notion Blog},
111
  year={2025}