Heegyu Kim PRO
heegyu
AI & ML interests
NLP
Recent Activity
liked
a dataset
2 days ago
data-agents/jupyter-agent-dataset
liked
a dataset
30 days ago
nvidia/Nemotron-Post-Training-Dataset-v1
liked
a dataset
about 1 month ago
bethgelab/CiteME
Organizations
Korean Reward Modeling
Korean Datasets, Reward Models for RLHF
-
heegyu/KoSafeGuard-8b-0503
Text Generation • 8B • Updated • 127 • 5 -
heegyu/ko-reward-model-helpful-1.3b-v0.2
Text Classification • 1B • Updated • 47 -
heegyu/ko-reward-model-safety-1.3b-v0.2
Text Classification • 1B • Updated • 2 • 5 -
heegyu/ko-reward-model-helpful-roberta-large-v0.1
Text Classification • 0.3B • Updated • 7 • 1
AjouBlue GPTs
AjouBlue-GPT: Small Korean Generative Language Models with 125M, 355M parameters
Synthetic Dataset
Reward Modeling Datasets
Vision LM
Domain Specific (Math, Code, etc)
Safety LM
R1-like Datasets
Korean Pretraining Dataset
Datasets Translated to Korean
RLHF papers
-
Pairwise Proximal Policy Optimization: Harnessing Relative Feedback for LLM Alignment
Paper • 2310.00212 • Published • 2 -
Stabilizing RLHF through Advantage Model and Selective Rehearsal
Paper • 2309.10202 • Published • 11 -
Aligning Language Models with Offline Reinforcement Learning from Human Feedback
Paper • 2308.12050 • Published • 1 -
Secrets of RLHF in Large Language Models Part I: PPO
Paper • 2307.04964 • Published • 29
Pre-training Dataset
Image Generation
Machine Translation
Text2SQL
SWE Agent
R1-like Datasets
Korean Reward Modeling
Korean Datasets, Reward Models for RLHF
-
heegyu/KoSafeGuard-8b-0503
Text Generation • 8B • Updated • 127 • 5 -
heegyu/ko-reward-model-helpful-1.3b-v0.2
Text Classification • 1B • Updated • 47 -
heegyu/ko-reward-model-safety-1.3b-v0.2
Text Classification • 1B • Updated • 2 • 5 -
heegyu/ko-reward-model-helpful-roberta-large-v0.1
Text Classification • 0.3B • Updated • 7 • 1
Korean Pretraining Dataset
AjouBlue GPTs
AjouBlue-GPT: Small Korean Generative Language Models with 125M, 355M parameters
Datasets Translated to Korean
Synthetic Dataset
RLHF papers
-
Pairwise Proximal Policy Optimization: Harnessing Relative Feedback for LLM Alignment
Paper • 2310.00212 • Published • 2 -
Stabilizing RLHF through Advantage Model and Selective Rehearsal
Paper • 2309.10202 • Published • 11 -
Aligning Language Models with Offline Reinforcement Learning from Human Feedback
Paper • 2308.12050 • Published • 1 -
Secrets of RLHF in Large Language Models Part I: PPO
Paper • 2307.04964 • Published • 29
Reward Modeling Datasets
Pre-training Dataset
Vision LM
Image Generation
Domain Specific (Math, Code, etc)
Machine Translation
Safety LM
Text2SQL