CodeBERT: A Pre-Trained Model for Programming and Natural Languages Paper • 2002.08155 • Published Feb 19, 2020 • 2
Text Generation with Diffusion Language Models: A Pre-training Approach with Continuous Paragraph Denoise Paper • 2212.11685 • Published Dec 22, 2022 • 2
ByT5: Towards a token-free future with pre-trained byte-to-byte models Paper • 2105.13626 • Published May 28, 2021 • 3