-
BlackMamba: Mixture of Experts for State-Space Models
Paper • 2402.01771 • Published • 24 -
OpenMoE: An Early Effort on Open Mixture-of-Experts Language Models
Paper • 2402.01739 • Published • 27 -
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
Paper • 2401.06066 • Published • 50
David Samuel
Davidsamuel101
AI & ML interests
NLP, Computer Vision
Recent Activity
updated
a collection
8 days ago
Bookbot Recording's Dataset
updated
a dataset
13 days ago
bookbot/en_snapshot_madison_vc
published
a dataset
15 days ago
bookbot/en_snapshot_madison_vc