Byron Gibson
bgibson
·
AI & ML interests
None yet
Organizations
None yet
papers
-
Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model
Paper • 2401.09417 • Published • 63 -
MoE-Mamba: Efficient Selective State Space Models with Mixture of Experts
Paper • 2401.04081 • Published • 73 -
SwitchHead: Accelerating Transformers with Mixture-of-Experts Attention
Paper • 2312.07987 • Published • 41 -
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
Paper • 2401.06066 • Published • 55
llm-models
llm-local
-
LLM in a flash: Efficient Large Language Model Inference with Limited Memory
Paper • 2312.11514 • Published • 258 -
Running9595
Transformers to Core ML
⚡ -
enterprise-explorers/Llama-2-7b-chat-coreml
Text Generation • Updated • 2.62k • 138 -
tiiuae/falcon-7b-instruct
Text Generation • 7B • Updated • 143k • 1k
llm-datasets
llm-analysis
llm-local
-
LLM in a flash: Efficient Large Language Model Inference with Limited Memory
Paper • 2312.11514 • Published • 258 -
Running9595
Transformers to Core ML
⚡ -
enterprise-explorers/Llama-2-7b-chat-coreml
Text Generation • Updated • 2.62k • 138 -
tiiuae/falcon-7b-instruct
Text Generation • 7B • Updated • 143k • 1k
papers
-
Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model
Paper • 2401.09417 • Published • 63 -
MoE-Mamba: Efficient Selective State Space Models with Mixture of Experts
Paper • 2401.04081 • Published • 73 -
SwitchHead: Accelerating Transformers with Mixture-of-Experts Attention
Paper • 2312.07987 • Published • 41 -
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
Paper • 2401.06066 • Published • 55
llm-datasets
llm-models