Byron Gibson
bgibson
·
AI & ML interests
None yet
Organizations
None yet
papers
-
Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model
Paper • 2401.09417 • Published • 62 -
MoE-Mamba: Efficient Selective State Space Models with Mixture of Experts
Paper • 2401.04081 • Published • 73 -
SwitchHead: Accelerating Transformers with Mixture-of-Experts Attention
Paper • 2312.07987 • Published • 41 -
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
Paper • 2401.06066 • Published • 56
llm-models
llm-local
-
LLM in a flash: Efficient Large Language Model Inference with Limited Memory
Paper • 2312.11514 • Published • 260 -
Paused9797
Transformers to Core ML
⚡Display a loading screen with a spinner
-
enterprise-explorers/Llama-2-7b-chat-coreml
Text Generation • Updated • 5.54k • 138 -
tiiuae/falcon-7b-instruct
Text Generation • 7B • Updated • 94.1k • 1.02k
llm-datasets
llm-analysis
llm-local
-
LLM in a flash: Efficient Large Language Model Inference with Limited Memory
Paper • 2312.11514 • Published • 260 -
Paused9797
Transformers to Core ML
⚡Display a loading screen with a spinner
-
enterprise-explorers/Llama-2-7b-chat-coreml
Text Generation • Updated • 5.54k • 138 -
tiiuae/falcon-7b-instruct
Text Generation • 7B • Updated • 94.1k • 1.02k
papers
-
Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model
Paper • 2401.09417 • Published • 62 -
MoE-Mamba: Efficient Selective State Space Models with Mixture of Experts
Paper • 2401.04081 • Published • 73 -
SwitchHead: Accelerating Transformers with Mixture-of-Experts Attention
Paper • 2312.07987 • Published • 41 -
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
Paper • 2401.06066 • Published • 56
llm-datasets
llm-models